CN111833858A - Voice interaction state display method and device based on loudspeaker box - Google Patents

Voice interaction state display method and device based on loudspeaker box Download PDF

Info

Publication number
CN111833858A
CN111833858A CN201910310025.4A CN201910310025A CN111833858A CN 111833858 A CN111833858 A CN 111833858A CN 201910310025 A CN201910310025 A CN 201910310025A CN 111833858 A CN111833858 A CN 111833858A
Authority
CN
China
Prior art keywords
light
voice
sound box
user
preset
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201910310025.4A
Other languages
Chinese (zh)
Inventor
翟月
程晓雯
李士岩
张晓东
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Baidu Netcom Science and Technology Co Ltd
Shanghai Xiaodu Technology Co Ltd
Original Assignee
Beijing Baidu Netcom Science and Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Baidu Netcom Science and Technology Co Ltd filed Critical Beijing Baidu Netcom Science and Technology Co Ltd
Priority to CN201910310025.4A priority Critical patent/CN111833858A/en
Publication of CN111833858A publication Critical patent/CN111833858A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/08Speech classification or search
    • G10L15/18Speech classification or search using natural language modelling
    • G10L15/1822Parsing for meaning understanding
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R1/00Details of transducers, loudspeakers or microphones
    • H04R1/02Casings; Cabinets ; Supports therefor; Mountings therein
    • H04R1/028Casings; Cabinets ; Supports therefor; Mountings therein associated with devices performing functions other than acoustics, e.g. electric candles
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R3/00Circuits for transducers, loudspeakers or microphones
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • G10L2015/223Execution procedure of a spoken command
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • G10L2015/225Feedback of the input speech

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • Computational Linguistics (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Human Computer Interaction (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Artificial Intelligence (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

The invention provides a voice interaction state display method and device based on a sound box, wherein the method comprises the following steps: detecting whether a current scene meets a preset sound box awakening state or not, if so, acquiring a preset first light display parameter, and controlling the sound box light to display the sound box light to be in the awakening state according to the first light display parameter; acquiring preset second light display parameters after voice information input by a user is collected, controlling the light display of the sound box to be in a voice analysis state, performing semantic analysis on the voice information input by the user to acquire search keywords, and inquiring a preset instruction information base to judge whether response information corresponding to the search keywords can be acquired or not; if the response information corresponding to the search keyword can be acquired, the preset third light display parameter is acquired, the light of the sound box is controlled to display a voice broadcast state according to the third light display parameter, and the response information is played to the user. Therefore, the voice interaction state is visually prompted, and the human-computer interaction experience is improved.

Description

Voice interaction state display method and device based on loudspeaker box
Technical Field
The invention relates to the technical field of information prompt, in particular to a voice interaction state display method and device based on a sound box.
Background
Along with the development of computer technology, the intelligent sound box gradually enters the life of people, and the functions of the intelligent sound box are enriched, wherein the intelligent sound box can perform voice interaction with a user so as to provide intelligent services for the user.
In the related art, the smart speaker only provides voice-related services, such as receiving the voice of the user, performing voice response according to the voice of the user, and the like, but the voice service mode is too rigid, and the man-machine interaction between the user and the speaker is not strong.
Disclosure of Invention
The present invention is directed to solving, at least to some extent, one of the technical problems in the related art described above.
Therefore, a first objective of the present invention is to provide a method for displaying a voice interaction state based on a sound box, so as to solve the technical problem that a user cannot know a current voice interaction state of the sound box in the prior art.
The second purpose of the invention is to provide a voice interaction state display device based on a sound box.
A third object of the invention is to propose an electronic device.
A fourth object of the invention is to propose a non-transitory computer-readable storage medium.
In order to achieve the above object, an embodiment of a first aspect of the present invention provides a method for displaying a voice interaction state based on a speaker, including the following steps: detecting whether a current scene meets a preset sound box awakening state or not, if so, acquiring a preset first light display parameter, and controlling the sound box light to display in the awakening state according to the first light display parameter; acquiring preset second light display parameters after voice information input by a user is collected, controlling the sound box light to be in a voice analysis state according to the second light display parameters, performing semantic analysis on the voice information input by the user to acquire search keywords, and inquiring a preset instruction information base to judge whether response information corresponding to the search keywords can be acquired or not; if the response information corresponding to the search keyword can be acquired, acquiring a preset third light display parameter, controlling the light display of the sound box to be in a voice broadcast state according to the third light display parameter, and playing the response information to the user.
The embodiment of the second aspect of the present invention provides a voice interaction state display device based on a sound box, including: the detection module is used for detecting whether the current scene meets a preset sound box awakening state or not; the first control module is used for acquiring a preset first light display parameter when the sound box awakening state is met, and controlling the sound box light to display the sound box light to be in the awakening state according to the first light display parameter; the second control module is used for acquiring preset second light display parameters after voice information input by a user is acquired, and controlling the sound box light to display in a voice analysis state according to the second light display parameters; the semantic recognition module is used for performing semantic analysis on the voice information input by the user to obtain search keywords, and inquiring a preset instruction information base to judge whether response information corresponding to the search keywords can be obtained or not; and the third control module is used for acquiring preset third light display parameters when response information corresponding to the search keywords can be acquired, controlling the light display of the sound box to be in a voice broadcast state according to the third light display parameters, and playing the response information to the user.
An embodiment of a third aspect of the present invention provides an electronic device, including a memory, a processor, and a computer program stored on the memory and executable on the processor, where the processor executes the computer program to implement the method for displaying a voice interaction state based on a sound box according to the embodiment of the first aspect.
A fourth aspect of the present invention provides a non-transitory computer-readable storage medium, on which a computer program is stored, where the computer program, when executed by a processor, implements the speaker-based voice interaction state display method according to the first aspect.
The technical scheme provided by the embodiment of the invention has the following beneficial effects:
different voice interaction states of the loudspeaker box are reminded based on different light display parameters, so that a user can clearly determine the voice interaction state of the loudspeaker box, when the user and the loudspeaker box perform voice interaction, the effect of conversation between the user and the loudspeaker box is formed, and human-computer interaction experience is improved.
Additional aspects and advantages of the invention will be set forth in part in the description which follows and, in part, will be obvious from the description, or may be learned by practice of the invention.
Drawings
The foregoing and/or additional aspects and advantages of the present invention will become apparent and readily appreciated from the following description of the embodiments, taken in conjunction with the accompanying drawings of which:
fig. 1 is a schematic flow chart of a voice interaction state display method based on a sound box according to an embodiment of the present invention;
FIG. 2 is a schematic view of a light element according to one embodiment of the present invention;
fig. 3 is a schematic view of an application scenario of a first method for displaying a voice interaction state based on a sound box according to an embodiment of the present invention;
fig. 4 is a schematic view of an application scenario of a second speaker-based voice interaction state display method according to an embodiment of the present invention;
fig. 5 is a schematic view of an application scenario of a third method for displaying a voice interaction state based on a sound box according to an embodiment of the present invention;
fig. 6 is a schematic view of an application scenario of a fourth method for displaying a voice interaction state based on a sound box according to an embodiment of the present invention;
FIG. 7 is a schematic structural diagram of a voice interaction state display device based on a speaker according to an embodiment of the present invention; and
fig. 8 is a schematic structural diagram of a voice interaction state display device based on a loudspeaker box according to another embodiment of the invention.
Detailed Description
Reference will now be made in detail to embodiments of the present invention, examples of which are illustrated in the accompanying drawings, wherein like or similar reference numerals refer to the same or similar elements or elements having the same or similar function throughout. The embodiments described below with reference to the drawings are illustrative and intended to be illustrative of the invention and are not to be construed as limiting the invention.
The following describes a method and an apparatus for displaying a voice interaction state based on a speaker according to an embodiment of the present invention with reference to the accompanying drawings.
In view of the technical problems mentioned in the background art that the current intelligent sound box is in a state and the user knows that the mode is relatively rigid, the invention provides a technical scheme for assisting the user to understand the sound box voice recognition state more vividly and helping the user to operate quickly by using light, wherein the light in the intelligent sound box is controlled to form a mode which can provide more information and better express user experience intuitively to the user by changing colors, lighting positions, dynamic effects of lamps and the like, and the user can have better visual experience, and acquiring the current specific working state of the sound box.
Specifically, fig. 1 is a schematic flow chart of a voice interaction state display method based on a sound box according to an embodiment of the present invention. As shown in fig. 1, the method includes:
step 101, detecting whether a current scene meets a preset sound box awakening state, if so, acquiring a preset first light display parameter, and controlling the sound box light to display the sound box light to be in the awakening state according to the first light display parameter.
It should be noted that, according to different wake-up mechanisms of the sound boxes, it is detected whether a current scene meets different preset sound box wake-up state modes, as a possible implementation mode, after receiving a wake-up instruction, user face information is detected, where the wake-up instruction may be sent by triggering a corresponding wake-up control, or a voice instruction including a wake-up keyword is sent by a user, after receiving the wake-up instruction, in order to ensure validity of sound box control, it is necessary to know the user face information, for example, the user face information is collected by a camera in the sound box, and then it is determined whether the user face information is legal face information, that is, the user face information is matched with the preset legal face information, and it is determined whether the user face information is legal face information according to a matching result.
Particularly, if the current scene is detected to meet the preset sound box awakening state, the preset first light display parameter is acquired, the sound box light display is controlled to be in the awakening state according to the first light display parameter, therefore, the user can know that the current sound box is in the awakening state, and the user is prevented from being awakened repeatedly.
The lighting display parameters comprise one or more of lighting display position, lighting display frequency, lighting display period, lighting color, lighting display intensity and the like.
It should be understood that, in the embodiment of the present invention, the light display parameter corresponds to the voice interaction state, so that the light display parameter has the characteristics of scene and semantic, the light display parameter and the voice interaction state have an obvious semantic corresponding relationship, light can be visually conveyed to a user to obtain clear information, and the user can know the voice interaction state of the current smart sound box based on the light display parameter.
Obviously, the intelligent sound box has a light element therein, wherein, as shown in fig. 2, the light element may be a circular LED strip, or a light array, etc., without limitation.
Specifically, the reminding mode of the voice interaction state commonly used on other terminal equipment can be learned according to a large amount of experimental data, and the light display parameters which are more convenient for the user to understand are determined according to the learning of the reminding mode.
As a possible implementation manner, the determined first lighting display parameter includes two lighting display colors, an initial display position corresponding to one of the lighting display colors, a first display path and a second display path corresponding to the lighting display color and a display speed corresponding to the two display paths according to the initial position, and a first display position sequence and a second display position sequence corresponding to another lighting display color, so that the user is prompted that the current loudspeaker box is in the wake-up state based on the first lighting display parameter.
For example, as shown in fig. 3, when the light element is a circular ring-shaped light strip, and when the two colors include blue and pink (not shown in the figure, the difference in gray values represents the difference in color), the opposite position of the closest position of the received wake-up command is determined as the initial position, the blue light is controlled to converge from the initial position to both sides until the closest position of the received wake-up command is converged, and meanwhile, the two arc lights in pink converge to the closest position of the received wake-up command in two directions.
In an embodiment of the invention, in order to further improve the human-computer interaction effect and facilitate the operation of the user on the sound box, after the sound box is displayed in the awakening state, an awakening guide voice can be played to the user to prompt the user to take in voice, and the guide voice can be set in an anthropomorphic manner to provide good interaction experience for the user. For example, a "lovely host, i.e., ready to ask you what you need or to ask me" is played to the user. Of course, this guiding manner can also be realized by the lamp tube display parameters, which are not illustrated here.
Step 102, acquiring preset second light display parameters after voice information input by a user is collected, controlling the light display of the sound box to be in a voice analysis state according to the second light display parameters, performing semantic analysis on the voice information input by the user to acquire search keywords, and inquiring a preset instruction information base to judge whether response information corresponding to the search keywords can be acquired.
Specifically, after voice input by a user is collected, a preset second light display parameter is obtained, and the light display of the sound box is controlled to be in a voice analysis state according to the second light display parameter, so that the user cannot feel that the sound box is in a non-response state.
As a possible implementation, two colors corresponding to the voice analysis state, and the display position of one color, and the display position sequence and display position sequence period of the other color are determined, whereby in the voice analysis state, the sound box is indicated in the voice analysis state with a load lighting effect.
For example, as shown in fig. 4, when the light element is a circular ring-shaped light strip, and when the two colors include blue and pink (not shown in the figure, the difference of the gray values represents the difference of the colors), the display mode of blue is determined to display the whole circular ring, and the red light is displayed circularly in the form of a circular arc along the pointer.
Furthermore, the sound box performs semantic analysis on voice information input by a user to obtain search keywords, and queries a preset instruction information base to judge whether response information corresponding to the search keywords can be obtained, namely, a preset execution information base, wherein the instruction information base stores the corresponding relation between the search keywords and the response information, and the response information is further obtained based on the search keywords.
In an embodiment of the present invention, in order to clearly define the current state of the sound box for the user, the voice receiving state is also defined based on the light display parameter to indicate that the user is currently receiving the input voice, and the user is requested to continue inputting, that is, before the voice information input by the user is collected, when the voice input by the user is detected, a preset fourth light display parameter is obtained, and the light display of the sound box is controlled to be the voice receiving state according to the fourth light display parameter.
As a possible implementation manner, the determining the fourth lighting display parameter includes: and determining two colors corresponding to the voice analysis state, the display position of one color, the display path sequence of the other color and the display period corresponding to the display path sequence, thereby indicating that the sound box is in the voice receiving state with the breathing light effect in the voice analysis state.
For example, as shown in fig. 5, when the light element is a circular ring-shaped light strip, and when the two colors include blue and pink (not shown in the figure, the difference of gray values represents the difference of colors), the display mode of blue is determined to display the whole circular ring, the display mode of red light extends from an initial position to two sides with a light spot, and the device is covered on the whole circular ring and then returns to a light spot again, and this cycle forms a breathing effect to indicate that the speaker box is receiving voice.
Step 103, if the response information corresponding to the search keyword can be obtained, obtaining a preset third light display parameter, controlling the light display of the sound box to be in a voice broadcast state according to the third light display parameter, and playing the response information to the user.
Specifically, if the response information corresponding to the search keyword can be acquired, the preset third light display parameter is acquired, the light display of the sound box is controlled to be in a voice broadcast state according to the third light display parameter, the response information is played to the user, and therefore the user is indicated to be currently in voice broadcast.
As a possible implementation method, the determining the third display light parameter includes: confirm two colours that correspond with the speech analysis state, confirm the light position of every colour and show the sequence, show the start time and show frequency, and then, when voice broadcast, instruct the audio amplifier to be in the voice broadcast state with breathing lamp effect in turn.
For example, as shown in fig. 6, when the light element is a circular ring-shaped light strip, and when the two colors include blue and pink (not shown in the figure, the difference in gray scale values represents the difference in color), the light position display sequence for determining blue is a left-right symmetric arc, and an up-down symmetric arc, and the light position display sequence for determining red is a left-right symmetric arc, and an up-down symmetric arc, the blue circular arc-shaped light strip and the pink circular arc-shaped light strip are alternately breathed based on the setting of the display start time and the display frequency.
In an embodiment of the present invention, if the response information corresponding to the search keyword cannot be obtained, a preset fifth lighting display parameter is obtained, and the lighting display of the sound box is controlled to be the query response failure state according to the fifth lighting display parameter, so that the user can know that the response cannot be performed based on the control display of the fifth lighting display parameter on the lighting, wherein the fifth lighting display parameter may be set according to an application scene, and is not limited herein.
It should be noted that, in order to make the above dynamic effect more clear, in the above example, the determination of the colors of the two lights needs to be determined according to the color contrast, so as to ensure that the color contrast of the two lights is higher than a certain value.
In summary, the sound box-based voice interaction state display method provided by the embodiment of the invention reminds different voice interaction states of the sound box based on different light display parameters, so that a user can clearly determine the voice interaction state of the sound box, a conversation effect between the user and the sound box is formed when the user and the sound box perform voice interaction, and human-computer interaction experience is improved.
In order to implement the above embodiments, the present invention further provides a voice interaction state display device based on a speaker.
Fig. 7 is a schematic structural diagram of a speaker-based voice interaction state display device according to an embodiment of the present invention, and as shown in fig. 7, the speaker-based voice interaction state display device includes: a detection module 10, a first control module 20, a second control module 30, a semantic recognition module 40, and a third control module 50, wherein,
and the detection module 10 is configured to detect whether the current scene meets a preset sound box wake-up state.
In an embodiment of the present invention, the detecting module 10 is specifically configured to detect the user face information after receiving the wake-up instruction, and determine whether the user face information is legal face information.
And the first control module 20 is used for acquiring a preset first light display parameter when the sound box awakening state is met, and controlling the sound box light to display the sound box light to be in the awakening state according to the first light display parameter.
And the second control module 30 is used for acquiring preset second light display parameters after acquiring the voice information input by the user, and controlling the light display of the sound box to be in a voice analysis state according to the second light display parameters.
And the semantic recognition module 40 is configured to perform semantic analysis on the voice information input by the user to obtain a search keyword, and query a preset instruction information base to determine whether response information corresponding to the search keyword can be obtained.
And the third control module 50 is configured to, when response information corresponding to the search keyword can be acquired, acquire a preset third light display parameter, control the sound box light to display a voice broadcast state according to the third light display parameter, and play the response information to the user.
Fig. 8 is a schematic structural diagram of a voice interaction state display device based on a sound box according to another embodiment of the present invention, as shown in fig. 8, on the basis of fig. 7, the device further includes a fourth control module 60, wherein,
and the fourth control module 60 is configured to, when detecting that the user performs voice input, acquire a preset fourth light display parameter, and control the light display of the sound box to be in a voice receiving state according to the fourth light display parameter.
It should be noted that the foregoing explanation of the embodiment of the method for displaying a voice interaction state based on a sound box is also applicable to the device for displaying a voice interaction state based on a sound box of the embodiment, and is not repeated here.
In summary, the voice interaction state display device based on the sound box according to the embodiment of the invention reminds different voice interaction states of the sound box based on different light display parameters, so that a user can clearly determine the voice interaction state of the sound box, a conversation effect between the user and the sound box is formed when the user and the sound box perform voice interaction, and human-computer interaction experience is improved.
In order to implement the foregoing embodiments, the present invention further provides an electronic device, which includes a memory, a processor, and a computer program stored in the memory and executable on the processor, and when the processor executes the computer program, the method for displaying a voice interaction state based on a speaker as described in the foregoing embodiments is implemented.
In order to implement the above embodiments, the present invention further proposes a non-transitory computer-readable storage medium, on which a computer program is stored, which, when being executed by a processor, implements the loudspeaker-based voice interaction state display method as described in the foregoing method embodiments.
In the description herein, references to the description of the term "one embodiment," "some embodiments," "an example," "a specific example," or "some examples," etc., mean that a particular feature, structure, material, or characteristic described in connection with the embodiment or example is included in at least one embodiment or example of the invention. In this specification, the schematic representations of the terms used above are not necessarily intended to refer to the same embodiment or example. Furthermore, the particular features, structures, materials, or characteristics described may be combined in any suitable manner in any one or more embodiments or examples. Furthermore, various embodiments or examples and features of different embodiments or examples described in this specification can be combined and combined by one skilled in the art without contradiction.
Furthermore, the terms "first", "second" and "first" are used for descriptive purposes only and are not to be construed as indicating or implying relative importance or implicitly indicating the number of technical features indicated. Thus, a feature defined as "first" or "second" may explicitly or implicitly include at least one such feature. In the description of the present invention, "a plurality" means at least two, e.g., two, three, etc., unless specifically limited otherwise.
Any process or method descriptions in flow charts or otherwise described herein may be understood as representing modules, segments, or portions of code which include one or more executable instructions for implementing steps of a custom logic function or process, and alternate implementations are included within the scope of the preferred embodiment of the present invention in which functions may be executed out of order from that shown or discussed, including substantially concurrently or in reverse order, depending on the functionality involved, as would be understood by those reasonably skilled in the art of the present invention.
The logic and/or steps represented in the flowcharts or otherwise described herein, e.g., an ordered listing of executable instructions that can be considered to implement logical functions, can be embodied in any computer-readable medium for use by or in connection with an instruction execution system, apparatus, or device, such as a computer-based system, processor-containing system, or other system that can fetch the instructions from the instruction execution system, apparatus, or device and execute the instructions. For the purposes of this description, a "computer-readable medium" can be any means that can contain, store, communicate, propagate, or transport the program for use by or in connection with the instruction execution system, apparatus, or device. More specific examples (a non-exhaustive list) of the computer-readable medium would include the following: an electrical connection (electronic device) having one or more wires, a portable computer diskette (magnetic device), a Random Access Memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or flash memory), an optical fiber device, and a portable compact disc read-only memory (CDROM). Additionally, the computer-readable medium could even be paper or another suitable medium upon which the program is printed, as the program can be electronically captured, via for instance optical scanning of the paper or other medium, then compiled, interpreted or otherwise processed in a suitable manner if necessary, and then stored in a computer memory.
It should be understood that portions of the present invention may be implemented in hardware, software, firmware, or a combination thereof. In the above embodiments, the various steps or methods may be implemented in software or firmware stored in memory and executed by a suitable instruction execution system. If implemented in hardware, as in another embodiment, any one or combination of the following techniques, which are known in the art, may be used: a discrete logic circuit having a logic gate circuit for implementing a logic function on a data signal, an application specific integrated circuit having an appropriate combinational logic gate circuit, a Programmable Gate Array (PGA), a Field Programmable Gate Array (FPGA), or the like.
It will be understood by those skilled in the art that all or part of the steps carried by the method for implementing the above embodiments may be implemented by hardware related to instructions of a program, which may be stored in a computer readable storage medium, and when the program is executed, the program includes one or a combination of the steps of the method embodiments.
In addition, functional units in the embodiments of the present invention may be integrated into one processing module, or each unit may exist alone physically, or two or more units are integrated into one module. The integrated module can be realized in a hardware mode, and can also be realized in a software functional module mode. The integrated module, if implemented in the form of a software functional module and sold or used as a stand-alone product, may also be stored in a computer readable storage medium.
The storage medium mentioned above may be a read-only memory, a magnetic or optical disk, etc. Although embodiments of the present invention have been shown and described above, it is understood that the above embodiments are exemplary and should not be construed as limiting the present invention, and that variations, modifications, substitutions and alterations can be made to the above embodiments by those of ordinary skill in the art within the scope of the present invention.

Claims (10)

1. A voice interaction state display method based on a sound box is characterized by comprising the following steps:
detecting whether a current scene meets a preset sound box awakening state or not, if so, acquiring a preset first light display parameter, and controlling the sound box light to display in the awakening state according to the first light display parameter;
acquiring preset second light display parameters after voice information input by a user is collected, controlling the sound box light to be in a voice analysis state according to the second light display parameters, performing semantic analysis on the voice information input by the user to acquire search keywords, and inquiring a preset instruction information base to judge whether response information corresponding to the search keywords can be acquired or not;
if the response information corresponding to the search keyword can be acquired, acquiring a preset third light display parameter, controlling the light display of the sound box to be in a voice broadcast state according to the third light display parameter, and playing the response information to the user.
2. The method of claim 1, wherein the detecting whether the current scene satisfies a preset speaker wakeup state comprises:
after receiving a wake-up instruction, detecting user face information;
it is determined whether the user face information is legitimate face information.
3. The method of claim 1, prior to said collecting user-entered speech information, further comprising:
and when detecting that the user inputs voice, acquiring a preset fourth light display parameter, and controlling the sound box light to display a voice receiving state according to the fourth light display parameter.
4. The method of claim 1, wherein after controlling the speaker light to display in the wake-up state according to the first light display parameter, further comprising:
and playing a wake-up guide voice to the user and prompting the user to perform voice input.
5. The method of claim 1, after the querying a preset instruction information base to determine whether response information corresponding to the search keyword can be acquired, further comprising:
and if the response information corresponding to the search keyword cannot be acquired, acquiring a preset fifth light display parameter, and controlling the light display of the sound box to be in a query response failure state according to the fifth light display parameter.
6. The utility model provides a voice interaction state display device based on audio amplifier which characterized in that includes:
the detection module is used for detecting whether the current scene meets a preset sound box awakening state or not;
the first control module is used for acquiring a preset first light display parameter when the sound box awakening state is met, and controlling the sound box light to display the sound box light to be in the awakening state according to the first light display parameter;
the second control module is used for acquiring preset second light display parameters after voice information input by a user is acquired, and controlling the sound box light to display in a voice analysis state according to the second light display parameters;
the semantic recognition module is used for performing semantic analysis on the voice information input by the user to obtain search keywords, and inquiring a preset instruction information base to judge whether response information corresponding to the search keywords can be obtained or not;
and the third control module is used for acquiring preset third light display parameters when response information corresponding to the search keywords can be acquired, controlling the light display of the sound box to be in a voice broadcast state according to the third light display parameters, and playing the response information to the user.
7. The apparatus of claim 6, wherein the detection module is specifically configured to:
after receiving a wake-up instruction, detecting user face information;
it is determined whether the user face information is legitimate face information.
8. The apparatus of claim 6, further comprising:
and the fourth control module is used for acquiring a preset fourth light display parameter when detecting that the user inputs voice, and controlling the sound box light to display a voice receiving state according to the fourth light display parameter.
9. An electronic device comprising a memory, a processor, and a computer program stored on the memory and executable on the processor, wherein the processor, when executing the computer program, implements the speaker-based voice interaction state display method according to any one of claims 1-5.
10. A non-transitory computer readable storage medium having stored thereon a computer program, wherein the computer program, when executed by a processor, implements the speaker-based voice interaction state display method according to any one of claims 1-5.
CN201910310025.4A 2019-04-17 2019-04-17 Voice interaction state display method and device based on loudspeaker box Pending CN111833858A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910310025.4A CN111833858A (en) 2019-04-17 2019-04-17 Voice interaction state display method and device based on loudspeaker box

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910310025.4A CN111833858A (en) 2019-04-17 2019-04-17 Voice interaction state display method and device based on loudspeaker box

Publications (1)

Publication Number Publication Date
CN111833858A true CN111833858A (en) 2020-10-27

Family

ID=72915504

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910310025.4A Pending CN111833858A (en) 2019-04-17 2019-04-17 Voice interaction state display method and device based on loudspeaker box

Country Status (1)

Country Link
CN (1) CN111833858A (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112998538A (en) * 2021-03-10 2021-06-22 上海松下微波炉有限公司 Steaming and baking oven
WO2023160087A1 (en) * 2022-02-28 2023-08-31 海信视像科技股份有限公司 Prompting method for response state of voice instruction and display device

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106297780A (en) * 2015-06-03 2017-01-04 深圳市轻生活科技有限公司 A kind of voice interactive method and system and Intelligent voice broadcasting terminal
CN106448672A (en) * 2016-10-27 2017-02-22 Tcl通力电子(惠州)有限公司 Sound system and control method
CN107180631A (en) * 2017-05-24 2017-09-19 刘平舟 A kind of voice interactive method and device
CN107369446A (en) * 2017-06-28 2017-11-21 北京小米移动软件有限公司 Handle state prompt method, device and computer-readable recording medium
CN108039172A (en) * 2017-12-01 2018-05-15 Tcl通力电子(惠州)有限公司 Smart bluetooth speaker voice interactive method, smart bluetooth speaker and storage medium
CN109410951A (en) * 2018-11-21 2019-03-01 广州番禺巨大汽车音响设备有限公司 Audio controlling method, system and stereo set based on Alexa voice control
CN109471537A (en) * 2017-09-08 2019-03-15 腾讯科技(深圳)有限公司 Pronunciation inputting method, device, computer equipment and storage medium

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106297780A (en) * 2015-06-03 2017-01-04 深圳市轻生活科技有限公司 A kind of voice interactive method and system and Intelligent voice broadcasting terminal
CN106448672A (en) * 2016-10-27 2017-02-22 Tcl通力电子(惠州)有限公司 Sound system and control method
CN107180631A (en) * 2017-05-24 2017-09-19 刘平舟 A kind of voice interactive method and device
CN107369446A (en) * 2017-06-28 2017-11-21 北京小米移动软件有限公司 Handle state prompt method, device and computer-readable recording medium
CN109471537A (en) * 2017-09-08 2019-03-15 腾讯科技(深圳)有限公司 Pronunciation inputting method, device, computer equipment and storage medium
CN108039172A (en) * 2017-12-01 2018-05-15 Tcl通力电子(惠州)有限公司 Smart bluetooth speaker voice interactive method, smart bluetooth speaker and storage medium
CN109410951A (en) * 2018-11-21 2019-03-01 广州番禺巨大汽车音响设备有限公司 Audio controlling method, system and stereo set based on Alexa voice control

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112998538A (en) * 2021-03-10 2021-06-22 上海松下微波炉有限公司 Steaming and baking oven
WO2023160087A1 (en) * 2022-02-28 2023-08-31 海信视像科技股份有限公司 Prompting method for response state of voice instruction and display device

Similar Documents

Publication Publication Date Title
CN108009303B (en) Search method and device based on voice recognition, electronic equipment and storage medium
US10657959B2 (en) Information processing device, information processing method, and program
CN113460070B (en) Vehicle control method and device
WO2017141530A1 (en) Information processing device, information processing method and program
CN111833858A (en) Voice interaction state display method and device based on loudspeaker box
KR101763717B1 (en) Smart Watch To Change The Display Color Based On User's State
CN109863744B (en) Light output system
US20200050074A1 (en) Systems and methods to transform events and/or mood associated with playing media into lighting effects
CN106303018A (en) A kind of alarm clock control method and terminal
CN109559488B (en) Control method, remote control terminal, household appliance, system and storage medium
CN112331195B (en) Voice interaction method, device and system
CN109166571A (en) Wake-up word training method, device and the household appliance of household appliance
CN112735411A (en) Control method, client, vehicle, voice system, and storage medium
CN110871813A (en) Control method and device of virtual robot, vehicle, equipment and storage medium
CN112185415A (en) Sound visualization method and device, storage medium and MR mixed reality equipment
CN105979378B (en) A kind of control method and system of player fast forwarding and fast rewinding
CN109416905A (en) Performance assistant apparatus and method
CN111820479A (en) Control method and device of electronic cigarette
CN109508403A (en) Matching process, device and the vehicle intelligent controller of vehicle-mounted music
CN110491384B (en) Voice data processing method and device
CN105654971B (en) Singing assisting method and equipment in song recording process
CN112017651A (en) Voice control method and device of electronic equipment, computer equipment and storage medium
CN201731425U (en) LED intelligent voice tree lamp
CN113771738A (en) Vehicle-mounted light control method, device, system, equipment and storage medium
CN111045636A (en) Vehicle function display method and system

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
TA01 Transfer of patent application right

Effective date of registration: 20210508

Address after: 100085 Baidu Building, 10 Shangdi Tenth Street, Haidian District, Beijing

Applicant after: BEIJING BAIDU NETCOM SCIENCE AND TECHNOLOGY Co.,Ltd.

Applicant after: Shanghai Xiaodu Technology Co.,Ltd.

Address before: 100085 Baidu Building, 10 Shangdi Tenth Street, Haidian District, Beijing

Applicant before: BEIJING BAIDU NETCOM SCIENCE AND TECHNOLOGY Co.,Ltd.

TA01 Transfer of patent application right