CN110287365B - Data processing method and electronic equipment - Google Patents

Data processing method and electronic equipment Download PDF

Info

Publication number
CN110287365B
CN110287365B CN201910581137.3A CN201910581137A CN110287365B CN 110287365 B CN110287365 B CN 110287365B CN 201910581137 A CN201910581137 A CN 201910581137A CN 110287365 B CN110287365 B CN 110287365B
Authority
CN
China
Prior art keywords
voice data
application
voice
retrieval instruction
keyword information
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201910581137.3A
Other languages
Chinese (zh)
Other versions
CN110287365A (en
Inventor
徐培来
汪俊杰
张传良
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Lenovo Beijing Ltd
Original Assignee
Lenovo Beijing Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Lenovo Beijing Ltd filed Critical Lenovo Beijing Ltd
Priority to CN201910581137.3A priority Critical patent/CN110287365B/en
Publication of CN110287365A publication Critical patent/CN110287365A/en
Application granted granted Critical
Publication of CN110287365B publication Critical patent/CN110287365B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/60Information retrieval; Database structures therefor; File system structures therefor of audio data
    • G06F16/63Querying
    • G06F16/635Filtering based on additional data, e.g. user or group profiles
    • G06F16/637Administration of user profiles, e.g. generation, initialization, adaptation or distribution
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/60Information retrieval; Database structures therefor; File system structures therefor of audio data
    • G06F16/63Querying
    • G06F16/638Presentation of query results
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/60Information retrieval; Database structures therefor; File system structures therefor of audio data
    • G06F16/68Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/686Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using information manually generated, e.g. tags, keywords, comments, title or artist information, time, location or usage information, user ratings

Abstract

The present disclosure discloses a data processing method and an electronic device, wherein the method is applied to the electronic device, the electronic device at least comprises an application capable of realizing voice interaction, and the method comprises the following steps: receiving a voice data retrieval instruction, and acquiring keyword information corresponding to the voice data retrieval instruction; and responding to the voice data retrieval instruction, retrieving the voice data in the application based on the keyword information, and outputting a retrieval result. The method and the device can effectively search the voice content in the application capable of realizing voice interaction, and improve user experience.

Description

Data processing method and electronic equipment
Technical Field
The present disclosure relates to the field of electronic technologies, and in particular, to a data processing method and an electronic device.
Background
With the popularization of intelligent terminal equipment and the wide use of voice interaction modes, more and more voice contents are owned by users on the intelligent terminal equipment.
Therefore, how to effectively search the voice content is an urgent problem to be solved.
Disclosure of Invention
In view of this, the present disclosure provides the following technical solutions:
a data processing method is applied to electronic equipment, wherein the electronic equipment at least comprises an application capable of realizing voice interaction, and the method comprises the following steps:
receiving a voice data retrieval instruction;
acquiring keyword information corresponding to the voice data retrieval instruction;
and responding to the voice data retrieval instruction, retrieving the voice data in the application based on the keyword information, and outputting a retrieval result.
Preferably, the retrieving, in response to the voice data retrieval instruction, the voice data in the application based on the keyword information and outputting a retrieval result includes:
responding to the voice data retrieval instruction, and triggering the voice data in the application to carry out playing operation;
acquiring voice data in the process of playing the voice data in the application;
and matching the information content of each piece of voice data based on the keyword information, and outputting a retrieval result.
Preferably, the triggering, in response to the voice data retrieval instruction, the voice data in the application to perform a play operation includes:
and responding to the voice data retrieval instruction, and controlling the playing volume of the voice data in the application to be smaller than a preset volume threshold when the voice data in the application is triggered to be played.
Preferably, the triggering, in response to the voice data retrieval instruction, the voice data in the application to perform a play operation further includes:
responding to the voice data retrieval instruction, and controlling the playing speed of the voice data in the application to be larger than a first preset threshold when the voice data in the application is triggered to be played.
Preferably, the triggering, in response to the voice data retrieval instruction, the voice data in the application to perform a play operation includes:
and responding to the voice data retrieval instruction, and controlling the voice data in the application to be played in the target equipment when the voice data in the application is triggered to be played, wherein the volume of the voice data in the application played by the target equipment is zero.
Preferably, the triggering, in response to the voice data retrieval instruction, the voice data in the application to perform a play operation further includes:
and controlling the playing speed of the voice data in the application to be played in the target equipment to be greater than a second preset threshold value.
Preferably, the acquiring of the keyword information corresponding to the voice data retrieval instruction includes:
and acquiring voice keyword information or text keyword information corresponding to the voice data retrieval instruction.
Preferably, the matching in the information content of each piece of voice data based on the keyword information and outputting a search result include:
converting each piece of acquired voice data into corresponding text information;
and matching the text keyword information with the text information obtained by converting each piece of voice data, and outputting a retrieval result.
An electronic device including at least one application capable of voice interaction, the electronic device comprising:
the receiving module is used for receiving a voice data retrieval instruction;
the acquisition module is used for acquiring keyword information corresponding to the voice data retrieval instruction;
and the retrieval module is used for responding to the voice data retrieval instruction, retrieving the voice data in the application based on the keyword information and outputting a retrieval result.
Preferably, when the retrieval module executes the retrieval instruction in response to the voice data, retrieves the voice data in the application based on the keyword information, and outputs a retrieval result, the retrieval module is specifically configured to:
responding to the voice data retrieval instruction, and triggering the voice data in the application to carry out playing operation;
acquiring voice data in the process of playing the voice data in the application;
and matching the information content of each piece of voice data based on the keyword information, and outputting a retrieval result.
Preferably, when the retrieval module executes the voice data retrieval instruction in response to the voice data retrieval instruction and triggers the voice data in the application to perform the play operation, the retrieval module is specifically configured to:
and responding to the voice data retrieval instruction, and controlling the playing volume of the voice data in the application to be smaller than a preset volume threshold when the voice data in the application is triggered to be played.
Preferably, when the retrieval module executes the operation of triggering the voice data in the application to play in response to the voice data retrieval instruction, the retrieval module is further configured to:
responding to the voice data retrieval instruction, and controlling the playing speed of the voice data in the application to be larger than a first preset threshold when the voice data in the application is triggered to be played.
Preferably, when the retrieval module executes the voice data retrieval instruction in response to the voice data retrieval instruction and triggers the voice data in the application to perform the play operation, the retrieval module is specifically configured to:
and responding to the voice data retrieval instruction, and controlling the voice data in the application to be played in the target equipment when the voice data in the application is triggered to be played, wherein the volume of the voice data in the application played by the target equipment is zero.
Preferably, when the retrieval module executes the operation of triggering the voice data in the application to play in response to the voice data retrieval instruction, the retrieval module is further configured to:
and controlling the playing speed of the voice data in the application to be played in the target equipment to be greater than a second preset threshold value.
Preferably, when executing the process of acquiring the keyword information corresponding to the voice data retrieval instruction, the acquiring module is specifically configured to:
and acquiring voice keyword information or text keyword information corresponding to the voice data retrieval instruction.
Preferably, when the retrieval module performs matching in the information content of each piece of voice data based on the keyword information and outputs a retrieval result, the retrieval module is specifically configured to:
converting each piece of acquired voice data into corresponding text information;
and matching the text keyword information with the text information obtained by converting each piece of voice data, and outputting a retrieval result.
An electronic device including at least one application capable of voice interaction, the electronic device comprising: memory, a processor and a computer program stored on the memory and executable on the processor, the processor when executing the program for implementing:
receiving a voice data retrieval instruction;
acquiring keyword information corresponding to the voice data retrieval instruction;
and responding to the voice data retrieval instruction, retrieving the voice data in the application based on the keyword information, and outputting a retrieval result.
According to the technical scheme, the data processing method is applied to the electronic equipment, the electronic equipment at least comprises one application capable of realizing voice interaction, when data processing is carried out, the voice data retrieval instruction is received firstly, then the keyword information corresponding to the voice data retrieval instruction is obtained, the voice data retrieval instruction is responded, the voice data in the application is retrieved based on the keyword information, and the retrieval result is output. The method and the device can effectively search the voice content in the application capable of realizing voice interaction, and improve user experience.
Drawings
In order to more clearly illustrate the embodiments of the present disclosure or the technical solutions in the prior art, the drawings used in the description of the embodiments or the prior art will be briefly described below, it is obvious that the drawings in the following description are only embodiments of the present disclosure, and for those skilled in the art, other drawings can be obtained according to the provided drawings without creative efforts.
Fig. 1 is a method flowchart of an embodiment 1 of a data processing method disclosed in the present disclosure;
FIG. 2 is a flow chart of a method of an embodiment 2 of the data processing method disclosed in the present disclosure;
FIG. 3 is a flow chart of a method of embodiment 3 of a data processing method of the present disclosure;
FIG. 4 is a method flow diagram of embodiment 4 of a data processing method of the present disclosure;
fig. 5 is a schematic structural diagram of an electronic device embodiment 1 according to the present disclosure;
fig. 6 is a schematic structural diagram of an electronic device embodiment 2 according to the present disclosure;
fig. 7 is a schematic structural diagram of an electronic device embodiment 3 according to the present disclosure;
fig. 8 is a schematic structural diagram of an electronic device embodiment 4 according to the present disclosure;
fig. 9 is a schematic structural diagram of an electronic device embodiment 5 according to the present disclosure.
Detailed Description
The technical solutions in the embodiments of the present disclosure will be clearly and completely described below with reference to the drawings in the embodiments of the present disclosure, and it is obvious that the described embodiments are only a part of the embodiments of the present disclosure, and not all of the embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments disclosed herein without making any creative effort, shall fall within the protection scope of the present disclosure. The technical solutions in the embodiments of the present disclosure will be clearly and completely described below with reference to the drawings in the embodiments of the present disclosure, and it is obvious that the described embodiments are only a part of the embodiments of the present disclosure, and not all of the embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments disclosed herein without making any creative effort, shall fall within the protection scope of the present disclosure.
As shown in fig. 1, which is a flowchart of a method in an embodiment 1 of a data processing method disclosed in the present disclosure, the method is applied to an electronic device, where the electronic device may be a smart device such as a smart phone and a tablet computer, and the electronic device includes at least one application capable of implementing voice interaction, for example: applications such as "WeChat", "QQ", etc.; the method may comprise the steps of:
s101, receiving a voice data retrieval instruction;
when the voice content of the application capable of realizing voice interaction in the electronic equipment needs to be searched, a voice data retrieval instruction is received firstly, namely, an instruction for starting a voice data retrieval function is received. It should be noted that the voice data retrieval instruction may be generated by user triggering. For example, the user may input the related content that needs to be retrieved through a search bar of the electronic device, and generate a corresponding voice data retrieval instruction when the user inputs the related content that needs to be retrieved.
In a specific application scenario, taking an electronic device as a smart phone and an application capable of realizing voice interaction as WeChat as an example, when a user needs to search for voice content in the WeChat application in the electronic device, related content needing to be retrieved can be input through a search bar of the smart phone, for example, the user inputs a "recognition rate of a voice recognition engine" in the search bar, and when the user inputs the "recognition rate of the voice recognition engine", a corresponding voice data retrieval instruction is generated.
S102, acquiring keyword information corresponding to the voice data retrieval instruction;
after receiving the voice data retrieval instruction, further analyzing the voice data retrieval instruction to obtain keyword information corresponding to the voice data retrieval instruction. The acquired keyword information may be a part of the content for search input by the user or may be all of the content for search according to different search requirements. For example, when the content for performing the voice data search input by the user is "recognition rate of voice recognition engine", the keyword information may be content such as "voice recognition engine" or "recognition rate of engine", or may be "recognition rate of voice recognition engine". In addition, when the keyword information is a part of the content used for searching, the obtained related search result is wider, and when the keyword information is the whole content of the content used for searching, the obtained related search result is more accurate. Therefore, the keyword information corresponding to the voice data retrieval instruction can be flexibly set according to different requirements.
S103, responding to the voice data retrieval instruction, retrieving the voice data in the application based on the keyword information, and outputting a retrieval result.
And after the keyword information corresponding to the voice data retrieval instruction is acquired, further responding to the received voice data retrieval instruction, retrieving the voice data in the application according to the acquired keyword information, and finally outputting the retrieved result. For example, when the keyword information is "recognition rate of engine", the speech data in the WeChat application is searched for by "recognition rate of engine", and finally the search result "recognition rate of speech recognition engine" is output.
It should be noted that, when outputting the search result, the implementation manner may be to jump to the corresponding position of the voice data to be used, and further identify the searched voice data, which is convenient for the user to identify; the implementation mode can also be that the retrieved results are displayed in a list form; that is to say, when the retrieval result is output, a corresponding retrieval result presentation mode can be designed according to actual requirements.
In summary, in the above embodiments, when the voice data in the application needs to be retrieved, the voice data retrieval instruction is received first, then the keyword information corresponding to the voice data retrieval instruction is obtained, the voice data in the application is retrieved based on the keyword information in response to the voice data retrieval instruction, and the retrieval result is output. The method and the device can effectively search the voice content in the application capable of realizing voice interaction, and improve user experience.
As shown in fig. 2, which is a flowchart of a method in an embodiment 2 of a data processing method disclosed in the present disclosure, the method is applied to an electronic device, where the electronic device may be a smart device such as a smart phone and a tablet computer, and the electronic device includes at least one application capable of implementing voice interaction, for example: applications such as "WeChat", "QQ", etc.; the method may comprise the steps of:
s201, receiving a voice data retrieval instruction;
when the voice content of the application capable of realizing voice interaction in the electronic equipment needs to be searched, a voice data retrieval instruction is received firstly, namely, an instruction for starting a voice data retrieval function is received. It should be noted that the voice data retrieval instruction may be generated by user triggering. For example, the user may input the related content that needs to be retrieved through a search bar of the electronic device, and generate a corresponding voice data retrieval instruction when the user inputs the related content that needs to be retrieved.
In a specific application scenario, taking an electronic device as a smart phone and an application capable of realizing voice interaction as WeChat as an example, when a user needs to search for voice content in the WeChat application in the electronic device, related content needing to be retrieved can be input through a search bar of the smart phone, for example, the user inputs a "recognition rate of a voice recognition engine" in the search bar, and when the user inputs the "recognition rate of the voice recognition engine", a corresponding voice data retrieval instruction is generated.
S202, acquiring keyword information corresponding to the voice data retrieval instruction;
after receiving the voice data retrieval instruction, further analyzing the voice data retrieval instruction to obtain keyword information corresponding to the voice data retrieval instruction. The acquired keyword information may be a part of the content for search input by the user or may be all of the content for search according to different search requirements. For example, when the content for performing the voice data search input by the user is "recognition rate of voice recognition engine", the keyword information may be content such as "voice recognition engine" or "recognition rate of engine", or may be "recognition rate of voice recognition engine". In addition, when the keyword information is a part of the content used for searching, the obtained related search result is wider, and when the keyword information is the whole content of the content used for searching, the obtained related search result is more accurate. Therefore, the keyword information corresponding to the voice data retrieval instruction can be flexibly set according to different requirements.
S203, responding to the voice data retrieval instruction, and triggering the voice data in the application to carry out playing operation;
and after the keyword information corresponding to the voice data retrieval instruction is acquired, further responding to the received voice data retrieval instruction, and triggering the internal voice data in the application to carry out playing operation. For example, in a specific application scenario, a play operation is triggered to be performed on historical voice data in the WeChat application in response to a voice data retrieval instruction.
S204, acquiring voice data in the process of playing the voice data in the application;
in the process of triggering the playing operation of the voice data in the application, the audio stream in the application is made to flow in the code frame of the operating system, and in the process of flowing the audio stream, the audio data is obtained, namely the voice data is obtained.
And S205, matching the information content of each piece of voice data based on the keyword information, and outputting a retrieval result.
After the voice data in the application is acquired, the information content of each piece of acquired voice data is further matched according to the acquired keyword information, and finally, a matching result is output, namely a retrieval result is output.
For example, when the keyword information is "recognition rate of engine", the information content of each piece of acquired voice data is matched by "recognition rate of engine", and finally, a matching result "recognition rate of voice recognition engine" is output.
It should be noted that, when outputting the search result, the implementation manner may be to jump to the corresponding position of the voice data to be used, and further identify the searched voice data, which is convenient for the user to identify; the implementation mode can also be that the retrieved results are displayed in a list form; that is to say, when the retrieval result is output, a corresponding retrieval result presentation mode can be designed according to actual requirements.
In summary, in the above embodiments, when the voice data in the application needs to be retrieved, the voice data in the application can be triggered to perform a playing operation by responding to the voice data retrieval instruction, the voice data is obtained during the playing operation, the information content of each piece of voice data is matched according to the keyword information in the retrieval instruction, and a retrieval result is output, so that the voice content in the voice interactive application is searched, and the user experience is improved.
Specifically, in the above embodiment, the received voice data retrieval instruction may be generated by user triggering. For example, the user may input the related content to be retrieved through a search bar of the electronic device, or the user may input the related content to be retrieved through voice through a voice interaction function of the electronic device, and generate a corresponding voice data retrieval instruction when the user inputs the related content to be retrieved. That is, the voice data retrieval instruction may be generated based on text information input by the user or based on voice information input by the user.
Therefore, when the voice data retrieval instruction is generated based on text information input by a user, when keyword information corresponding to the voice data retrieval instruction is acquired, text keyword information corresponding to the voice data retrieval instruction is acquired; when the voice data retrieval instruction is generated based on voice information input by a user, the voice keyword information corresponding to the voice data retrieval instruction is acquired when the keyword information corresponding to the voice data retrieval instruction is acquired.
Specifically, in the above embodiment, when the keyword information corresponding to the acquired voice data retrieval instruction is the text keyword information corresponding to the acquired voice data retrieval instruction, correspondingly, matching is performed in the information content of each piece of voice data based on the keyword information, and when the retrieval result is output, each piece of voice data acquired during the process of playing the voice data in the application may be first converted into corresponding text information, and then the text keyword information may be matched with the text information obtained by converting each piece of voice data, and the retrieval result is output.
In addition, when the keyword information corresponding to the acquired voice data retrieval instruction is the voice keyword information corresponding to the acquired voice data retrieval instruction, and accordingly, when the information content of each piece of voice data is matched based on the keyword information and the retrieval result is output, the voice keyword information may be first converted into corresponding keyword text information, and meanwhile, each piece of voice data acquired during the playing operation of the voice data in the application is converted into corresponding text information, and then the keyword text information corresponding to the voice keyword information is matched with the text information obtained by converting each piece of voice data, and the retrieval result is output.
As shown in fig. 3, which is a flowchart of a method in embodiment 3 of a data processing method disclosed in the present disclosure, the method is applied to an electronic device, where the electronic device may be a smart device such as a smart phone and a tablet computer, and the electronic device includes at least one application capable of implementing voice interaction, for example: applications such as "WeChat", "QQ", etc.; the method may comprise the steps of:
s301, receiving a voice data retrieval instruction;
when the voice content of the application capable of realizing voice interaction in the electronic equipment needs to be searched, a voice data retrieval instruction is received firstly, namely, an instruction for starting a voice data retrieval function is received. It should be noted that the voice data retrieval instruction may be generated by user triggering. For example, the user may input the related content that needs to be retrieved through a search bar of the electronic device, and generate a corresponding voice data retrieval instruction when the user inputs the related content that needs to be retrieved.
In a specific application scenario, taking an electronic device as a smart phone and an application capable of realizing voice interaction as WeChat as an example, when a user needs to search for voice content in the WeChat application in the electronic device, related content needing to be retrieved can be input through a search bar of the smart phone, for example, the user inputs a "recognition rate of a voice recognition engine" in the search bar, and when the user inputs the "recognition rate of the voice recognition engine", a corresponding voice data retrieval instruction is generated.
S302, acquiring keyword information corresponding to the voice data retrieval instruction;
after receiving the voice data retrieval instruction, further analyzing the voice data retrieval instruction to obtain keyword information corresponding to the voice data retrieval instruction. The acquired keyword information may be a part of the content for search input by the user or may be all of the content for search according to different search requirements. For example, when the content for performing the voice data search input by the user is "recognition rate of voice recognition engine", the keyword information may be content such as "voice recognition engine" or "recognition rate of engine", or may be "recognition rate of voice recognition engine". In addition, when the keyword information is a part of the content used for searching, the obtained related search result is wider, and when the keyword information is the whole content of the content used for searching, the obtained related search result is more accurate. Therefore, the keyword information corresponding to the voice data retrieval instruction can be flexibly set according to different requirements.
S303, responding to the voice data retrieval instruction, triggering the voice data in the application to perform playing operation, and controlling the playing volume of the voice data in the application to be smaller than a preset volume threshold value and/or controlling the playing speed of the voice data in the application to be larger than a first preset threshold value when the voice data in the application is triggered to perform playing operation;
and after the keyword information corresponding to the voice data retrieval instruction is acquired, further responding to the received voice data retrieval instruction, and triggering the internal voice data in the application to carry out playing operation. For example, in a specific application scenario, a play operation is triggered to be performed on historical voice data in the WeChat application in response to a voice data retrieval instruction.
In the process of triggering the voice data in the application to perform the playing operation, in order to further reduce the influence on the user in the process of playing the voice data, the playing volume of the voice data in the application can be further controlled to be smaller than the preset volume threshold when the playing operation is performed. The preset volume threshold may be a volume value at which the play volume has the least influence on the user.
In the process of triggering the voice data in the application to perform the playing operation, in order to further improve the retrieval efficiency, the playing speed of the voice data in the application may be further controlled to be greater than the first preset threshold when the playing operation is performed. The first preset threshold may be a speed value of normal play of voice data.
It should be noted that, when the voice data in the application is triggered to perform the playing operation, according to different retrieval requirements, only the playing volume of the voice data in the application may be controlled to be smaller than a preset volume threshold, or only the playing speed of the voice data in the application is controlled to be larger than a first preset threshold; and simultaneously controlling the playing volume of the voice data in the application to be smaller than a preset volume threshold value and controlling the playing speed of the voice data in the application to be larger than a first preset threshold value.
S304, acquiring voice data in the process of playing the voice data in the application;
in the process of triggering the playing operation of the voice data in the application, the audio stream in the application is made to flow in the code frame of the operating system, and in the process of flowing the audio stream, the audio data is obtained, namely the voice data is obtained.
S305, matching is carried out in the information content of each piece of voice data based on the keyword information, and a retrieval result is output.
After the voice data in the application is acquired, the information content of each piece of acquired voice data is further matched according to the acquired keyword information, and finally, a matching result is output, namely a retrieval result is output.
For example, when the keyword information is "recognition rate of engine", the information content of each piece of acquired voice data is matched by "recognition rate of engine", and finally, a matching result "recognition rate of voice recognition engine" is output.
It should be noted that, when outputting the search result, the implementation manner may be to jump to the corresponding position of the voice data to be used, and further identify the searched voice data, which is convenient for the user to identify; the implementation mode can also be that the retrieved results are displayed in a list form; that is to say, when the retrieval result is output, a corresponding retrieval result presentation mode can be designed according to actual requirements.
In summary, in this embodiment, on the basis of the method embodiment 2, when the voice data in the application is triggered to be played in response to the voice data retrieval instruction, and when the voice data in the application is triggered to be played, the playing volume of the voice data in the application can be further controlled to be smaller than the preset volume threshold and/or the playing speed of the voice data in the application is controlled to be greater than the first preset threshold, so that on the basis of implementing searching for the voice content in the application of voice interaction, the influence of the retrieval process on the user is further reduced and/or the retrieval efficiency is improved.
As shown in fig. 4, which is a flowchart of a method in embodiment 4 of a data processing method disclosed in the present disclosure, the method is applied to an electronic device, where the electronic device may be a smart device such as a smart phone and a tablet computer, and the electronic device includes at least one application capable of implementing voice interaction, for example: applications such as "WeChat", "QQ", etc.; the method may comprise the steps of:
s401, receiving a voice data retrieval instruction;
when the voice content of the application capable of realizing voice interaction in the electronic equipment needs to be searched, a voice data retrieval instruction is received firstly, namely, an instruction for starting a voice data retrieval function is received. It should be noted that the voice data retrieval instruction may be generated by user triggering. For example, the user may input the related content that needs to be retrieved through a search bar of the electronic device, and generate a corresponding voice data retrieval instruction when the user inputs the related content that needs to be retrieved.
In a specific application scenario, taking an electronic device as a smart phone and an application capable of realizing voice interaction as WeChat as an example, when a user needs to search for voice content in the WeChat application in the electronic device, related content needing to be retrieved can be input through a search bar of the smart phone, for example, the user inputs a "recognition rate of a voice recognition engine" in the search bar, and when the user inputs the "recognition rate of the voice recognition engine", a corresponding voice data retrieval instruction is generated.
S402, acquiring keyword information corresponding to the voice data retrieval instruction;
after receiving the voice data retrieval instruction, further analyzing the voice data retrieval instruction to obtain keyword information corresponding to the voice data retrieval instruction. The acquired keyword information may be a part of the content for search input by the user or may be all of the content for search according to different search requirements. For example, when the content for performing the voice data search input by the user is "recognition rate of voice recognition engine", the keyword information may be content such as "voice recognition engine" or "recognition rate of engine", or may be "recognition rate of voice recognition engine". In addition, when the keyword information is a part of the content used for searching, the obtained related search result is wider, and when the keyword information is the whole content of the content used for searching, the obtained related search result is more accurate. Therefore, the keyword information corresponding to the voice data retrieval instruction can be flexibly set according to different requirements.
S403, responding to the voice data retrieval instruction, triggering the voice data in the application to perform playing operation, and controlling the voice data in the application to be played in the target device and/or controlling the playing speed of the voice data in the application to be greater than a second preset threshold when the voice data in the application is triggered to perform playing operation, wherein the volume of the voice data in the application played by the target device is zero;
and after the keyword information corresponding to the voice data retrieval instruction is acquired, further responding to the received voice data retrieval instruction, and triggering the internal voice data in the application to carry out playing operation. For example, in a specific application scenario, a play operation is triggered to be performed on historical voice data in the WeChat application in response to a voice data retrieval instruction.
In the process of triggering the playing operation of the voice data in the application, in order to further reduce the influence on the user in the playing process of the voice data, the target device can be redirected to play the voice data in the application when the playing operation is executed, and the target device does not actually make a playing sound when the voice data in the application is played, that is, for the user, the volume of the voice data in the application played by the target device is zero.
In the process of triggering the voice data in the application to perform the playing operation, in order to further improve the retrieval efficiency, the playing speed of the voice data in the application may be further controlled to be greater than the second preset threshold when the playing operation is performed. The second preset threshold may be a speed value of normal playing of the voice data, or a speed value faster than the speed of normal playing.
It should be noted that, when the voice data in the application is triggered to perform the playing operation, according to different retrieval requirements, only the voice data in the application may be controlled to be played in the target device, or only the playing speed of the voice data in the application is controlled to be greater than a second preset threshold; and simultaneously controlling the voice data in the application to be played in the target device, and controlling the playing speed of the voice data in the application to be greater than a second preset threshold value.
S404, acquiring voice data in the process of playing the voice data in the application;
in the process of triggering the playing operation of the voice data in the application, the audio stream in the application is made to flow in the code frame of the operating system, and in the process of flowing the audio stream, the audio data is obtained, namely the voice data is obtained.
S405, matching is carried out in the information content of each piece of voice data based on the keyword information, and a retrieval result is output.
After the voice data in the application is acquired, the information content of each piece of acquired voice data is further matched according to the acquired keyword information, and finally, a matching result is output, namely a retrieval result is output.
For example, when the keyword information is "recognition rate of engine", the information content of each piece of acquired voice data is matched by "recognition rate of engine", and finally, a matching result "recognition rate of voice recognition engine" is output.
It should be noted that, when outputting the search result, the implementation manner may be to jump to the corresponding position of the voice data to be used, and further identify the searched voice data, which is convenient for the user to identify; the implementation mode can also be that the retrieved results are displayed in a list form; that is to say, when the retrieval result is output, a corresponding retrieval result presentation mode can be designed according to actual requirements.
In summary, in this embodiment, on the basis of the method embodiment 2, when the voice data in the application is triggered to be played in response to the voice data retrieval instruction, and when the voice data in the application is triggered to be played, the voice data in the application can be further controlled to be played in the target device and/or the playing speed of the voice data in the application is controlled to be greater than the second preset threshold, so that on the basis of implementing searching for the voice content in the application of voice interaction, the influence of the retrieval process on the user is reduced and/or the retrieval efficiency is improved.
As shown in fig. 5, which is a schematic structural diagram of an embodiment 1 of an electronic device disclosed in the present disclosure, the electronic device may be an intelligent device such as a smart phone and a tablet computer, and the electronic device at least includes an application capable of implementing voice interaction, for example: applications such as "WeChat", "QQ", etc.; the electronic device may include:
a receiving module 501, configured to receive a voice data retrieval instruction;
when the voice content of the application capable of realizing voice interaction in the electronic equipment needs to be searched, a voice data retrieval instruction is received firstly, namely, an instruction for starting a voice data retrieval function is received. It should be noted that the voice data retrieval instruction may be generated by user triggering. For example, the user may input the related content that needs to be retrieved through a search bar of the electronic device, and generate a corresponding voice data retrieval instruction when the user inputs the related content that needs to be retrieved.
In a specific application scenario, taking an electronic device as a smart phone and an application capable of realizing voice interaction as WeChat as an example, when a user needs to search for voice content in the WeChat application in the electronic device, related content needing to be retrieved can be input through a search bar of the smart phone, for example, the user inputs a "recognition rate of a voice recognition engine" in the search bar, and when the user inputs the "recognition rate of the voice recognition engine", a corresponding voice data retrieval instruction is generated.
An obtaining module 502, configured to obtain keyword information corresponding to a voice data retrieval instruction;
after receiving the voice data retrieval instruction, further analyzing the voice data retrieval instruction to obtain keyword information corresponding to the voice data retrieval instruction. The acquired keyword information may be a part of the content for search input by the user or may be all of the content for search according to different search requirements. For example, when the content for performing the voice data search input by the user is "recognition rate of voice recognition engine", the keyword information may be content such as "voice recognition engine" or "recognition rate of engine", or may be "recognition rate of voice recognition engine". In addition, when the keyword information is a part of the content used for searching, the obtained related search result is wider, and when the keyword information is the whole content of the content used for searching, the obtained related search result is more accurate. Therefore, the keyword information corresponding to the voice data retrieval instruction can be flexibly set according to different requirements.
And the retrieval module 503 is configured to respond to the voice data retrieval instruction, retrieve the voice data in the application based on the keyword information, and output a retrieval result.
And after the keyword information corresponding to the voice data retrieval instruction is acquired, further responding to the received voice data retrieval instruction, retrieving the voice data in the application according to the acquired keyword information, and finally outputting the retrieved result. For example, when the keyword information is "recognition rate of engine", the speech data in the WeChat application is searched for by "recognition rate of engine", and finally the search result "recognition rate of speech recognition engine" is output.
It should be noted that, when outputting the search result, the implementation manner may be to jump to the corresponding position of the voice data to be used, and further identify the searched voice data, which is convenient for the user to identify; the implementation mode can also be that the retrieved results are displayed in a list form; that is to say, when the retrieval result is output, a corresponding retrieval result presentation mode can be designed according to actual requirements.
In summary, in the above embodiments, when the voice data in the application needs to be retrieved, the voice data retrieval instruction is received first, then the keyword information corresponding to the voice data retrieval instruction is obtained, the voice data in the application is retrieved based on the keyword information in response to the voice data retrieval instruction, and the retrieval result is output. The method and the device can effectively search the voice content in the application capable of realizing voice interaction, and improve user experience.
As shown in fig. 6, which is a schematic structural diagram of an embodiment 2 of an electronic device disclosed in the present disclosure, the electronic device may be an intelligent device such as a smart phone and a tablet computer, and the electronic device at least includes an application capable of implementing voice interaction, for example: applications such as "WeChat", "QQ", etc.; the electronic device may include:
a receiving module 601, configured to receive a voice data retrieval instruction;
when the voice content of the application capable of realizing voice interaction in the electronic equipment needs to be searched, a voice data retrieval instruction is received firstly, namely, an instruction for starting a voice data retrieval function is received. It should be noted that the voice data retrieval instruction may be generated by user triggering. For example, the user may input the related content that needs to be retrieved through a search bar of the electronic device, and generate a corresponding voice data retrieval instruction when the user inputs the related content that needs to be retrieved.
In a specific application scenario, taking an electronic device as a smart phone and an application capable of realizing voice interaction as WeChat as an example, when a user needs to search for voice content in the WeChat application in the electronic device, related content needing to be retrieved can be input through a search bar of the smart phone, for example, the user inputs a "recognition rate of a voice recognition engine" in the search bar, and when the user inputs the "recognition rate of the voice recognition engine", a corresponding voice data retrieval instruction is generated.
An obtaining module 602, configured to obtain keyword information corresponding to a voice data retrieval instruction;
after receiving the voice data retrieval instruction, further analyzing the voice data retrieval instruction to obtain keyword information corresponding to the voice data retrieval instruction. The acquired keyword information may be a part of the content for search input by the user or may be all of the content for search according to different search requirements. For example, when the content for performing the voice data search input by the user is "recognition rate of voice recognition engine", the keyword information may be content such as "voice recognition engine" or "recognition rate of engine", or may be "recognition rate of voice recognition engine". In addition, when the keyword information is a part of the content used for searching, the obtained related search result is wider, and when the keyword information is the whole content of the content used for searching, the obtained related search result is more accurate. Therefore, the keyword information corresponding to the voice data retrieval instruction can be flexibly set according to different requirements.
The retrieval module 603 is configured to respond to the voice data retrieval instruction, and trigger the voice data in the application to perform a playing operation;
and after the keyword information corresponding to the voice data retrieval instruction is acquired, further responding to the received voice data retrieval instruction, and triggering the internal voice data in the application to carry out playing operation. For example, in a specific application scenario, a play operation is triggered to be performed on historical voice data in the WeChat application in response to a voice data retrieval instruction.
The retrieving module 603 is further configured to obtain voice data during a process of playing the voice data in the application;
in the process of triggering the playing operation of the voice data in the application, the audio stream in the application is made to flow in the code frame of the operating system, and in the process of flowing the audio stream, the audio data is obtained, namely the voice data is obtained.
The retrieving module 603 is further configured to match information content of each piece of voice data based on the keyword information, and output a retrieval result.
After the voice data in the application is acquired, the information content of each piece of acquired voice data is further matched according to the acquired keyword information, and finally, a matching result is output, namely a retrieval result is output.
For example, when the keyword information is "recognition rate of engine", the information content of each piece of acquired voice data is matched by "recognition rate of engine", and finally, a matching result "recognition rate of voice recognition engine" is output.
It should be noted that, when outputting the search result, the implementation manner may be to jump to the corresponding position of the voice data to be used, and further identify the searched voice data, which is convenient for the user to identify; the implementation mode can also be that the retrieved results are displayed in a list form; that is to say, when the retrieval result is output, a corresponding retrieval result presentation mode can be designed according to actual requirements.
In summary, in the above embodiments, when the voice data in the application needs to be retrieved, the voice data in the application can be triggered to perform a playing operation by responding to the voice data retrieval instruction, the voice data is obtained during the playing operation, the information content of each piece of voice data is matched according to the keyword information in the retrieval instruction, and a retrieval result is output, so that the voice content in the voice interactive application is searched, and the user experience is improved.
Specifically, in the above embodiment, the received voice data retrieval instruction may be generated by user triggering. For example, the user may input the related content to be retrieved through a search bar of the electronic device, or the user may input the related content to be retrieved through voice through a voice interaction function of the electronic device, and generate a corresponding voice data retrieval instruction when the user inputs the related content to be retrieved. That is, the voice data retrieval instruction may be generated based on text information input by the user or based on voice information input by the user.
Therefore, when the voice data retrieval instruction is generated based on text information input by a user, when keyword information corresponding to the voice data retrieval instruction is acquired, text keyword information corresponding to the voice data retrieval instruction is acquired; when the voice data retrieval instruction is generated based on voice information input by a user, the voice keyword information corresponding to the voice data retrieval instruction is acquired when the keyword information corresponding to the voice data retrieval instruction is acquired.
Specifically, in the above embodiment, when the keyword information corresponding to the acquired voice data retrieval instruction is the text keyword information corresponding to the acquired voice data retrieval instruction, correspondingly, matching is performed in the information content of each piece of voice data based on the keyword information, and when the retrieval result is output, each piece of voice data acquired during the process of playing the voice data in the application may be first converted into corresponding text information, and then the text keyword information may be matched with the text information obtained by converting each piece of voice data, and the retrieval result is output.
In addition, when the keyword information corresponding to the acquired voice data retrieval instruction is the voice keyword information corresponding to the acquired voice data retrieval instruction, and accordingly, when the information content of each piece of voice data is matched based on the keyword information and the retrieval result is output, the voice keyword information may be first converted into corresponding keyword text information, and meanwhile, each piece of voice data acquired during the playing operation of the voice data in the application is converted into corresponding text information, and then the keyword text information corresponding to the voice keyword information is matched with the text information obtained by converting each piece of voice data, and the retrieval result is output.
As shown in fig. 7, which is a schematic structural diagram of an electronic device in embodiment 3 of the present disclosure, the electronic device may be an intelligent device such as a smart phone or a tablet computer, and the electronic device at least includes an application capable of implementing voice interaction, for example: applications such as "WeChat", "QQ", etc.; the electronic device may include:
a receiving module 701, configured to receive a voice data retrieval instruction;
when the voice content of the application capable of realizing voice interaction in the electronic equipment needs to be searched, a voice data retrieval instruction is received firstly, namely, an instruction for starting a voice data retrieval function is received. It should be noted that the voice data retrieval instruction may be generated by user triggering. For example, the user may input the related content that needs to be retrieved through a search bar of the electronic device, and generate a corresponding voice data retrieval instruction when the user inputs the related content that needs to be retrieved.
In a specific application scenario, taking an electronic device as a smart phone and an application capable of realizing voice interaction as WeChat as an example, when a user needs to search for voice content in the WeChat application in the electronic device, related content needing to be retrieved can be input through a search bar of the smart phone, for example, the user inputs a "recognition rate of a voice recognition engine" in the search bar, and when the user inputs the "recognition rate of the voice recognition engine", a corresponding voice data retrieval instruction is generated.
An obtaining module 702, configured to obtain keyword information corresponding to a voice data retrieval instruction;
after receiving the voice data retrieval instruction, further analyzing the voice data retrieval instruction to obtain keyword information corresponding to the voice data retrieval instruction. The acquired keyword information may be a part of the content for search input by the user or may be all of the content for search according to different search requirements. For example, when the content for performing the voice data search input by the user is "recognition rate of voice recognition engine", the keyword information may be content such as "voice recognition engine" or "recognition rate of engine", or may be "recognition rate of voice recognition engine". In addition, when the keyword information is a part of the content used for searching, the obtained related search result is wider, and when the keyword information is the whole content of the content used for searching, the obtained related search result is more accurate. Therefore, the keyword information corresponding to the voice data retrieval instruction can be flexibly set according to different requirements.
The retrieval module 703 is configured to respond to the voice data retrieval instruction, trigger the voice data in the application to perform a playing operation, and control a playing volume of the voice data in the application to be smaller than a preset volume threshold and/or control a playing speed of the voice data in the application to be larger than a first preset threshold when the voice data in the application is triggered to perform the playing operation;
and after the keyword information corresponding to the voice data retrieval instruction is acquired, further responding to the received voice data retrieval instruction, and triggering the internal voice data in the application to carry out playing operation. For example, in a specific application scenario, a play operation is triggered to be performed on historical voice data in the WeChat application in response to a voice data retrieval instruction.
In the process of triggering the voice data in the application to perform the playing operation, in order to further reduce the influence on the user in the process of playing the voice data, the playing volume of the voice data in the application can be further controlled to be smaller than the preset volume threshold when the playing operation is performed. The preset volume threshold may be a volume value at which the play volume has the least influence on the user.
In the process of triggering the voice data in the application to perform the playing operation, in order to further improve the retrieval efficiency, the playing speed of the voice data in the application may be further controlled to be greater than the first preset threshold when the playing operation is performed. The first preset threshold may be a speed value of normal play of voice data.
It should be noted that, when the voice data in the application is triggered to perform the playing operation, according to different retrieval requirements, only the playing volume of the voice data in the application may be controlled to be smaller than a preset volume threshold, or only the playing speed of the voice data in the application is controlled to be larger than a first preset threshold; and simultaneously controlling the playing volume of the voice data in the application to be smaller than a preset volume threshold value and controlling the playing speed of the voice data in the application to be larger than a first preset threshold value.
The retrieving module 703 is further configured to obtain voice data during a process of playing the voice data in the application;
in the process of triggering the playing operation of the voice data in the application, the audio stream in the application is made to flow in the code frame of the operating system, and in the process of flowing the audio stream, the audio data is obtained, namely the voice data is obtained.
The retrieving module 703 is further configured to match the information content of each piece of voice data based on the keyword information, and output a retrieval result.
After the voice data in the application is acquired, the information content of each piece of acquired voice data is further matched according to the acquired keyword information, and finally, a matching result is output, namely a retrieval result is output.
For example, when the keyword information is "recognition rate of engine", the information content of each piece of acquired voice data is matched by "recognition rate of engine", and finally, a matching result "recognition rate of voice recognition engine" is output.
It should be noted that, when outputting the search result, the implementation manner may be to jump to the corresponding position of the voice data to be used, and further identify the searched voice data, which is convenient for the user to identify; the implementation mode can also be that the retrieved results are displayed in a list form; that is to say, when the retrieval result is output, a corresponding retrieval result presentation mode can be designed according to actual requirements.
In summary, in this embodiment, on the basis of embodiment 2 of the electronic device, when the voice data search instruction is responded, the voice data in the application is triggered to perform the play operation, and when the voice data in the application is triggered to perform the play operation, the play volume of the voice data in the application can be further controlled to be smaller than the preset volume threshold and/or the play speed of the voice data in the application is controlled to be greater than the first preset threshold, so that on the basis of implementing searching for the voice content in the application of voice interaction, the influence of the search process on the user is further reduced and/or the search efficiency is improved.
As shown in fig. 8, which is a schematic structural diagram of an embodiment 4 of an electronic device disclosed in the present disclosure, the electronic device may be an intelligent device such as a smart phone and a tablet computer, and the electronic device at least includes an application capable of implementing voice interaction, for example: applications such as "WeChat", "QQ", etc.; the electronic device may include:
a receiving module 801, configured to receive a voice data retrieval instruction;
when the voice content of the application capable of realizing voice interaction in the electronic equipment needs to be searched, a voice data retrieval instruction is received firstly, namely, an instruction for starting a voice data retrieval function is received. It should be noted that the voice data retrieval instruction may be generated by user triggering. For example, the user may input the related content that needs to be retrieved through a search bar of the electronic device, and generate a corresponding voice data retrieval instruction when the user inputs the related content that needs to be retrieved.
In a specific application scenario, taking an electronic device as a smart phone and an application capable of realizing voice interaction as WeChat as an example, when a user needs to search for voice content in the WeChat application in the electronic device, related content needing to be retrieved can be input through a search bar of the smart phone, for example, the user inputs a "recognition rate of a voice recognition engine" in the search bar, and when the user inputs the "recognition rate of the voice recognition engine", a corresponding voice data retrieval instruction is generated.
An obtaining module 802, configured to obtain keyword information corresponding to a voice data retrieval instruction;
after receiving the voice data retrieval instruction, further analyzing the voice data retrieval instruction to obtain keyword information corresponding to the voice data retrieval instruction. The acquired keyword information may be a part of the content for search input by the user or may be all of the content for search according to different search requirements. For example, when the content for performing the voice data search input by the user is "recognition rate of voice recognition engine", the keyword information may be content such as "voice recognition engine" or "recognition rate of engine", or may be "recognition rate of voice recognition engine". In addition, when the keyword information is a part of the content used for searching, the obtained related search result is wider, and when the keyword information is the whole content of the content used for searching, the obtained related search result is more accurate. Therefore, the keyword information corresponding to the voice data retrieval instruction can be flexibly set according to different requirements.
The retrieval module 803 is configured to respond to the voice data retrieval instruction, trigger the voice data in the application to perform a playing operation, and control the voice data in the application to be played in the target device and/or control the playing speed of the voice data in the application to be greater than a second preset threshold when the voice data in the application is triggered to perform the playing operation, where a volume of the voice data in the application played by the target device is zero;
and after the keyword information corresponding to the voice data retrieval instruction is acquired, further responding to the received voice data retrieval instruction, and triggering the internal voice data in the application to carry out playing operation. For example, in a specific application scenario, a play operation is triggered to be performed on historical voice data in the WeChat application in response to a voice data retrieval instruction.
In the process of triggering the playing operation of the voice data in the application, in order to further reduce the influence on the user in the playing process of the voice data, the target device can be redirected to play the voice data in the application when the playing operation is executed, and the target device does not actually make a playing sound when the voice data in the application is played, that is, for the user, the volume of the voice data in the application played by the target device is zero.
In the process of triggering the voice data in the application to perform the playing operation, in order to further improve the retrieval efficiency, the playing speed of the voice data in the application may be further controlled to be greater than the second preset threshold when the playing operation is performed. The second preset threshold may be a speed value of normal playing of the voice data, or a speed value faster than the speed of normal playing.
It should be noted that, when the voice data in the application is triggered to perform the playing operation, according to different retrieval requirements, only the voice data in the application may be controlled to be played in the target device, or only the playing speed of the voice data in the application is controlled to be greater than a second preset threshold; and simultaneously controlling the voice data in the application to be played in the target device, and controlling the playing speed of the voice data in the application to be greater than a second preset threshold value.
The retrieving module 803 is further configured to obtain the voice data during a process of playing the voice data in the application;
in the process of triggering the playing operation of the voice data in the application, the audio stream in the application is made to flow in the code frame of the operating system, and in the process of flowing the audio stream, the audio data is obtained, namely the voice data is obtained.
The retrieving module 803 is further configured to match the information content of each piece of voice data based on the keyword information, and output a retrieval result.
After the voice data in the application is acquired, the information content of each piece of acquired voice data is further matched according to the acquired keyword information, and finally, a matching result is output, namely a retrieval result is output.
For example, when the keyword information is "recognition rate of engine", the information content of each piece of acquired voice data is matched by "recognition rate of engine", and finally, a matching result "recognition rate of voice recognition engine" is output.
It should be noted that, when outputting the search result, the implementation manner may be to jump to the corresponding position of the voice data to be used, and further identify the searched voice data, which is convenient for the user to identify; the implementation mode can also be that the retrieved results are displayed in a list form; that is to say, when the retrieval result is output, a corresponding retrieval result presentation mode can be designed according to actual requirements.
In summary, in this embodiment, on the basis of the embodiment 2 of the electronic device, when the voice data in the application is triggered to be played in response to the voice data retrieval instruction, the voice data in the application can be further controlled to be played in the target device and/or the playing speed of the voice data in the application is controlled to be greater than the second preset threshold, so that on the basis of implementing searching for the voice content in the application of voice interaction, the influence of the retrieval process on the user is reduced and/or the retrieval efficiency is improved.
As shown in fig. 9, which is a schematic structural diagram of an embodiment 5 of an electronic device disclosed in the present disclosure, the electronic device may be an intelligent device such as a smart phone or a tablet computer, and the electronic device at least includes an application capable of implementing voice interaction, for example: applications such as "WeChat", "QQ", etc.; the electronic device may include: a memory 901, a processor 902 and a computer program stored on the memory 901 and executable on the processor 902, the program when executed by the processor 902 for effecting:
receiving a voice data retrieval instruction;
acquiring keyword information corresponding to a voice data retrieval instruction;
and responding to the voice data retrieval instruction, retrieving the voice data in the application based on the keyword information, and outputting a retrieval result.
When the voice content of the application capable of realizing voice interaction in the electronic equipment needs to be searched, a voice data retrieval instruction is received firstly, namely, an instruction for starting a voice data retrieval function is received. It should be noted that the voice data retrieval instruction may be generated by user triggering. For example, the user may input the related content that needs to be retrieved through a search bar of the electronic device, and generate a corresponding voice data retrieval instruction when the user inputs the related content that needs to be retrieved.
In a specific application scenario, taking an electronic device as a smart phone and an application capable of realizing voice interaction as WeChat as an example, when a user needs to search for voice content in the WeChat application in the electronic device, related content needing to be retrieved can be input through a search bar of the smart phone, for example, the user inputs a "recognition rate of a voice recognition engine" in the search bar, and when the user inputs the "recognition rate of the voice recognition engine", a corresponding voice data retrieval instruction is generated.
After receiving the voice data retrieval instruction, further analyzing the voice data retrieval instruction to obtain keyword information corresponding to the voice data retrieval instruction. The acquired keyword information may be a part of the content for search input by the user or may be all of the content for search according to different search requirements. For example, when the content for performing the voice data search input by the user is "recognition rate of voice recognition engine", the keyword information may be content such as "voice recognition engine" or "recognition rate of engine", or may be "recognition rate of voice recognition engine". In addition, when the keyword information is a part of the content used for searching, the obtained related search result is wider, and when the keyword information is the whole content of the content used for searching, the obtained related search result is more accurate. Therefore, the keyword information corresponding to the voice data retrieval instruction can be flexibly set according to different requirements.
And after the keyword information corresponding to the voice data retrieval instruction is acquired, further responding to the received voice data retrieval instruction, retrieving the voice data in the application according to the acquired keyword information, and finally outputting the retrieved result. For example, when the keyword information is "recognition rate of engine", the speech data in the WeChat application is searched for by "recognition rate of engine", and finally the search result "recognition rate of speech recognition engine" is output.
It should be noted that, when outputting the search result, the implementation manner may be to jump to the corresponding position of the voice data to be used, and further identify the searched voice data, which is convenient for the user to identify; the implementation mode can also be that the retrieved results are displayed in a list form; that is to say, when the retrieval result is output, a corresponding retrieval result presentation mode can be designed according to actual requirements.
In summary, in the above embodiments, when the voice data in the application needs to be retrieved, the voice data retrieval instruction is received first, then the keyword information corresponding to the voice data retrieval instruction is obtained, the voice data in the application is retrieved based on the keyword information in response to the voice data retrieval instruction, and the retrieval result is output. The method and the device can effectively search the voice content in the application capable of realizing voice interaction, and improve user experience.
It should be noted that, in the present specification, the embodiments are all described in a progressive manner, each embodiment focuses on differences from other embodiments, and the same and similar parts among the embodiments may be referred to each other. For the device or system type embodiment, since it is basically similar to the method embodiment, the description is simple, and for the relevant points, refer to the partial description of the method embodiment.
It is further noted that, herein, relational terms such as first and second, and the like may be used solely to distinguish one entity or action from another entity or action without necessarily requiring or implying any actual such relationship or order between such entities or actions. Also, the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Without further limitation, an element defined by the phrase "comprising an … …" does not exclude the presence of other identical elements in a process, method, article, or apparatus that comprises the element.
The steps of a method or algorithm described in connection with the embodiments disclosed herein may be embodied directly in hardware, in a software module executed by a processor, or in a combination of the two. A software module may reside in Random Access Memory (RAM), memory, Read Only Memory (ROM), electrically programmable ROM, electrically erasable programmable ROM, registers, hard disk, a removable disk, a CD-ROM, or any other form of storage medium known in the art.
The previous description of the disclosed embodiments is provided to enable any person skilled in the art to make or use the present disclosure. Various modifications to these embodiments will be readily apparent to those skilled in the art, and the generic principles defined herein may be applied to other embodiments without departing from the spirit or scope of the disclosure. Thus, the present disclosure is not intended to be limited to the embodiments shown herein but is to be accorded the widest scope consistent with the principles and novel features disclosed herein.

Claims (9)

1. A data processing method is applied to electronic equipment, wherein the electronic equipment at least comprises an application capable of realizing voice interaction, and the method comprises the following steps:
receiving a voice data retrieval instruction;
acquiring keyword information corresponding to the voice data retrieval instruction;
responding to the voice data retrieval instruction, retrieving the voice data in the application based on the keyword information, and outputting a retrieval result;
wherein, the responding the voice data retrieval instruction, retrieving the voice data in the application based on the keyword information, and outputting a retrieval result comprises:
responding to the voice data retrieval instruction, and triggering the voice data in the application to carry out playing operation;
acquiring voice data in the process of playing the voice data in the application;
and matching the information content of each piece of voice data based on the keyword information, and outputting a retrieval result.
2. The method of claim 1, wherein the triggering, in response to the voice data retrieval instruction, the voice data in the application to perform a play operation comprises:
and responding to the voice data retrieval instruction, and controlling the playing volume of the voice data in the application to be smaller than a preset volume threshold when the voice data in the application is triggered to be played.
3. The method of claim 2, wherein the triggering of the voice data in the application to play back in response to the voice data retrieval instruction further comprises:
responding to the voice data retrieval instruction, and controlling the playing speed of the voice data in the application to be larger than a first preset threshold when the voice data in the application is triggered to be played.
4. The method of claim 1, wherein the triggering, in response to the voice data retrieval instruction, the voice data in the application to perform a play operation comprises:
and responding to the voice data retrieval instruction, and controlling the voice data in the application to be played in the target equipment when the voice data in the application is triggered to be played, wherein the volume of the voice data in the application played by the target equipment is zero.
5. The method of claim 4, wherein the triggering of the voice data in the application to play back in response to the voice data retrieval instruction further comprises:
and controlling the playing speed of the voice data in the application to be played in the target equipment to be greater than a second preset threshold value.
6. The method of claim 1, wherein the obtaining of the keyword information corresponding to the voice data retrieval instruction comprises:
and acquiring voice keyword information or text keyword information corresponding to the voice data retrieval instruction.
7. The method of claim 6, wherein the matching in the information content of each piece of the voice data based on the keyword information and outputting a retrieval result comprises:
converting each piece of acquired voice data into corresponding text information;
and matching the text keyword information with the text information obtained by converting each piece of voice data, and outputting a retrieval result.
8. An electronic device including at least one application capable of voice interaction, the electronic device comprising:
the receiving module is used for receiving a voice data retrieval instruction;
the acquisition module is used for acquiring keyword information corresponding to the voice data retrieval instruction;
the retrieval module is used for responding to the voice data retrieval instruction, retrieving the voice data in the application based on the keyword information and outputting a retrieval result;
wherein the retrieval module is configured to:
responding to the voice data retrieval instruction, and triggering the voice data in the application to carry out playing operation;
acquiring voice data in the process of playing the voice data in the application;
and matching the information content of each piece of voice data based on the keyword information, and outputting a retrieval result.
9. An electronic device including at least one application capable of voice interaction, the electronic device comprising: memory, a processor and a computer program stored on the memory and executable on the processor, the processor when executing the program for implementing:
receiving a voice data retrieval instruction;
acquiring keyword information corresponding to the voice data retrieval instruction;
responding to the voice data retrieval instruction, retrieving the voice data in the application based on the keyword information, and outputting a retrieval result;
wherein, the responding the voice data retrieval instruction, retrieving the voice data in the application based on the keyword information, and outputting a retrieval result comprises:
responding to the voice data retrieval instruction, and triggering the voice data in the application to carry out playing operation;
acquiring voice data in the process of playing the voice data in the application;
and matching the information content of each piece of voice data based on the keyword information, and outputting a retrieval result.
CN201910581137.3A 2019-06-29 2019-06-29 Data processing method and electronic equipment Active CN110287365B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910581137.3A CN110287365B (en) 2019-06-29 2019-06-29 Data processing method and electronic equipment

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910581137.3A CN110287365B (en) 2019-06-29 2019-06-29 Data processing method and electronic equipment

Publications (2)

Publication Number Publication Date
CN110287365A CN110287365A (en) 2019-09-27
CN110287365B true CN110287365B (en) 2021-07-16

Family

ID=68020174

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910581137.3A Active CN110287365B (en) 2019-06-29 2019-06-29 Data processing method and electronic equipment

Country Status (1)

Country Link
CN (1) CN110287365B (en)

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101329867A (en) * 2007-06-21 2008-12-24 西门子(中国)有限公司 Method and device for playing speech on demand
CN103425668A (en) * 2012-05-16 2013-12-04 联想(北京)有限公司 Information search method and electronic equipment
CN104239328A (en) * 2013-06-18 2014-12-24 联想(北京)有限公司 Multimedia processing method and multimedia system
CN105869623A (en) * 2015-12-07 2016-08-17 乐视网信息技术(北京)股份有限公司 Video playing method and device based on speech recognition
WO2017000808A1 (en) * 2015-06-30 2017-01-05 芋头科技(杭州)有限公司 Method for adjusting music playback
CN109543102A (en) * 2018-11-12 2019-03-29 百度在线网络技术(北京)有限公司 Information recommendation method, device and storage medium based on video playing

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102833617B (en) * 2012-09-20 2016-03-23 北京酷云互动科技有限公司 A kind of method of simultaneous display multimedia programming satellite information, system and terminal
US8862593B1 (en) * 2013-03-15 2014-10-14 Sowt International Ltd. System and method for creating, managing, and publishing audio microposts
CN109640164A (en) * 2018-11-02 2019-04-16 重庆爱奇艺智能科技有限公司 A kind of playback method and device between multiple virtual reality devices
CN109714111A (en) * 2018-12-05 2019-05-03 肇庆市诚一网络技术有限公司 The searching method of keyword Internet-based

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101329867A (en) * 2007-06-21 2008-12-24 西门子(中国)有限公司 Method and device for playing speech on demand
CN103425668A (en) * 2012-05-16 2013-12-04 联想(北京)有限公司 Information search method and electronic equipment
CN104239328A (en) * 2013-06-18 2014-12-24 联想(北京)有限公司 Multimedia processing method and multimedia system
WO2017000808A1 (en) * 2015-06-30 2017-01-05 芋头科技(杭州)有限公司 Method for adjusting music playback
CN105869623A (en) * 2015-12-07 2016-08-17 乐视网信息技术(北京)股份有限公司 Video playing method and device based on speech recognition
CN109543102A (en) * 2018-11-12 2019-03-29 百度在线网络技术(北京)有限公司 Information recommendation method, device and storage medium based on video playing

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
人工智能语音技术在广电媒体的应用;刘晓曦;《广播电视信息》;20180315;30-35 *

Also Published As

Publication number Publication date
CN110287365A (en) 2019-09-27

Similar Documents

Publication Publication Date Title
US10692504B2 (en) User profiling for voice input processing
CN106658129B (en) Terminal control method and device based on emotion and terminal
CN107591155B (en) Voice recognition method and device, terminal and computer readable storage medium
CN106098063B (en) Voice control method, terminal device and server
TW201635134A (en) Method and apparatus for voice control
CN109979450B (en) Information processing method and device and electronic equipment
AU2017216520A1 (en) Common data repository for improving transactional efficiencies of user interactions with a computing device
CN105893351B (en) Audio recognition method and device
CN112463106A (en) Voice interaction method, device and equipment based on intelligent screen and storage medium
US8868419B2 (en) Generalizing text content summary from speech content
JP2014513828A (en) Automatic conversation support
CN109656444B (en) List positioning method, device, equipment and storage medium
CN111949240A (en) Interaction method, storage medium, service program, and device
JP2023506087A (en) Voice Wakeup Method and Apparatus for Skills
CN111324330B (en) Electronic book playing processing method, computing device and computer storage medium
CN111142993A (en) Information acquisition method, terminal and computer storage medium
CN108093311B (en) Multimedia file processing method and device, storage medium and electronic equipment
CN112652302A (en) Voice control method, device, terminal and storage medium
CN109325180B (en) Article abstract pushing method and device, terminal equipment, server and storage medium
CN108492826B (en) Audio processing method and device, intelligent equipment and medium
CN103500234A (en) Method for downloading multi-media files and electronic equipment
CN110287365B (en) Data processing method and electronic equipment
CN111814492A (en) Translation method, terminal and computer storage medium
CN110874176B (en) Interaction method, storage medium, operating system and device
CN113672154B (en) Page interaction method, medium, device and computing equipment

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant