WO2018036392A1 - Procédé, dispositif et terminal mobile de partage d'informations par voie vocale - Google Patents

Procédé, dispositif et terminal mobile de partage d'informations par voie vocale Download PDF

Info

Publication number
WO2018036392A1
WO2018036392A1 PCT/CN2017/097012 CN2017097012W WO2018036392A1 WO 2018036392 A1 WO2018036392 A1 WO 2018036392A1 CN 2017097012 W CN2017097012 W CN 2017097012W WO 2018036392 A1 WO2018036392 A1 WO 2018036392A1
Authority
WO
WIPO (PCT)
Prior art keywords
information
target
user
target object
identifier
Prior art date
Application number
PCT/CN2017/097012
Other languages
English (en)
Chinese (zh)
Inventor
王亮
Original Assignee
阿里巴巴集团控股有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 阿里巴巴集团控股有限公司 filed Critical 阿里巴巴集团控股有限公司
Publication of WO2018036392A1 publication Critical patent/WO2018036392A1/fr

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue

Definitions

  • the present invention relates to voice control technologies, and in particular, to a method and apparatus for sharing information based on voice and a mobile terminal.
  • the manner in which the terminal shares information is as follows: clicking the share button on the target page, selecting one of the displayed sharing channels, such as address book, WeChat, Weibo, and then determining the specific sharing from the sharing channel.
  • Object finally click to confirm, share the information of the target page to the object, such sharing process requires at least 3 operations, the sharing process is cumbersome, causing great inconvenience to the user.
  • the invention provides a method, a device and a mobile terminal based on voice sharing information, so as to solve the problem that the sharing process in the prior art is cumbersome.
  • the present invention provides a method for sharing information based on voice, including:
  • voice data includes at least: information about a target object to be shared;
  • Information related to the target page is transmitted to the target object by the target application.
  • the present invention provides a voice sharing based device, including:
  • a display module for displaying a target page
  • a receiving module configured to receive a sharing instruction input by a user
  • An acquiring module configured to acquire voice data according to the sharing instruction, where the voice data includes at least: Information about the target audience of the share;
  • a determining module configured to determine a target object to be shared based on at least the voice data, and determine a target application to be shared;
  • a sending module configured to send, by the target application, information related to the target page to the target object.
  • the present invention provides an apparatus for sharing information based on voice, including: an input device, a processor, and a display screen;
  • the processor is configured to control the display screen to display a target page
  • the input device is configured to receive a sharing instruction input by the user, and acquire voice data according to the sharing instruction, where the voice data includes at least: information related to the target object to be shared;
  • the processor is further configured to determine a target object to be shared based on the voice data, and determine a target application to be shared, and send information related to the target page to the target object by using the target application.
  • the invention provides a mobile terminal comprising the apparatus of any of the preceding claims.
  • the user wants to share the information of the target page, by transmitting the voice including the information of the target object in the target application, it is possible to transmit the information related to the target page to the target object through the target application, so as much as possible There is no need to manually share the information of the target page, thereby avoiding the complicated sharing process and bringing great convenience to the user.
  • FIG. 1A is a schematic flowchart of a method for sharing information based on voice according to an embodiment of the present invention
  • FIG. 1B is a schematic flowchart of a method for sharing information based on voice according to another embodiment of the present invention.
  • FIG. 2A is a schematic flowchart of a method for sharing information based on voice according to another embodiment of the present invention.
  • 2B to 2G are schematic diagrams showing display pages of respective steps of a method for sharing information based on voice according to another embodiment of the present invention.
  • 3A is a schematic flowchart of a method for sharing information based on voice according to another embodiment of the present invention.
  • FIG. 3B and FIG. 3C are schematic diagrams showing display pages of respective steps of a voice sharing information-based method according to still another embodiment of the present invention.
  • FIG. 4A is a schematic structural diagram of an apparatus for sharing information based on voice according to an embodiment of the present invention.
  • FIG. 4B is a schematic structural diagram of an apparatus for sharing information based on voice according to another embodiment of the present invention.
  • FIG. 5 is a schematic structural diagram of an apparatus for sharing information based on voice according to another embodiment of the present invention.
  • FIG. 5B is a schematic structural diagram of an apparatus for sharing information based on voice according to another embodiment of the present invention.
  • FIG. 6 is a schematic structural diagram of an apparatus for sharing information based on voice according to another embodiment of the present invention.
  • This embodiment provides a method for sharing information based on voice to implement information sharing by using voice.
  • the execution subject of this embodiment is a device based on voice sharing information.
  • FIG. 1A it is a schematic flowchart of a method for sharing information based on voice according to the present embodiment.
  • the method includes:
  • step 111 the target page is displayed, and a sharing instruction input by the user is received.
  • the target page is the web page currently being displayed on the terminal display.
  • the sharing instruction of this embodiment is used to share information, such as information related to a target page.
  • the receiving user sends a sharing instruction by pressing or clicking a preset button on the target page, or inputting a sharing instruction by performing a preset operation on the target page, for example, long pressing the target.
  • the page, or the terminal that displays the target page performs a preset operation to input a sharing command, such as shaking the terminal, or inputting a sharing command by sending a voice command, which may be selected according to actual needs, and details are not described herein again.
  • Step 112 Acquire voice data according to the sharing instruction, where the voice data includes at least: information related to the target object to be shared.
  • the information related to the target object to be shared is some information of the target object.
  • the voice data is the voice emitted by the user, and the target object included in the voice data is the user to whom the target page is to be sent.
  • the information of the target object is, for example, information such as an account, a user name, an avatar, a custom nickname, and the like of the target object in the target application.
  • the voice data input by the user during the continuous pressing of the button may be acquired, and if the sharing instruction input by other methods is used, the user may recognize After inputting the instruction, the voice input by the user in the preset time is acquired to obtain the voice data.
  • Step 113 Determine a target object to be shared based on at least the voice data, and determine a target application to be shared.
  • the required information content can be extracted from the voice data by using various technical methods.
  • the target object to be shared may be extracted and/or determined. This solution does not limit the use of certain specific ways to determine the target object to be shared from the voice data.
  • a person skilled in the art can arbitrarily select various methods in the prior art to implement the step according to an implementation environment (such as, but not limited to, hardware limitation, software limitation, security restriction, etc.), for example, can be based on voice content. The way to retrieve audio is fine.
  • the information of the target object to be shared in the voice data may be acquired according to the keyword extraction manner.
  • Keyword extraction is to identify a given word, that is, a keyword, in a continuous utterance, while ignoring other non-speech words other than keywords in the utterance, such as breathing, coughing, music, background noise, and the like.
  • the front end processing is performed on the voice data to eliminate the influence of noise and different speakers, and then feature extraction is performed to identify the keyword.
  • keyword extraction can be performed based on the Filter model.
  • Hidden Marcov Models (HMM) can also be used for keyword extraction. It can also be based on the On-Line Garbage (OLG) model.
  • HMM Hidden Marcov Models
  • OOG On-Line Garbage
  • the word extraction can be selected according to actual needs, and is not limited herein.
  • the required information content may also be extracted from the voice data based on other audio processing techniques.
  • ASR automatic speech recognition
  • word segmentation techniques are used to intercept keywords. This solution is not limited herein and will not be described in detail.
  • the device simultaneously identifies a plurality of target objects based on the voice data, and may determine the required target object according to the user's selection, for example, displaying the identified target objects to the user for the user to select.
  • the target object may be determined according to the priority of the target object, for example, the target object with the highest priority is used as the final target object.
  • the identified multiple target objects may be determined as finalized objects. target.
  • the user inputs two kinds of voice data, namely “Wang Xiaohua” and “Share to Wang Xiaohua”.
  • the device recognizes three objects “Wang Xiaohua”, “School Flower” and “Fen Xiang”, and the most frequently contacted objects can be selected.
  • the final target object and given to the user first the user can also be provided with multiple target objects for selection.
  • the final target object can also be determined by other means, which can be determined according to actual needs.
  • the device may acquire all the information of each object in the various target applications in advance, and after receiving the information of the target object, all the pre-existing target objects are matched in the target application.
  • the device can obtain contact information in an Instant Messaging (IM) application and match the objects “Wang Xiaohua”, “School Flower”, and “Fen” in the voice data input by the user in the previous example. ", it is determined that the two target objects "Wang Xiaohua” and "School Flower” which are pre-existing are matched, and the two target objects are provided to the user for selection.
  • IM Instant Messaging
  • the device may convert information of each object in the various target applications into text information, and when the device determines the information of the target object, matches the corresponding target object in all of the text information. This scenario is not described here.
  • the target application is matched, for example, if an application is matched, step 114 may be performed. If multiple applications are matched, one application may be randomly selected as the target application and step 114 is performed. It is also possible to select an application as the target application according to the frequency of use and perform step 114. Of course, the target application may be selected in other manners, and may be set according to actual needs.
  • the target application can also be determined manually by the user. Specifically, after the device that shares the information based on the voice determines the target object, acquiring a plurality of applications related to the target object, displaying the first identifiers of the plurality of applications, and determining, according to the selection of the first identifier by the user, determining The selected first identifier corresponds to the target application. For example, if the information of the target object is the name of the target object, for example, “Wang Xiaohua”, and the target object uses the name in three applications such as Weibo, WeChat, and Taobao, the device based on the voice sharing information is based on the target object.
  • the identifiers or names of the three applications may be displayed, and the user selects one of them as the target application, for example, clicking the name of one of the applications to send the first confirmation information, and the clicked name corresponds to The application is the target application.
  • Step 114 Send information about the target page to the target object through the target application.
  • the information related to the target page may include at least one of the following information: a link of the target page, a snapshot of the target page, and information of the user-defined target page.
  • the user corresponding to the target object can open the target page through the link, and the related information of the target page can be obtained through the snapshot of the target page.
  • the user who inputs the sharing instruction can customize the information of the target page, for example, edit the text or input the voice data on the snapshot of the target page, so that the target user corresponding to the target object can see the customized information or hear the Custom information, that is, custom information can be presented in a voice and/or text manner.
  • the method further includes:
  • Custom information includes custom voice messages or custom text messages
  • the user can edit the related information of the target page that he wants to send, which can be voice editing or text editing.
  • the custom voice information is input in the process of displaying the edit prompt information
  • the custom text information is input in the input box displaying the edit prompt information
  • the device generates the information to be sent to the target object according to the customized information and the related information of the target page.
  • the information of the target page is a registered trademark of Microsoft Corporation.
  • the method further includes:
  • the cancel instruction sent by the user is received, and the operation triggered according to the share instruction is suspended according to the cancel instruction.
  • the user incorrectly triggers the sharing instruction.
  • the user does not want to share the related information of the target page.
  • the user can send a cancel instruction, for example, clicking a preset button on the target page to trigger the cancellation instruction.
  • triggering a sharing instruction by performing a preset operation on the target page, such as continuously clicking the target page, or performing a preset operation on the terminal displaying the target page to trigger a sharing instruction, such as shaking the terminal, or triggering by sending a voice command.
  • Share the instructions which can be selected according to actual needs, and will not be described here.
  • the device based on the voice sharing information recognizes the cancel instruction, the operation triggered according to the sharing instruction is terminated, that is, the sharing operation is terminated.
  • determining the target object to be shared based on at least the voice data includes:
  • a target object corresponding to the selected second identifier is determined corresponding to the user's selection of the second identifier.
  • determining the target object corresponding to the selected second identifier includes:
  • a plurality of target objects corresponding to the plurality of second identifiers are determined corresponding to the plurality of second identifiers selected by the user.
  • the target object selected by the user may be one or more.
  • the device based on the voice sharing information matches the partial name of the target object.
  • Matching a plurality of related objects namely "Li Xiaohua”, “Zhao Xiaohua” and “Wang Xiaohua”, respectively, displaying a second identifier of the plurality of objects to the user, for example, displaying respective avatars of the plurality of objects, and the user clicks on the Liu Xiaohua
  • the avatar based on the voice sharing information, determines Liu Xiaohua as the target object according to the avatar selected by the user.
  • the information of the target page can be sent to the target object through the target application, so that the user does not need to pass through as much as possible.
  • Manual operation to share the information of the target page thereby avoiding complicated sharing processes and bringing great convenience to users.
  • This embodiment provides a method for sharing information based on voice to implement information sharing by using voice.
  • the execution subject of this embodiment is a device based on voice sharing information.
  • FIG. 1B it is a schematic flowchart of a method for sharing information based on voice according to the present embodiment.
  • the method includes:
  • step 101 the target page is displayed, and a sharing instruction input by the user is received.
  • the target page is the web page currently being displayed on the terminal display.
  • the sharing instruction of this embodiment is used to share information, such as information related to a target page.
  • the receiving user sends a sharing instruction by pressing or clicking a preset button on the target page, or inputting a sharing instruction by performing a preset operation on the target page, for example, long pressing the target.
  • the page, or the terminal that displays the target page performs a preset operation to input a sharing command, such as shaking the terminal, or inputting a sharing command by sending a voice command, which may be selected according to actual needs, and details are not described herein again.
  • Step 102 Acquire voice data according to a sharing instruction, where the voice data includes at least information of the target application and the target object.
  • the voice data is the voice sent by the user, and the target application included in the voice data is to share the target page with the application used by other users, and the target object is the user to whom the target page is to be sent.
  • the target application of this embodiment is one of the following applications: a short message, a social application.
  • the social application may be an application such as WeChat, Weibo, nailing, or other applications, such as Taobao, Alipay, facebook, twitter, instagram, etc., as long as the application that can perform information interaction with other objects can be used as the target application.
  • the voice data input by the user during the continuous pressing of the button may be acquired, and if the sharing instruction input by other methods is used, the user may recognize After inputting the instruction, the voice input by the user in the preset time is acquired to obtain the voice data.
  • the information of the target application is information related to the target application
  • the information of the target object is information related to the target object.
  • the device can acquire the information of the target application and the target object according to the keyword extraction manner. For example, the device can obtain information and target information of various target applications in advance. The information of each object in the application, such as counting the names of some popular applications, and then analyzing and identifying the voice data. After identifying the target application, the voice data input by the user can be matched in the target application to identify the target. Object.
  • Step 103 Send, according to the voice data, information related to the target page to the target object by using the target application.
  • the information related to the target page may include at least one of the following information: a link of the target page, a snapshot of the target page, and information of the user-defined target page.
  • the user corresponding to the target object can open the target page through the link, and the related information of the target page can be obtained through the snapshot of the target page.
  • the user who inputs the sharing instruction can customize the information of the target page, for example, edit the text or input the voice data on the snapshot of the target page, so that the target user corresponding to the target object can see the customized information or hear the Custom information, that is, custom information can be presented in a voice and/or text manner.
  • the method further includes:
  • the user can edit the related information of the target page that he wants to send, which can be voice editing or text editing.
  • the custom voice information is input in the process of displaying the edit prompt information
  • the custom text information is input in the input box displaying the edit prompt information
  • the device generates the information to be sent to the target object according to the customized information and the related information of the target page.
  • the information of the target page is a registered trademark of Microsoft Corporation.
  • the method further includes:
  • the cancel instruction sent by the user is received, and the operation triggered according to the share instruction is suspended according to the cancel instruction.
  • the user incorrectly triggers the sharing instruction.
  • the user does not want to share the related information of the target page.
  • the user can send a cancel instruction, for example, clicking a preset button on the target page to trigger the cancellation instruction.
  • triggering a sharing instruction by performing a preset operation on the target page, such as continuously clicking the target page, or performing a preset operation on the terminal displaying the target page to trigger a sharing instruction, such as shaking the terminal, or triggering by sending a voice command.
  • Share the instructions which can be selected according to actual needs, and will not be described here.
  • the device based on the voice sharing information recognizes the cancel instruction, the operation triggered according to the sharing instruction is terminated, that is, the sharing operation is terminated.
  • the method further includes:
  • the target application corresponding to the selected third identifier is determined.
  • the names of the two applications are “Tmall” and “Daily Express”
  • the device of the information matches the part name of the target application, and matches two applications to be determined, namely “Tmall” and “Daily Express” respectively, and then displays the plurality of to-be-determined applications to the user, for example, displaying the application to be determined.
  • the identifier is that the user clicks the identifier of the Tmall to send a third confirmation message, and the device based on the voice sharing information identifies the third confirmation information, and determines that Tmall is the target application.
  • a target object corresponding to the selected second identifier is determined corresponding to the user's selection of the second identifier.
  • the device based on the voice sharing information matches the partial name of the target object.
  • Matching multiple objects namely "Li Xiaohua”, “Zhao Xiaohua” and "Wang Xiaohua”
  • display the plurality of objects to the user for example, displaying the avatar of the object to be determined, the user clicking the avatar of Liu Xiaohua, and the device for sharing information based on voice Identify the choice and determine Liu Xiaohua as the target.
  • the information of the target page is sent to the target object through the target application by issuing the voice including the target application and the target object, so that the manual operation is not required as much as possible. Share the information on the target page to avoid complicated sharing processes and bring great convenience to users.
  • This embodiment further supplements the method for voice sharing information according to the second embodiment.
  • FIG. 2A it is a schematic flowchart of a method for sharing information based on voice according to the present embodiment.
  • the method includes:
  • Step 201 Display a target page, and receive a sharing instruction sent by the user by pressing a preset button on the target page, and perform step 202.
  • the target page is the page currently being displayed on the terminal display, such as a web page.
  • the sharing instruction of this embodiment is for sharing information, such as sharing information related to the target page.
  • the preset position on the target page may be provided with a virtual button that triggers a sharing instruction when the user presses the virtual button.
  • Step 202 If it is recognized that the time when the user continues to press the preset button exceeds the first preset threshold, the first prompt information is displayed, and the first voice data input by the user during the display of the first prompt information is obtained, and the first The voice data is used as the information of the target application, and step 203 is performed.
  • the voice data is the voice sent by the user.
  • the target application mentioned in the voice data is to share the target page with the application used by the target user, and the target object is the target user who wants to send the information of the target page to the target application. Account.
  • the target application of this embodiment is one of the following applications: a short message, a social application.
  • the social application may be an application such as WeChat, Weibo, nailing, or other applications, such as Taobao, Alipay, facebook, twitter, instagram, etc., as long as the application that can perform information interaction with other objects can be used as the target application.
  • the first preset threshold of the embodiment may be set according to actual needs, for example, 0.5 seconds, and the first prompt information is displayed on the display screen of the terminal, where the first prompt information is used to remind the user to input information of the target application. This allows the user to enter information about the target application.
  • step 203 the second prompt information is displayed, and the second voice data input by the user in the second prompt information display process is acquired, and the second voice data is used as the target object information, and step 204 is performed.
  • the second prompt information is displayed to prompt the user to input the information of the target object.
  • Step 204 Analyze the voice data, obtain the information of the target application, and the information of the target object, and perform step 205.
  • the device may analyze the voice data acquired by the user during the continuous pressing of the preset button to acquire information of the target application and the target object. Specifically, the device may store the voice data to different locations according to the first prompt information and the second prompt information to distinguish the information of the target application and the target object.
  • Step 205 If it is found that the target application is installed, the target application queries whether there is a target object. If the query result is yes, step 206 is performed; otherwise, step 207 is performed.
  • step 205 Before performing step 205, it may be determined whether the target application has been installed by the terminal. If the determination result is yes, step 205 is performed. Otherwise, the target application can be automatically downloaded and installed through the network, and then step 205 is executed, or information for prompting the user not to install the target application is displayed.
  • Step 206 Send information related to the target page to the target object.
  • the device opens the dialog interface corresponding to the target object, and sends the information of the target page to the target object through the dialog interface, and the process can be displayed to the user and then returned to the target page.
  • the dialog boundary corresponding to the target object is opened Between the faces, it also includes:
  • the information of the target object in the target application such as avatar information
  • the user sends a confirmation instruction by clicking the displayed information, for example, clicking the displayed avatar.
  • the user can further confirm the target object to avoid sending errors.
  • Step 207 Display query failure information indicating that there is no target object in the target application.
  • the information is, for example, a name, and it is very likely that the user who inputs the sharing instruction misreads the information of the target user in the target application, so that the query failure information can be displayed. To prompt the user that the target object does not have the target object.
  • the user A opens the target page 210 as shown in FIG. 2B, and feels that the item displayed on the target page is very good. If he wants to recommend to the target user B, the "Share" button 211 on the target page can be pressed.
  • the first prompt information 212 as shown in FIG. 2C is displayed.
  • the content of the first prompt information 212 is “Please input the name of the target application”, and the user can input the voice “WeChat” at the device.
  • the second prompt information 213 shown in FIG. 2D is displayed, and the content of the second prompt information 213 is “Please enter the name of the target object”, and the user can input the voice.
  • the editing prompt information 214 can be displayed, the content is "please input custom information", the user A inputs the voice "this thing is very good”, or the prompt box 215 is displayed as shown in Fig. 2F, the user input Customizing the information, the user enters the text "This is a very good thing” and then clicks on the confirmation button 216.
  • the device After the device recognizes that the user clicks the confirmation button 216, the information of the target application input by the user is retrieved, and the name of the target object is matched in the target application. If the matching is successful, the target object shown in FIG. 2G is displayed in the target application.
  • the avatar 217 in the middle may also display some other related information.
  • User B opens the information sent by User A through WeChat, and recognizes that this is a photo through the preview image, and then clicks the photo to enlarge the display.
  • the terminal corresponding to User B plays the user A input on the colleague who presents the snapshot of the target page. Customize voice messages or display a snapshot of the target page with User A custom information.
  • the sharing instruction is input, and the first prompt information and the second prompt information are displayed to prompt the user to input the information of the target application and the target object, which not only makes the user try to There is no need to share the information of the target page through complicated manual operations, which brings great convenience to the user, and enables the device to accurately identify the information of the target application and the target object, thereby avoiding sending the information of the target page to the wrong target object as much as possible.
  • This embodiment further supplements the method for voice sharing information according to the second embodiment.
  • FIG. 3A it is a schematic flowchart of a method for sharing information based on voice according to the present embodiment.
  • the method includes:
  • Step 301 Display a target page, and receive a sharing instruction sent by the user by pressing a preset button on the target page, and perform step 302.
  • the target page is the page currently being displayed on the terminal display, such as a web page.
  • the sharing instruction of this embodiment is for sharing information, such as sharing information related to the target page.
  • Step 302 Identify whether two pieces of voice sub-data are included in the voice data generated by the user during the continuous pressing of the preset button, and the time interval between the two pieces of voice sub-data exceeds a second preset threshold, and if the recognition result is yes, Go to step 303.
  • the voice data is the voice sent by the user.
  • the target application mentioned in the voice data is to share the target page with the application used by the target user, and the target object is the target user who wants to send the information of the target page to the target application. Account.
  • the target application of this embodiment is one of the following applications: a short message, a social application.
  • the social application may be an application such as WeChat, Weibo, nailing, or other applications, such as Taobao, Alipay, facebook, twitter, instagram, etc., as long as the application that can perform information interaction with other objects can be used as the target application.
  • the second preset threshold of the embodiment may be set as needed, for example, 0.5 seconds or 1 second, and may of course be set to other times, which is not limited herein.
  • the user is displayed with information for prompting the unrecognizable to inform the user to re-enter the voice data, and the time interval between the two voice sub-data is extended when the two voice sub-data are input. .
  • step 303 the first piece of voice sub-data is used as the information of the target application, and the second piece of voice sub-data is used as the information of the target object, and step 304 is performed.
  • Step 304 Analyze the voice data, obtain the information of the target application, and the information of the target object, and perform step 305.
  • the device may analyze the voice data acquired by the user during the continuous pressing of the preset button to acquire information of the target application and the target object. Specifically, the device may store the voice data to different locations according to the first prompt information and the second prompt information to distinguish the mesh. Information about the application and target object.
  • the method of analyzing the voice data may adopt a keyword extraction manner, and how to implement the technology belongs to the prior art, and details are not described herein again.
  • Step 305 If it is found that the target application is installed, the target application queries whether there is a target object. If the query result is yes, step 306 is performed; otherwise, step 307 is performed.
  • step 305 it may be determined whether the target application has been installed by the terminal. If the determination result is yes, step 305 is performed. Otherwise, the target application can be automatically downloaded and installed through the network, and then step 305 is performed, or information for prompting the user not to install the target application is displayed.
  • Step 306 sending information related to the target page to the target object.
  • the device opens the dialog interface corresponding to the target object, and sends information related to the target page to the target object through the dialog interface, and the process can be displayed to the user and then returned to the target page.
  • the method further includes:
  • the target object may be displayed in the target application, and the information is, for example, avatar information, and the user sends a confirmation command by clicking the displayed information, for example, clicking the displayed avatar to send. Confirm the order. In this way, the user can further confirm the target object to avoid sending errors.
  • Step 307 displaying query failure information indicating that there is no target object in the target application.
  • the information is, for example, a name, and it is very likely that the user who inputs the sharing instruction misreads the information of the target user in the target application, so that the query failure may be displayed.
  • Information to prompt the user that the target object does not have the target object.
  • the second preset threshold is 0.5 seconds.
  • the user A opens the target page 310 as shown in FIG. 3B, and feels that the item displayed on the target page is very good. If you want to recommend to the user B, you can continuously press the “Share” button 311 on the target page. At this time, the user can Enter two pieces of speech sub-data, which are “WeChat” and “Xiaofang” respectively. The time interval between the two pieces of speech sub-data is 1 second. Next, the "Share" button 311 is no longer pressed.
  • the device After the device recognizes that the user releases the "Share" button 311, the information of the target application input by the user is retrieved, and the name of the target object is matched in the target application. If the matching is successful, the target object as shown in FIG. 3C is displayed. In the avatar 312 in the target application, the user A clicks the avatar 312 to send an acknowledgment command, and after receiving the acknowledgment command, the device sends a snapshot of the target page with the customized information to the target object.
  • User B opens the information sent by User A through WeChat, and recognizes that this is a photo through the preview image, and then clicks the photo to enlarge the display.
  • the terminal corresponding to User B plays the user A input on the colleague who presents the snapshot of the target page. Customize voice messages or display a snapshot of the target page with User A custom information.
  • the sharing instruction is triggered, and by identifying whether the two pieces of voice sub data are included in the voice data generated by the user during the continuous pressing of the preset button, the user does not need to pass the complicated
  • the manual operation to share the information of the target page brings great convenience to the user, and enables the device to accurately identify the information of the target application and the target object, thereby avoiding sending the information of the target page to the wrong target object as much as possible.
  • This embodiment provides an apparatus for sharing information based on voice, for performing the foregoing method.
  • FIG. 4A it is a schematic structural diagram of a device for sharing information based on voice according to the present embodiment.
  • the apparatus of this embodiment includes a display module 401, a receiving module 402, an obtaining module 403, a determining module 404, and a sending module 405.
  • the display module 401 is configured to display a target page; the receiving module 402 is configured to receive a sharing instruction input by the user; the obtaining module 403 is configured to acquire voice data according to the sharing instruction, where the voice data includes at least: information related to the target object to be shared.
  • the determining module 404 is configured to determine a target object to be shared based on the voice data, and determine a target application to be shared; the sending module 405 is configured to send, by the target application, information related to the target page to the target object.
  • the determining module 404 includes a target obtaining submodule 412, a first display identifier submodule 413, and a first determining submodule 414, wherein the target obtaining submodule 412 is configured to acquire the target object. a plurality of applications related to the information; the first display identifier sub-module 413 is configured to display a first identifier of the plurality of applications; the first determining sub-module 414 is configured to determine, according to the user's selection of the first identifier, the selected first Identify the corresponding target application.
  • the first determining submodule 414 is specifically configured to:
  • a target application corresponding to the selected first identifier is determined corresponding to the first identifier selected by the user via a voice input or a click operation.
  • the first determining submodule 414 is specifically configured to:
  • a plurality of target applications corresponding to the plurality of first identifiers are determined.
  • the determining module 404 further includes a first matching submodule 415, a second display identifier submodule 416, and a second determining submodule 417, wherein the first matching submodule 415 is used in the target.
  • the application matches a plurality of objects related to the information of the target object; the second display identifier sub-module 416 is configured to display the second identifier of the plurality of objects; and the second determining sub-module 417 is configured to select the second identifier corresponding to the user, A target object corresponding to the selected second identifier is determined.
  • the second determining submodule 417 is specifically configured to:
  • a plurality of target objects corresponding to the plurality of second identifiers are determined corresponding to the plurality of second identifiers selected by the user.
  • the voice data includes information of the target application.
  • the determining module 404 further includes a second matching submodule 420, a third display identifier submodule 421, and a third determining submodule 422.
  • the second matching sub-module 420 is configured to acquire a plurality of applications related to information of the target application;
  • the third display identifier sub-module 421 is configured to display a third identifier of the plurality of applications;
  • the third determining sub-module 422 is configured to correspond to The user selects the third identifier to determine a target application corresponding to the selected third identifier.
  • the target obtaining submodule 412, the first display identifier submodule 413, and the first determining submodule 414 may exist simultaneously with the first matching submodule 415, the second display identifier submodule 416, and the second determining submodule 417.
  • the target obtaining sub-module 412, the first display identifier sub-module 413, and the first determining sub-module 414 may be associated with the second matching sub-module 420, the third display identifier sub-module 421, and the third determining.
  • the sub-modules 422 exist at the same time, and may also exist separately.
  • the first matching sub-module 415, the second display identifier sub-module 416, and the second determining sub-module 417 may be combined with the second matching sub-module 420, the third display identifier sub-module 421, and the The three determining sub-modules 422 exist simultaneously, and may also exist separately, and the case where the above modules exist is shown in FIG. 4B.
  • the receiving module 402 is specifically configured to: receive a sharing instruction sent by the user by pressing a preset button on the target page;
  • the obtaining module 403 is specifically configured to: acquire voice data input by the user during the continuous pressing of the button.
  • the information of the target object includes at least one of the following information: a name, an avatar of the target object in the target application.
  • the information of the target page includes at least one of the following information: a link of the target page, a snapshot of the target page, and information of the user-defined target page.
  • the target application is one of the following applications: SMS, social application.
  • the voice data further includes user-defined information.
  • the information of the target page also includes custom information.
  • the information of the target page is sent to the target object through the target application by issuing the voice including the target application and the target object, so that the manual operation is not required as much as possible. Share the information on the target page to avoid complicated sharing processes and bring great convenience to users.
  • This embodiment further supplements the description of the apparatus of the fourth embodiment.
  • This embodiment mainly describes the specific operation mode of the acquisition module.
  • the acquisition module 404 in the device includes a first acquisition submodule 4031 and a second acquisition submodule 4032.
  • the first obtaining sub-module 4031 is configured to display the first prompt information if the time when the user continuously presses the preset button exceeds the first preset threshold, and obtain the first input by the user during the display process of the first prompt information.
  • the determining module 404 uses the first voice data as the information of the target application; the second obtaining sub-module 4032 is configured to display the second prompt information, and acquire the second input by the user during the second prompt information display process.
  • the voice data correspondingly, the determination module 404 uses the second voice data as the information of the target object.
  • the determining module 404 is specifically configured to:
  • the recognition result is yes, the first piece of speech subdata is used as the information of the target application, and the second piece of speech subdata is used as the information of the target object.
  • the determining module 404 is specifically configured to:
  • the keyword extraction method is used to obtain the information of the target application and the target object in the voice data.
  • the sending module 405 of this embodiment is specifically configured to:
  • the target application is queried whether there is a target object
  • the dialog interface corresponding to the target object is opened, and the information of the target page is sent to the target object through the dialog interface.
  • the sending module 405 is further configured to:
  • the query failure information indicating that there is no target object in the target application is displayed.
  • the sending module 405 is further configured to:
  • Trigger display module 401 displays information of the target application and/or the target object
  • the apparatus of this embodiment further includes an editing module 502, where the editing module 502 is configured to:
  • the device further includes a canceling module 503, configured to receive a cancel instruction sent by the user, and suspend the operation triggered according to the sharing instruction according to the cancel instruction.
  • the cancellation module is shown coupled to display module 503 in Figure 5B, which can be coupled to any of the modules of Figure 5B to cancel the operation of the corresponding module.
  • the sharing instruction is triggered, and the information of the target application and the target object are identified in different manners, so that the user does not need to share the information of the target page through complicated manual operations, and gives the user It brings great convenience and enables the device to accurately identify the target application and the target object, so as to avoid sending the information of the target page to the wrong target object.
  • This embodiment provides another apparatus for sharing information based on voice for performing the foregoing method.
  • the device of this embodiment includes an input device, a processor, and a display screen;
  • the processor is configured to control the display screen to display the target page;
  • the input device is configured to receive a sharing instruction input by the user, and obtain voice data according to the sharing instruction, where the voice data includes at least information related to the target object to be shared;
  • the processor is further configured to: Determining a target object to be shared based on at least the voice data, and determining a target application to be shared, and transmitting information related to the target page to the target object through the target application.
  • the processor is specifically configured to:
  • the control display screen displays the first identification of the plurality of applications
  • a target application corresponding to the selected first identifier is determined corresponding to the user's selection of the first identifier.
  • the processor is specifically configured to:
  • a target object corresponding to the selected second identifier is determined corresponding to the user's selection of the second identifier.
  • the voice data further includes information of the target application.
  • the processor is specifically configured to:
  • the target application corresponding to the selected third identifier is determined.
  • the processor is specifically used to:
  • the display screen is displayed to display the first prompt information, and the user is obtained.
  • the first voice data input during the display of the first prompt information, and the first voice data is used as the information of the target application;
  • the processor is specifically configured to:
  • the recognition result is yes, the first piece of speech subdata is used as the information of the target application, and the second piece of speech subdata is used as the information of the target object.
  • the processor is further configured to control the display screen to display the editing prompt information, and obtain the customized information input by the user during the display process of the editing prompt information, where the customized information includes the customized voice information or the customized text information;
  • the input device, the processor, the display screen, the processor, and the display screen of the embodiment are also used to perform the foregoing corresponding methods, and details are not described herein.
  • FIG. 6 is a schematic structural diagram of a device for sharing information based on voice according to the present embodiment.
  • device 600 can be a mobile phone, a computer, a digital broadcast terminal, a messaging device, a gaming console, a tablet device, a medical device, a fitness device, a personal digital assistant, and the like.
  • apparatus 600 can include one or more of the following components: processing component 602, memory 604, power component 606, multimedia component 608, audio component 610, input/output (I/O) interface 612, sensor group A piece 614, and a communication component 616.
  • Processing component 602 typically controls the overall operation of device 600, such as operations associated with display, telephone calls, data communications, camera operations, and recording operations.
  • Processing component 602 can include one or more processors 620 to execute instructions to perform all or part of the steps of the above described methods.
  • processing component 602 can include one or more modules to facilitate interaction between component 602 and other components.
  • processing component 602 can include a multimedia module to facilitate interaction between multimedia component 608 and processing component 602.
  • Memory 604 is configured to store various types of data to support operation at device 600. Examples of such data include instructions for any application or method operating on device 600, contact data, phone book data, messages, pictures, videos, and the like.
  • the memory 604 can be implemented by any type of volatile or non-volatile storage device or a combination thereof, such as static random access memory (SRAM), electrically erasable programmable read only memory (EEPROM), erasable Programmable Read Only Memory (EPROM), Programmable Read Only Memory (PROM), Read Only Memory (ROM), Magnetic Memory, Flash Memory, Disk or Optical Disk.
  • SRAM static random access memory
  • EEPROM electrically erasable programmable read only memory
  • EPROM erasable Programmable Read Only Memory
  • PROM Programmable Read Only Memory
  • ROM Read Only Memory
  • Magnetic Memory Flash Memory
  • Disk Disk or Optical Disk.
  • Power component 606 provides power to various components of device 600.
  • Power component 606 can include a power management system, one or more power sources, and other components associated with generating, managing, and distributing power for device 600.
  • the multimedia component 608 includes a display screen between the device 600 and the user that provides an output interface.
  • the display screen can include a liquid crystal display (LCD) and a touch panel (TP). If the display includes a touch panel, the display can be implemented as a touch screen to receive input signals from the user.
  • the touch panel includes one or more touch sensors to sense touches, slides, and gestures on the touch panel. The touch sensor can sense not only the boundaries of the touch or sliding action, but also the duration and pressure associated with the touch or slide operation.
  • the multimedia component 608 includes a front camera and/or a rear camera. When the device 600 is in an operation mode, such as a shooting mode or a video mode, the front camera and/or the rear camera can receive external multimedia data. Each front and rear camera can be a fixed optical lens system or have focal length and optical zoom capabilities.
  • the audio component 610 is configured to output and/or input an audio signal.
  • audio component 610 includes a microphone (MIC) that is configured to receive an external audio signal when device 600 is in an operational mode, such as a call mode, a recording mode, and a voice recognition mode.
  • the received audio signal may be further stored in memory 604 or transmitted via communication component 616.
  • audio component 610 also includes a speaker for outputting an audio signal.
  • the I/O interface 612 provides an interface between the processing component 602 and the peripheral interface module, which may be a keyboard, a click wheel, a button, or the like. These buttons can include but are not limited to: home button, volume button, start button And lock button.
  • Sensor assembly 614 includes one or more sensors for providing device 600 with a status assessment of various aspects.
  • sensor component 614 can detect an open/closed state of device 600, a relative positioning of components, such as a display and a keypad of device 600, and sensor component 614 can also detect a change in position of one component of device 600 or device 600, user The presence or absence of contact with device 600, device 600 orientation or acceleration/deceleration and temperature variation of device 600.
  • Sensor assembly 614 can include a proximity sensor configured to detect the presence of nearby objects without any physical contact.
  • Sensor assembly 614 may also include a light sensor, such as a CMOS or CCD image sensor, for use in imaging applications.
  • the sensor component 614 can also include an acceleration sensor, a gyro sensor, a magnetic sensor, a pressure sensor, or a temperature sensor.
  • Communication component 616 is configured to facilitate wired or wireless communication between device 600 and other devices.
  • the device 600 can access a wireless network based on a communication standard, such as WiFi, 2G or 3G, or a combination thereof.
  • communication component 616 receives broadcast signals or broadcast associated information from an external broadcast management system via a broadcast channel.
  • communication component 616 also includes a near field communication (NFC) module to facilitate short range communication.
  • NFC near field communication
  • the NFC module can be implemented based on radio frequency identification (RFID) technology, infrared data association (IrDA) technology, ultra-wideband (UWB) technology, Bluetooth (BT) technology, and other technologies.
  • RFID radio frequency identification
  • IrDA infrared data association
  • UWB ultra-wideband
  • Bluetooth Bluetooth
  • device 600 may be implemented by one or more application specific integrated circuits (ASICs), digital signal processors (DSPs), digital signal processing devices (DSPDs), programmable logic devices (PLDs), field programmable A gate array (FPGA), controller, microcontroller, microprocessor, or other electronic component implementation for performing the above methods.
  • ASICs application specific integrated circuits
  • DSPs digital signal processors
  • DSPDs digital signal processing devices
  • PLDs programmable logic devices
  • FPGA field programmable A gate array
  • controller microcontroller, microprocessor, or other electronic component implementation for performing the above methods.
  • the communication component 616, the audio component 610, and the input/output interface 612 involved in the embodiment of FIG. 6 can be implemented as an input device.
  • the user wants to share the information of the target page, and can send the information of the target page to the target object through the target application by issuing the voice including the target application and the target object, so that the information of the target page is not required to be manually operated.
  • the user wants to share the information of the target page, and can send the information of the target page to the target object through the target application by issuing the voice including the target application and the target object, so that the information of the target page is not required to be manually operated.
  • the invention also includes a mobile terminal comprising the apparatus of any of the preceding embodiments.

Abstract

L'invention concerne un procédé, un dispositif et un terminal mobile de partage d'informations par voie vocale. Le procédé comprend les étapes suivantes : affichage d'une page cible et réception d'une instruction de partage saisie par un utilisateur (111); acquisition, conformément à l'instruction de partage, de données vocales comprenant des informations associées à un objet cible à partager (112); détermination, sur la base des données vocales, de l'objet cible à partager (113); et transmission, par une application cible et à l'objet cible, des informations de la page cible (114). Le mode de réalisation de l'invention peut empêcher un processus de partage compliqué, en apportant une commodité maximale à un utilisateur.
PCT/CN2017/097012 2016-08-23 2017-08-11 Procédé, dispositif et terminal mobile de partage d'informations par voie vocale WO2018036392A1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201610710046.1 2016-08-23
CN201610710046.1A CN107767864B (zh) 2016-08-23 2016-08-23 基于语音分享信息的方法、装置与移动终端

Publications (1)

Publication Number Publication Date
WO2018036392A1 true WO2018036392A1 (fr) 2018-03-01

Family

ID=61246382

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2017/097012 WO2018036392A1 (fr) 2016-08-23 2017-08-11 Procédé, dispositif et terminal mobile de partage d'informations par voie vocale

Country Status (3)

Country Link
CN (1) CN107767864B (fr)
TW (1) TW201807565A (fr)
WO (1) WO2018036392A1 (fr)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110333836A (zh) * 2019-07-05 2019-10-15 网易(杭州)网络有限公司 信息的投屏方法、装置、存储介质和电子装置

Families Citing this family (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108470566B (zh) * 2018-03-08 2020-09-15 腾讯科技(深圳)有限公司 一种应用操作方法以及装置
CN110347303A (zh) * 2018-04-04 2019-10-18 腾讯科技(深圳)有限公司 一种信息处理方法及相关设备
CN110544473B (zh) 2018-05-28 2022-11-08 百度在线网络技术(北京)有限公司 语音交互方法和装置
CN108920119A (zh) * 2018-06-29 2018-11-30 维沃移动通信有限公司 一种分享方法及移动终端
CN109065049A (zh) * 2018-09-13 2018-12-21 苏州思必驰信息科技有限公司 基于语音交互的智能终端的社交分享方法及系统、智能终端设备
CN110728586A (zh) * 2019-09-25 2020-01-24 支付宝(杭州)信息技术有限公司 数据的分享方法和装置、应用的推广方法和装置
CN111583929A (zh) * 2020-05-13 2020-08-25 军事科学院系统工程研究院后勤科学与技术研究所 使用离线语音的控制方法、装置及可识读设备
CN113037924B (zh) * 2021-01-27 2022-11-25 维沃移动通信有限公司 语音发送方法、装置、电子设备和可读存储介质
CN113113005A (zh) * 2021-03-19 2021-07-13 大众问问(北京)信息科技有限公司 语音数据处理方法、装置、计算机设备和存储介质

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6017219A (en) * 1997-06-18 2000-01-25 International Business Machines Corporation System and method for interactive reading and language instruction
CN103680497A (zh) * 2012-08-31 2014-03-26 百度在线网络技术(北京)有限公司 基于视频的语音识别系统及方法
CN104023040A (zh) * 2013-03-01 2014-09-03 联想(北京)有限公司 一种信息处理的方法及装置
CN104063155A (zh) * 2013-03-20 2014-09-24 腾讯科技(深圳)有限公司 内容分享方法、装置及电子设备
CN104065718A (zh) * 2014-06-19 2014-09-24 深圳米唐科技有限公司 实现智能音箱社交分享的方法及系统
CN104580534A (zh) * 2015-02-06 2015-04-29 联想(北京)有限公司 信息处理方法、装置及电子设备

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR100921554B1 (ko) * 2005-08-30 2009-10-14 주식회사 케이티 음성통화중에 다양한 콘텐츠를 공유 및 제어할 수 있는콘텐츠공유서비스를 제공하는 시스템 및 그 방법
US8447285B1 (en) * 2007-03-26 2013-05-21 Callwave Communications, Llc Methods and systems for managing telecommunications and for translating voice messages to text messages
CN105094801B (zh) * 2015-06-12 2019-12-24 阿里巴巴集团控股有限公司 应用功能激活方法及装置
CN105100449B (zh) * 2015-06-30 2018-01-23 广东欧珀移动通信有限公司 一种图片分享方法及移动终端
CN105656753A (zh) * 2015-12-16 2016-06-08 魅族科技(中国)有限公司 一种发送方法及装置

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6017219A (en) * 1997-06-18 2000-01-25 International Business Machines Corporation System and method for interactive reading and language instruction
CN103680497A (zh) * 2012-08-31 2014-03-26 百度在线网络技术(北京)有限公司 基于视频的语音识别系统及方法
CN104023040A (zh) * 2013-03-01 2014-09-03 联想(北京)有限公司 一种信息处理的方法及装置
CN104063155A (zh) * 2013-03-20 2014-09-24 腾讯科技(深圳)有限公司 内容分享方法、装置及电子设备
CN104065718A (zh) * 2014-06-19 2014-09-24 深圳米唐科技有限公司 实现智能音箱社交分享的方法及系统
CN104580534A (zh) * 2015-02-06 2015-04-29 联想(北京)有限公司 信息处理方法、装置及电子设备

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110333836A (zh) * 2019-07-05 2019-10-15 网易(杭州)网络有限公司 信息的投屏方法、装置、存储介质和电子装置
CN110333836B (zh) * 2019-07-05 2023-08-25 网易(杭州)网络有限公司 信息的投屏方法、装置、存储介质和电子装置

Also Published As

Publication number Publication date
CN107767864A (zh) 2018-03-06
TW201807565A (zh) 2018-03-01
CN107767864B (zh) 2021-06-29

Similar Documents

Publication Publication Date Title
WO2018036392A1 (fr) Procédé, dispositif et terminal mobile de partage d'informations par voie vocale
WO2017000490A1 (fr) Procédé et appareil d'alarme
CN104407592B (zh) 一种调整智能家居设备运行状态的方法及装置
CN107832036B (zh) 语音控制方法、装置及计算机可读存储介质
WO2017054356A1 (fr) Procédé et appareil de commande de dispositif électronique, et terminal
WO2016095448A1 (fr) Procédé et appareil de mise en réseau d'équipement
US20220295119A1 (en) Method and apparatus for interacting in live stream
WO2018098865A1 (fr) Procédé et appareil de lecture de message
WO2016107283A1 (fr) Procédé et dispositif de commande d'applications
WO2016082513A1 (fr) Procédé et dispositif d'annonce de demande d'appel
WO2017016148A1 (fr) Procédé et dispositif de communication à base d'images
CN106791921A (zh) 视频直播的处理方法及装置
WO2017036039A1 (fr) Procédé d'aide à distance et client
WO2016050038A1 (fr) Procédé et dispositif de reconnaissance de message de communication
US10242678B2 (en) Friend addition using voiceprint analysis method, device and medium
WO2016155304A1 (fr) Procédé et dispositif de commande de point d'accès sans fil
WO2017008400A1 (fr) Procédé et dispositif de commande de dispositif intelligent
WO2017000486A1 (fr) Procédé et dispositif pour l'obtention de résultats de recherche
WO2016065760A1 (fr) Procédé et appareil d'invite de connexion de dispositif
WO2015143817A1 (fr) Procédé et dispositif d'exécution d'instruction d'utilisateur
WO2017080205A1 (fr) Procédé d'obtention d'informations associées à un papier peint de verrouillage d'écran, et dispositif utilisant ce procédé
CN108495168A (zh) 弹幕信息的显示方法及装置
WO2015165172A1 (fr) Procédé et dispositif de transmission d'informations dans un service vocal
US10764418B2 (en) Method, device and medium for application switching
CN109388699A (zh) 输入方法、装置、设备及存储介质

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 17842815

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 17842815

Country of ref document: EP

Kind code of ref document: A1