WO2019119771A1 - 语音交互方法、装置及计算机可读存储介质 - Google Patents

语音交互方法、装置及计算机可读存储介质 Download PDF

Info

Publication number
WO2019119771A1
WO2019119771A1 PCT/CN2018/093750 CN2018093750W WO2019119771A1 WO 2019119771 A1 WO2019119771 A1 WO 2019119771A1 CN 2018093750 W CN2018093750 W CN 2018093750W WO 2019119771 A1 WO2019119771 A1 WO 2019119771A1
Authority
WO
WIPO (PCT)
Prior art keywords
voice
user
behavior data
operation behavior
received
Prior art date
Application number
PCT/CN2018/093750
Other languages
English (en)
French (fr)
Inventor
马小莉
Original Assignee
深圳Tcl新技术有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 深圳Tcl新技术有限公司 filed Critical 深圳Tcl新技术有限公司
Publication of WO2019119771A1 publication Critical patent/WO2019119771A1/zh

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/44Arrangements for executing specific programs
    • G06F9/4401Bootstrapping
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • G10L2015/223Execution procedure of a spoken command

Definitions

  • the present application relates to the field of communications, and in particular, to a voice interaction method, apparatus, and computer readable storage medium.
  • intelligent voice has become the first path for artificial intelligence to enter daily life.
  • Major businesses have launched intelligent voice products, such as smart voice TV, intelligent voice audio, and intelligent voice navigation. Instrument and voice air detectors, etc.
  • Intelligent voice products bring great convenience to consumers by simulating the voice mode of people and using voice instead of traditional operation methods.
  • the main purpose of the present application is to provide a voice interaction method, apparatus, and computer readable storage medium, which are intended to provide more user-friendly voice assistance and improve user experience during voice interaction.
  • the present application provides a voice interaction method, where the voice interaction method includes the following steps:
  • the voice interaction function When receiving the user-triggered voice interaction function wake-up instruction, the voice interaction function is activated according to the voice interaction function wake-up instruction, and the preset voice response file is played to respond;
  • the personal operation behavior data of the user is acquired, and a corresponding voice guidance file is generated and played according to the personal operation behavior data to guide the user.
  • the voice interaction method further includes:
  • the operation behavior data of each user on the line is acquired, and a corresponding voice help file is generated and played according to the operation behavior data to assist the user;
  • the voice command of the user is received within the second preset time, the voice command is identified, and a corresponding operation is performed according to the recognition result.
  • the step of acquiring the operation behavior data of each user on the line and generating a corresponding voice help file according to the operation behavior data and playing to assist the user includes:
  • the step of acquiring the personal operation behavior data of the user, and generating a corresponding voice guidance file according to the personal operation behavior data and playing to guide the user includes:
  • the method includes:
  • the voice command of the user is received within the first preset time, the voice command is identified, and a corresponding operation is performed according to the recognition result.
  • the present application further provides a voice interaction device, including: a memory, a processor, and a voice interaction program stored on the memory and operable on the processor, The following steps are performed when the speech interaction program is executed by the processor:
  • the voice interaction function When receiving the user-triggered voice interaction function wake-up instruction, the voice interaction function is activated according to the voice interaction function wake-up instruction, and the preset voice response file is played to respond;
  • the personal operation behavior data of the user is acquired, and a corresponding voice guidance file is generated and played according to the personal operation behavior data to guide the user.
  • the voice interaction program is further executed by the processor to implement the following steps:
  • the operation behavior data of each user on the line is acquired, and a corresponding voice help file is generated and played according to the operation behavior data to assist the user;
  • the voice command of the user is received within the second preset time, the voice command is identified, and a corresponding operation is performed according to the recognition result.
  • the voice interaction program is further executed by the processor to implement the following steps:
  • the voice interaction program is further executed by the processor to implement the following steps:
  • the present application further provides a computer readable storage medium, where the computer readable storage medium stores a voice interaction program, and when the voice interaction program is executed by the processor, the following steps are implemented:
  • the voice interaction function When receiving the user-triggered voice interaction function wake-up instruction, the voice interaction function is activated according to the voice interaction function wake-up instruction, and the preset voice response file is played to respond;
  • the personal operation behavior data of the user is acquired, and a corresponding voice guidance file is generated and played according to the personal operation behavior data to guide the user.
  • the present application provides a voice interaction method, apparatus, and computer readable storage medium.
  • the voice interaction function When receiving a user-triggered voice interaction function wake-up instruction, the voice interaction function is activated according to the voice interaction function wake-up instruction, and the preset voice response is played.
  • the file responds; monitors whether the user's voice instruction is received within the first preset time; if the user's voice instruction is not received within the first preset time, the user's personal operation behavior data is acquired, and according to the personal operation behavior The data generates a corresponding voice guidance file and plays it to guide the user.
  • the terminal of the present application when receiving the user-triggered voice interaction function wake-up instruction, starts the voice interaction function according to the voice interaction function wake-up instruction, and plays the preset voice response file to respond, and the application initiates the voice interaction function.
  • the interaction mode of the person can be simulated to respond promptly and actively, and the user experience effect can be improved; then, whether the user's voice command is received in the first preset time is received, and if the user is not received within the first preset time,
  • the voice command that is, when the user hesitates to perform the operation, the user's personal operation behavior data can be acquired at this time, and the corresponding voice guidance file is generated and played according to the personal operation behavior data to guide the user, compared with the prior art, the terminal
  • the feedback is only given when the voice command is received.
  • the application can perform statistical analysis according to the user's personal operation behavior data, thereby providing more humanized voice help and actively guiding the user. Therefore, the application can improve the voice interaction. Frequency, improve user experience during voice interaction
  • FIG. 1 is a schematic structural diagram of a terminal in a hardware operating environment according to an embodiment of the present application
  • FIG. 2 is a schematic flowchart of a first embodiment of a voice interaction method according to the present application
  • FIG. 3 is a schematic diagram of a process of refining a user's personal operation behavior data according to an embodiment of the present application, and generating a corresponding voice guidance file according to the personal operation behavior data and playing to guide the user;
  • FIG. 4 is a schematic flowchart of a second embodiment of a voice interaction method according to the present application.
  • FIG. 5 is a schematic diagram of a process of refining the user's operation behavior data on the line and generating a corresponding voice help file according to the operation behavior data and playing to assist the user in the embodiment.
  • the user when a user uses an intelligent voice product for communication, the user does not be as normal, timely, and available as the communication feedback between people. For example, after the user wakes up the smart voice product, the product does not take the initiative as a person. Communication, only when the user gives an order, the product will have feedback, the feedback is not timely and passive, and can not guide the user; when the product has not received the user instruction, it will directly end, or directly feedback product function help The feedback of the menu is not strong enough to really help the user. Therefore, in the voice interaction process of the existing intelligent voice product, the user is not provided with humanized voice help, and the user experience is poor.
  • the present application provides a voice interaction method, apparatus, and computer readable storage medium.
  • the voice interaction function is started according to the voice interaction function wake-up instruction. And playing a preset voice response file to respond; monitoring whether the user's voice instruction is received within the first preset time; if the user's voice instruction is not received within the first preset time, acquiring the user's personal operation behavior data, and Generating a corresponding voice guidance file according to the personal operation behavior data and playing to guide the user.
  • the terminal of the present application when receiving the user-triggered voice interaction function wake-up instruction, starts the voice interaction function according to the voice interaction function wake-up instruction, and plays the preset voice response file to respond, and the application initiates the voice interaction function.
  • the interaction mode of the person can be simulated to respond promptly and actively, and the user experience effect can be improved; then, whether the user's voice command is received in the first preset time is received, and if the user is not received within the first preset time,
  • the voice command that is, when the user hesitates to perform the operation, the user's personal operation behavior data can be acquired at this time, and the corresponding voice guidance file is generated and played according to the personal operation behavior data to guide the user, compared with the prior art, the terminal
  • the feedback is only given when the voice command is received.
  • the application can perform statistical analysis according to the user's personal operation behavior data, thereby providing more humanized voice help and actively guiding the user. Therefore, the application can improve the voice interaction. Frequency, improve user experience during voice interaction
  • FIG. 1 is a schematic structural diagram of a terminal in a hardware operating environment according to an embodiment of the present application.
  • the terminal in the embodiment of the present application may be a smart voice television, or an intelligent voice product such as an intelligent voice sound, an intelligent robot, a smart phone, a smart voice alarm clock, an intelligent voice navigator, and a voice air detecting machine.
  • an intelligent voice product such as an intelligent voice sound, an intelligent robot, a smart phone, a smart voice alarm clock, an intelligent voice navigator, and a voice air detecting machine.
  • the terminal may include a processor 1001, such as a CPU, a communication bus 1002, a user interface 1003, a network interface 1004, and a memory 1005.
  • the communication bus 1002 is used to implement connection communication between these components.
  • the user interface 1003 can include a display, an input unit such as a keyboard, and the optional user interface 1003 can also include a standard wired interface, a wireless interface.
  • the network interface 1004 can optionally include a standard wired interface, a wireless interface (such as a Wi-Fi interface).
  • the memory 1005 may be a high speed RAM memory or a stable memory (non-volatile) Memory), such as disk storage.
  • the memory 1005 can also optionally be a storage device independent of the aforementioned processor 1001.
  • the terminal may further include a camera, RF (Radio) Frequency, RF) circuits, sensors, audio circuits, Wi-Fi modules, and more.
  • sensors such as light sensors, motion sensors, and other sensors.
  • the light sensor may include an ambient light sensor and a proximity sensor, wherein the ambient light sensor may adjust the brightness of the display according to the brightness of the ambient light, and the proximity sensor may turn off the display and/or the backlight when the terminal moves to the ear. .
  • the gravity acceleration sensor can detect the magnitude of acceleration in each direction (usually three axes), and can detect the magnitude and direction of gravity when stationary, and can be used to identify the posture of the mobile terminal (such as horizontal and vertical screen switching, Related games, magnetometer attitude calibration), vibration recognition related functions (such as pedometer, tapping), etc.; of course, the terminal can also be equipped with other sensors such as gyroscope, barometer, hygrometer, thermometer, infrared sensor, etc. Let me repeat.
  • terminal structure shown in FIG. 1 does not constitute a limitation to the terminal, and may include more or less components than those illustrated, or a combination of certain components, or different component arrangements.
  • an operating system may be included in the memory 1005 as a computer storage medium.
  • a network communication module may be included in the memory 1005 as a computer storage medium.
  • a user interface module may be included in the memory 1005 as a computer storage medium.
  • a voice interaction program may be included in the memory 1005 as a computer storage medium.
  • the network interface 1004 is mainly used to connect to the background server for data communication with the background server;
  • the user interface 1003 is mainly used for connecting the client and performing data communication with the client; and
  • the processor 1001 can be used for
  • the voice interaction program stored in the memory 1005 is called and the following operations are performed:
  • the voice interaction function When receiving the user-triggered voice interaction function wake-up instruction, the voice interaction function is activated according to the voice interaction function wake-up instruction, and the preset voice response file is played to respond;
  • the personal operation behavior data of the user is acquired, and a corresponding voice guidance file is generated and played according to the personal operation behavior data to guide the user.
  • processor 1001 can call the voice interaction program stored in the memory 1005, and also performs the following operations:
  • the operation behavior data of each user on the line is acquired, and a corresponding voice help file is generated and played according to the operation behavior data to assist the user;
  • the voice command of the user is received within the second preset time, the voice command is identified, and a corresponding operation is performed according to the recognition result.
  • processor 1001 can call the voice interaction program stored in the memory 1005, and also performs the following operations:
  • processor 1001 can call the voice interaction program stored in the memory 1005, and also performs the following operations:
  • processor 1001 can call the voice interaction program stored in the memory 1005, and also performs the following operations:
  • the voice command of the user is received within the first preset time, the voice command is identified, and a corresponding operation is performed according to the recognition result.
  • the application provides a voice interaction method.
  • FIG. 2 is a schematic flowchart diagram of a voice interaction method according to a first embodiment of the present application.
  • the voice interaction method includes:
  • Step S10 when receiving the user-triggered voice interaction function wake-up instruction, start the voice interaction function according to the voice interaction function wake-up instruction, and play the preset voice response file to respond;
  • the voice interaction method can be used in a voice interaction process of an intelligent voice product, simulating a person-to-person interaction manner to respond promptly and actively, and performing statistical analysis according to a user's personal operation behavior, thereby providing more User-friendly voice help, proactively guide users, improve the frequency of voice interaction, and enhance the user experience during voice interaction.
  • the terminal in the embodiment of the present application may be a smart voice television, or an intelligent voice product such as an intelligent voice sound, an intelligent robot, a smart phone, a smart voice alarm clock, an intelligent voice navigator, and a voice air detecting machine.
  • the smart voice television is taken as an example for description in the embodiment of the present application.
  • the voice interaction function is started according to the voice interaction function wake-up instruction, and the preset voice response file is played to respond.
  • the manner in which the user triggers the voice interactive function wake-up command may include, but is not limited to, the following two types: 1) the user presses the power on button of the smart voice television, that is, the voice interactive function wake-up command is triggered when the smart voice television is turned on; 2) the user is in the smart The option to activate the voice interaction function is selected in the voice TV display interface.
  • the preset voice response file may be preset by the system or may be set by the user.
  • one or more voice response files may be stored in the voice response package.
  • the voice response package may include only one “activated voice interaction function, please command”.
  • the voice response file when the voice interaction function is activated, actively plays the voice response file; for example, the voice response package stores multiple voice response files, and one of them can be randomly called to play, and can also be based on current time information or other The information is selected to play the corresponding voice response information.
  • the voice response file includes “Master, happy weekend, what are you told me?” “Good evening, master, what do you have to order?” “Master, good morning, Please order” and wait for the voice response file, smart voice TV can choose to play according to the current time.
  • the voice interaction function When the voice interaction function is activated, the application can simulate the interaction mode of the person and the person to respond in a timely and active manner, thereby improving the user experience.
  • Step S20 Monitor whether a voice instruction of the user is received in the first preset time
  • step S30 is performed: acquiring personal operation behavior data of the user, and generating a corresponding voice guidance file according to the personal operation behavior data and playing to guide the user.
  • the smart voice television monitors whether the user's voice command is received within the first preset time. If the user's voice command is not received within the first preset time, the user's personal operation behavior data is acquired. And generating a corresponding voice guidance file according to the personal operation behavior data and playing to guide the user.
  • the first preset time is based on psychology and obtained through experiments, and can be set to 0.7s-1s.
  • the personal operation behavior data may include historical browsing records and browsing time, and then statistically analyze the user's personal operation behavior data. In addition, based on the statistical result of the user's personal operation behavior data, a more user-friendly voice guidance is provided, and the user is actively guided. Therefore, the application can improve the frequency of voice interaction and improve the user experience during the voice interaction process.
  • FIG. 3 is a schematic diagram of a process of refining the user's personal operation behavior data according to the embodiment of the present application, and generating a corresponding voice guidance file according to the personal operation behavior data and playing to guide the user.
  • Step S30 includes:
  • Step S31 acquiring personal operation behavior data of the user, and performing statistical analysis on the personal operation behavior data
  • Step S32 Generate a corresponding voice guidance file according to the statistical result and play it to guide the user.
  • the smart voice television terminal first obtains the personal operation behavior data of the user, and the personal operation behavior data may include a history browsing record and a browsing time, wherein the history browsing record may include the browsing type.
  • TV programs, TV series, movies, variety shows, etc. may also include program types, such as television programs including news, finance, sports, etc.
  • TV dramas include American dramas, Korean dramas, or love dramas, costume dramas, suspense dramas, etc.
  • the corresponding hosting or starring list may be included, and the browsing time may include a working day, a non-working day, an early, a middle, a late, etc., and then statistically analyze the personal operation behavior data, and generate a corresponding voice guidance file according to the statistical result. Play to guide the user. For example, statistical analysis of personal operational behavior data reveals that the user has been watching a certain TV series A and B recently. At this time, according to the statistical result, a voice guidance file "Master, do you want to continue watching A or B today", or Statistical analysis found that the user recently watched a movie starring someone C. At this time, the voice guidance file "master, C, starring movies, such as D, E, and F, you want to watch,” can be generated based on the statistical result.
  • the statistical result may be combined with the current time to generate a corresponding voice guidance file.
  • statistical analysis of the personal operation behavior data indicates that the user watches the news broadcast every night from 7:00 to 7:30. If the current time is between 7:00 and 7:30 or close to 7 o'clock, at this time, the voice guidance file "News simulcasting/coming soon, you need to watch it" can be generated according to the statistical result and the current time. Guided voice is formed based on the user's personal operational behavior data, which can give users more intimate and humanized help and service. And help guide this type of speech is actively occurring, you can improve the user's voice interaction frequency, allowing users to easily and naturally help enhance the user experience.
  • the voice interaction method may further include:
  • the voice command of the user is received within the first preset time, the voice command is identified, and a corresponding operation is performed according to the recognition result.
  • the smart voice television When the smart voice television receives the voice command of the user within the first preset time, the voice command is recognized, and the corresponding operation is performed according to the recognition result.
  • the specific identification technology can refer to the prior art, and details are not described herein. For example, if the user's voice command “play TV drama A” is received within the first preset time, the collected user voice may be identified, and then the television interface is controlled to jump to the episode selection interface of the TV drama A according to the recognition result; For another example, when the user's voice command “Play Movie D” is received within the first preset time, the TV interface is directly controlled to start playing the movie D after the recognition.
  • the present application provides a voice interaction method, when receiving a user-triggered voice interaction function wake-up instruction, starting a voice interaction function according to the voice interaction function wake-up instruction, and playing a preset voice response file to respond; monitoring the first pre- Whether the voice instruction of the user is received within the set time; if the voice instruction of the user is not received within the first preset time, the personal operation behavior data of the user is acquired, and the corresponding voice guidance file is generated according to the personal operation behavior data. Play to guide the user.
  • the terminal of the present application starts the voice interaction function according to the voice interaction function wake-up instruction, and plays the preset voice response file to respond, and the application initiates the voice interaction function.
  • the interaction mode of the person can be simulated to respond promptly and actively, and the user experience effect can be improved; then, whether the user's voice command is received in the first preset time is received, and if the user is not received within the first preset time,
  • the voice command that is, when the user hesitates to perform the operation, the user's personal operation behavior data can be acquired at this time, and the corresponding voice guidance file is generated and played according to the personal operation behavior data to guide the user, compared with the prior art, the terminal
  • the feedback is only given when the voice command is received.
  • the application can perform statistical analysis according to the user's personal operation behavior data, thereby providing more humanized voice help and actively guiding the user. Therefore, the application can improve the voice interaction. Frequency, improve user experience during voice interaction
  • FIG. 4 is a schematic flowchart diagram of a second embodiment of a voice interaction method according to the present application.
  • the voice interaction method further includes:
  • Step S40 Monitor whether a voice instruction of the user is received in the second preset time
  • Step S51 If the voice instruction of the user is not received within the second preset time, the operation behavior data of each user on the line is acquired, and a corresponding voice help file is generated and played according to the operation behavior data to assist the user;
  • the terminal continues to monitor whether the voice instruction of the user is received in the second preset time. If the voice command of the user is not received within the second preset time, the terminal guides the voice. It may not bring real help to the user, or the user may have no purpose.
  • the operation behavior data of each user on the line is obtained, and a corresponding voice help file is generated and played according to the operation behavior data to help the user.
  • the second preset time is based on psychology and obtained through experiments, and can be set to 2s-3s.
  • the operational behavior data can include video viewing records, web search records, and web pages within a certain time range (eg, within a month). Browse records, etc., and then perform statistical analysis on the operational behavior data of each user on the line, and then provide voice help based on the statistical results of the online user behavior data, thereby providing users with reference opinions and improving the user experience.
  • FIG. 5 is a schematic diagram of the process of refining the user's operation behavior data on the line and generating a corresponding voice help file according to the operation behavior data and playing to assist the user in the process of refinement.
  • Step S51 may include:
  • Step S511 acquiring operation behavior data of each user on the line, and performing statistical analysis on the operation behavior data;
  • Step S512 Generate a corresponding voice help file according to the statistical result and play it to help the user.
  • the smart voice television terminal first obtains the operation behavior data of each user on the line, and the operation behavior data of each user on the line.
  • the video viewing record, the network search record, and the web browsing record may be included in a certain time range (such as in the past one month), and then statistically analyzed the operation behavior data, and corresponding voice help files are generated and played according to the statistical result, thereby Provide users with a reference to help users make choices. For example, statistical analysis of the operational behavior data of each user on the line finds that most users have recently watched a certain movie X or searched for relevant information of the movie X. At this time, a voice help file can be generated based on the statistical result.
  • the selected voice service can provide more user-friendly help to users without purpose, and this kind of help voice is active, and can also improve the user's voice interaction frequency, and provide users with reference choices, which can be improved. user experience.
  • Step S52 If the voice command of the user is received in the second preset time, the voice command is identified, and the corresponding operation is performed according to the recognition result.
  • the smart voice television receives the voice command of the user within the second preset time, the voice command is recognized, and the corresponding operation is performed according to the recognition result.
  • the specific identification technology can refer to the prior art, and details are not described herein.
  • the present application also provides a voice interaction device including a memory, a processor, and a voice interaction program stored on the memory and operable on the processor, the voice interaction program being the processor.
  • a voice interaction device including a memory, a processor, and a voice interaction program stored on the memory and operable on the processor, the voice interaction program being the processor.
  • the specific embodiment of the voice interaction device of the present application is basically the same as the foregoing embodiments of the voice interaction method, and details are not described herein.
  • the present application further provides a computer readable storage medium having a voice interaction program stored thereon, the voice interaction program being executed by a processor to implement the voice interaction method according to any of the above embodiments. step.
  • the technical solution of the present application which is essential or contributes to the prior art, may be embodied in the form of a software product stored in a storage medium (such as ROM/RAM as described above). , a disk, an optical disk, including a number of instructions for causing a terminal device (which may be a mobile phone, a computer, a server, an air conditioner, or a network device, etc.) to perform the methods described in the various embodiments of the present application.
  • a terminal device which may be a mobile phone, a computer, a server, an air conditioner, or a network device, etc.

Abstract

本申请公开了一种语音交互方法。该语音交互方法包括:当接收到用户触发的语音交互功能唤醒指令时,根据所述语音交互功能唤醒指令启动语音交互功能,并播放预设语音回应文件进行回应;监测第一预设时间内是否接收到用户的语音指令;若第一预设时间内未接收到用户的语音指令,则获取用户的个人操作行为数据,并根据所述个人操作行为数据生成对应的语音引导文件并播放,以引导用户。本申请还公开了一种语音交互装置及计算机可读存储介质。本申请通过用户的个人操作行为数据生成对应的语音引导文件来引导用户,可给用户提供更人性化的语音帮助,从而提升语音交互过程中的用户体验感。

Description

语音交互方法、装置及计算机可读存储介质
本申请要求于2017年12月20日提交中国专利局、申请号为201711382608.5、申请名称为“语音交互方法、装置及计算机可读存储介质”的中国专利申请的优先权,其全部内容通过引用结合在申请中。
技术领域
本申请涉及通讯领域,尤其涉及一种语音交互方法、装置及计算机可读存储介质。
背景技术
随着语音技术的进步及互联网大数据的不断成熟,智能语音已经成为人工智能进入日常生活的第一条通路,各大商家纷纷推出智能语音产品,例如智能语音电视、智能语音音响、智能语音导航仪和语音空气检测机等。智能语音产品通过模拟人与人的语音方式,使用语音替代传统的操作方式,给消费者带来了极大的方便。
但用户在使用智能语音产品进行交流时,并不能像人与人之间的交流反馈一样正常、及时、可用,例如,用户唤醒智能语音产品后,产品不会像人一样主动交流,只有等到用户发出指令时,产品才有反馈,该反馈不及时且较为被动,无法对用户起到指引作用;当产品一直未收到用户指令时,会直接结束,或直接反馈产品功能帮助菜单,该反馈可用性不强,无法真正帮助到用户,因此,在现有智能语音产品的语音交互过程中,未给用户提供人性化的语音帮助,用户体验感较差。
申请内容
本申请的主要目的在于提供一种语音交互方法、装置及计算机可读存储介质,旨在提供更人性化的语音帮助,提升语音交互过程中的用户体验感。
为实现上述目的,本申请提供一种语音交互方法,所述语音交互方法包括以下步骤:
当接收到用户触发的语音交互功能唤醒指令时,根据所述语音交互功能唤醒指令启动语音交互功能,并播放预设语音回应文件进行回应;
监测第一预设时间内是否接收到用户的语音指令;
若第一预设时间内未接收到用户的语音指令,则获取用户的个人操作行为数据,并根据所述个人操作行为数据生成对应的语音引导文件并播放,以引导用户。
可选地,所述语音交互方法还包括:
监测第二预设时间内是否接收到用户的语音指令;
若第二预设时间内未接收到用户的语音指令,则获取线上各用户的操作行为数据,并根据所述操作行为数据生成对应的语音帮助文件并播放,以帮助用户;
若第二预设时间内接收到用户的语音指令,则对所述语音指令进行识别,并根据识别结果执行对应的操作。
可选地,所述获取线上各用户的操作行为数据,并根据所述操作行为数据生成对应的语音帮助文件并播放,以帮助用户的步骤,包括:
获取线上各用户的操作行为数据,并对所述操作行为数据进行统计分析;
根据统计结果生成对应的语音帮助文件并播放,以帮助用户。
可选地,所述获取用户的个人操作行为数据,并根据所述个人操作行为数据生成对应的语音引导文件并播放,以引导用户的步骤,包括:
获取用户的个人操作行为数据,并对所述个人操作行为数据进行统计分析;
根据统计结果生成对应的语音引导文件并播放,以引导用户。
可选地,所述监测第一预设时间内是否接收到用户的语音指令的步骤之后,包括:
若第一预设时间内接收到用户的语音指令,则对所述语音指令进行识别,并根据识别结果执行对应的操作。
此外,为实现上述目的,本申请还提供一种语音交互装置,所述语音交互装置包括:存储器、处理器及存储在所述存储器上并可在所述处理器上运行的语音交互程序,所述语音交互程序被所述处理器执行时实现以下步骤:
当接收到用户触发的语音交互功能唤醒指令时,根据所述语音交互功能唤醒指令启动语音交互功能,并播放预设语音回应文件进行回应;
监测第一预设时间内是否接收到用户的语音指令;
若第一预设时间内未接收到用户的语音指令,则获取用户的个人操作行为数据,并根据所述个人操作行为数据生成对应的语音引导文件并播放,以引导用户。
可选地,所述语音交互程序被所述处理器执行时还实现以下步骤:
监测第二预设时间内是否接收到用户的语音指令;
若第二预设时间内未接收到用户的语音指令,则获取线上各用户的操作行为数据,并根据所述操作行为数据生成对应的语音帮助文件并播放,以帮助用户;
若第二预设时间内接收到用户的语音指令,则对所述语音指令进行识别,并根据识别结果执行对应的操作。
可选地,所述语音交互程序被所述处理器执行时还实现以下步骤:
获取线上各用户的操作行为数据,并对所述操作行为数据进行统计分析;
根据统计结果生成对应的语音帮助文件并播放,以帮助用户。
可选地,所述语音交互程序被所述处理器执行时还实现以下步骤:
获取用户的个人操作行为数据,并对所述个人操作行为数据进行统计分析;
根据统计结果生成对应的语音引导文件并播放,以引导用户。
此外,为实现上述目的,本申请还提供一种计算机可读存储介质,所述计算机可读存储介质上存储有语音交互程序,所述语音交互程序被处理器执行时实现以下步骤:
当接收到用户触发的语音交互功能唤醒指令时,根据所述语音交互功能唤醒指令启动语音交互功能,并播放预设语音回应文件进行回应;
监测第一预设时间内是否接收到用户的语音指令;
若第一预设时间内未接收到用户的语音指令,则获取用户的个人操作行为数据,并根据所述个人操作行为数据生成对应的语音引导文件并播放,以引导用户。
本申请提供一种语音交互方法、装置及计算机可读存储介质,通过当接收到用户触发的语音交互功能唤醒指令时,根据所述语音交互功能唤醒指令启动语音交互功能,并播放预设语音回应文件进行回应;监测第一预设时间内是否接收到用户的语音指令;若第一预设时间内未接收到用户的语音指令,则获取用户的个人操作行为数据,并根据所述个人操作行为数据生成对应的语音引导文件并播放,以引导用户。通过上述方式,本申请终端在接收到用户触发的语音交互功能唤醒指令时,根据所述语音交互功能唤醒指令启动语音交互功能,并播放预设语音回应文件进行回应,本申请在启动语音交互功能时,可模拟人与人的互动方式及时主动地进行回应,可提高用户体验效果;然后监测第一预设时间内是否接收到用户的语音指令,若第一预设时间内未接收到用户的语音指令,即用户犹豫进行什么操作时,此时可获取用户的个人操作行为数据,并根据个人操作行为数据生成对应的语音引导文件并播放,以引导用户,相比于现有技术中,终端只有在接收到语音指令时才给予反馈回应,本申请可根据用户的个人操作行为数据进行统计分析,进而提供更人性化的语音帮助,主动对用户进行引导,因此,本申请可提高语音交互的频率,提升语音交互过程中的用户体验感。
附图说明
图1为本申请实施例方案涉及的硬件运行环境的终端结构示意图;
图2为本申请语音交互方法第一实施例的流程示意图;
图3为本申请实施例中获取用户的个人操作行为数据,并根据所述个人操作行为数据生成对应的语音引导文件并播放,以引导用户的细化流程示意图;
图4为本申请语音交互方法第二实施例的流程示意图;
图5为本申请实施例中获取线上各用户的操作行为数据,并根据所述操作行为数据生成对应的语音帮助文件并播放,以帮助用户的细化流程示意图。
本申请目的的实现、功能特点及优点将结合实施例,参照附图做进一步说明。
具体实施方式
应当理解,此处所描述的具体实施例仅仅用以解释本申请,并不用于限定本申请。
在现有技术中,用户在使用智能语音产品进行交流时,并不能像人与人之间的交流反馈一样正常、及时、可用,例如,用户唤醒智能语音产品后,产品不会像人一样主动交流,只有等到用户发出指令时,产品才有反馈,该反馈不及时且较为被动,无法对用户起到指引作用;当产品一直未收到用户指令时,会直接结束,或直接反馈产品功能帮助菜单,该反馈可用性不强,无法真正帮助到用户,因此,在现有智能语音产品的语音交互过程中,未给用户提供人性化的语音帮助,用户体验感较差。
为了解决上述技术问题,本申请提供一种语音交互方法、装置及计算机可读存储介质,通过当接收到用户触发的语音交互功能唤醒指令时,根据所述语音交互功能唤醒指令启动语音交互功能,并播放预设语音回应文件进行回应;监测第一预设时间内是否接收到用户的语音指令;若第一预设时间内未接收到用户的语音指令,则获取用户的个人操作行为数据,并根据所述个人操作行为数据生成对应的语音引导文件并播放,以引导用户。通过上述方式,本申请终端在接收到用户触发的语音交互功能唤醒指令时,根据所述语音交互功能唤醒指令启动语音交互功能,并播放预设语音回应文件进行回应,本申请在启动语音交互功能时,可模拟人与人的互动方式及时主动地进行回应,可提高用户体验效果;然后监测第一预设时间内是否接收到用户的语音指令,若第一预设时间内未接收到用户的语音指令,即用户犹豫进行什么操作时,此时可获取用户的个人操作行为数据,并根据个人操作行为数据生成对应的语音引导文件并播放,以引导用户,相比于现有技术中,终端只有在接收到语音指令时才给予反馈回应,本申请可根据用户的个人操作行为数据进行统计分析,进而提供更人性化的语音帮助,主动对用户进行引导,因此,本申请可提高语音交互的频率,提升语音交互过程中的用户体验感。
请参阅图1,图1为本申请实施例方案涉及的硬件运行环境的终端结构示意图。
本申请实施例终端可以是智能语音电视,也可以是智能语音音响、智能机器人、智能手机、智能语音闹钟、智能语音导航仪和语音空气检测机等智能语音产品。
如图1所示,该终端可以包括:处理器1001,例如CPU,通信总线1002,用户接口1003,网络接口1004,存储器1005。其中,通信总线1002用于实现这些组件之间的连接通信。用户接口1003可以包括显示屏(Display)、输入单元比如键盘(Keyboard),可选用户接口1003还可以包括标准的有线接口、无线接口。网络接口1004可选的可以包括标准的有线接口、无线接口(如Wi-Fi接口)。存储器1005可以是高速RAM存储器,也可以是稳定的存储器(non-volatile memory),例如磁盘存储器。存储器1005可选的还可以是独立于前述处理器1001的存储装置。
可选地,终端还可以包括摄像头、RF(Radio Frequency,射频)电路,传感器、音频电路、Wi-Fi模块等等。其中,传感器比如光传感器、运动传感器以及其他传感器。具体地,光传感器可包括环境光传感器及接近传感器,其中,环境光传感器可根据环境光线的明暗来调节显示屏的亮度,接近传感器可在终端移动到耳边时,关闭显示屏和/或背光。作为运动传感器的一种,重力加速度传感器可检测各个方向上(一般为三轴)加速度的大小,静止时可检测出重力的大小及方向,可用于识别移动终端姿态的应用(比如横竖屏切换、相关游戏、磁力计姿态校准)、振动识别相关功能(比如计步器、敲击)等;当然,终端还可配置陀螺仪、气压计、湿度计、温度计、红外线传感器等其他传感器,在此不再赘述。
本领域技术人员可以理解,图1中示出的终端结构并不构成对终端的限定,可以包括比图示更多或更少的部件,或者组合某些部件,或者不同的部件布置。
如图1所示,作为一种计算机存储介质的存储器1005中可以包括操作系统、网络通信模块、用户接口模块以及语音交互程序。
在图1所示的终端中,网络接口1004主要用于连接后台服务器,与后台服务器进行数据通信;用户接口1003主要用于连接客户端,与客户端进行数据通信;而处理器1001可以用于调用存储器1005中存储的语音交互程序,并执行以下操作:
当接收到用户触发的语音交互功能唤醒指令时,根据所述语音交互功能唤醒指令启动语音交互功能,并播放预设语音回应文件进行回应;
监测第一预设时间内是否接收到用户的语音指令;
若第一预设时间内未接收到用户的语音指令,则获取用户的个人操作行为数据,并根据所述个人操作行为数据生成对应的语音引导文件并播放,以引导用户。
进一步地,处理器1001可以调用存储器1005中存储的语音交互程序,还执行以下操作:
监测第二预设时间内是否接收到用户的语音指令;
若第二预设时间内未接收到用户的语音指令,则获取线上各用户的操作行为数据,并根据所述操作行为数据生成对应的语音帮助文件并播放,以帮助用户;
若第二预设时间内接收到用户的语音指令,则对所述语音指令进行识别,并根据识别结果执行对应的操作。
进一步地,处理器1001可以调用存储器1005中存储的语音交互程序,还执行以下操作:
获取线上各用户的操作行为数据,并对所述操作行为数据进行统计分析;
根据统计结果生成对应的语音帮助文件并播放,以帮助用户。
进一步地,处理器1001可以调用存储器1005中存储的语音交互程序,还执行以下操作:
获取用户的个人操作行为数据,并对所述个人操作行为数据进行统计分析;
根据统计结果生成对应的语音引导文件并播放,以引导用户。
进一步地,处理器1001可以调用存储器1005中存储的语音交互程序,还执行以下操作:
若第一预设时间内接收到用户的语音指令,则对所述语音指令进行识别,并根据识别结果执行对应的操作。
基于上述硬件结构,提出本申请语音交互方法各实施例。
本申请提供一种语音交互方法。
请参阅图2,图2为本申请语音交互方法第一实施例的流程示意图。
在本申请实施例中,该语音交互方法包括:
步骤S10,当接收到用户触发的语音交互功能唤醒指令时,根据所述语音交互功能唤醒指令启动语音交互功能,并播放预设语音回应文件进行回应;
在本申请实施例中,该语音交互方法可用于智能语音产品的语音交互过程中,模拟人与人的互动方式及时主动地进行回应,并且可根据用户的个人操作行为进行统计分析,进而提供更人性化的语音帮助,主动对用户进行引导帮助,提高语音交互的频率,提升语音交互过程中的用户体验感。本申请实施例终端可以是智能语音电视,也可以是智能语音音响、智能机器人、智能手机、智能语音闹钟、智能语音导航仪和语音空气检测机等智能语音产品。为方便说明,本申请实施例中以智能语音电视为例进行说明。
在本申请实施例中,当智能语音电视接收到用户触发的语音交互功能唤醒指令时,根据该语音交互功能唤醒指令启动语音交互功能,并播放预设语音回应文件进行回应。其中,用户触发语音交互功能唤醒指令的方式可以包括但不限于以下2种:1)用户按压智能语音电视的开机按钮,即开启智能语音电视时即触发语音交互功能唤醒指令;2)用户在智能语音电视显示界面中选择启动语音交互功能的选项。
需要说明的是,预设语音回应文件可以系统预先设定的,也可以是用户自己设定的。当预设语音回应文件为系统预先设定的时,可以在语音回应包中存储一个或多个语音回应文件,例如,该语音回应包可以只包括一个“已启动语音交互功能,请吩咐”的语音回应文件,当启动语音交互功能,主动播放该条语音回应文件;再例如,该语音回应包中存储有多个语音回应文件,可以随机调用其中一个进行播放,还可以根据当前时间信息或其他信息选取对应的语音回应信息进行播放,例如,该语音回应文件包括“主人,周末愉快,请问您有什么吩咐呢”,“晚上好,主人,您有什么吩咐呢”,“主人,早上好,请吩咐”等语音回应文件,智能语音电视可以根据当前时间进行选择播放。本申请在启动语音交互功能时,可模拟人与人的互动方式及时主动地进行回应,可提高用户体验效果。
步骤S20,监测第一预设时间内是否接收到用户的语音指令;
若第一预设时间内未接收到用户的语音指令,则执行步骤S30:获取用户的个人操作行为数据,并根据所述个人操作行为数据生成对应的语音引导文件并播放,以引导用户。
当启动语音交互功能之后,智能语音电视会监测第一预设时间内是否接收到用户的语音指令,若第一预设时间内未接收到用户的语音指令,则获取用户的个人操作行为数据,并根据所述个人操作行为数据生成对应的语音引导文件并播放,以引导用户。其中,第一预设时间是基于心理学并通过实验获得的,可以设为0.7s-1s,个人操作行为数据可以包括历史浏览记录和浏览时间,然后对用户的个人操作行为数据进行统计分析,进而基于用户个人操作行为数据的统计结果提供更人性化的语音引导,主动对用户进行引导,因此,本申请可提高语音交互的频率,提升语音交互过程中的用户体验感。
具体的,请参阅图3,图3为本申请实施例中获取用户的个人操作行为数据,并根据所述个人操作行为数据生成对应的语音引导文件并播放,以引导用户的细化流程示意图。步骤S30包括:
步骤S31,获取用户的个人操作行为数据,并对所述个人操作行为数据进行统计分析;
步骤S32,根据统计结果生成对应的语音引导文件并播放,以引导用户。
若第一预设时间内未接收到用户的语音指令,智能语音电视终端会先获取用户的个人操作行为数据,个人操作行为数据可以包括历史浏览记录和浏览时间,其中历史浏览记录可以包括浏览类型,例如电视节目、电视剧、电影、综艺等,还可以包括节目类型,例如电视节目包括新闻类、财经类、体育类等,电视剧包括美剧、韩剧,或者爱情剧、古装剧、悬疑剧等,还可以包括对应的主持或者主演名单,浏览时间可以包括工作日、非工作日,早、中、晚等分类,然后对这些个人操作行为数据进行统计分析,并根据统计结果生成对应的语音引导文件并播放,以引导用户。例如,对个人操作行为数据进行统计分析发现用户最近连续在看某电视剧A和B,此时可以根据该统计结果生成语音引导文件“主人,您今天想继续观看A或B吗”,或者,经统计分析发现用户最近连续看某人C主演的电影,此时可以根据该统计结果生成语音引导文件“主人,C主演的电影有很多不错的,例如D、E和F,您想观看吗”,此外,在具体实施例中,还可以将统计结果与当前时间相结合生成对应的语音引导文件,例如,对个人操作行为数据进行统计分析发现用户每天晚上7点至7点半都会观看新闻联播,如果获取到当前时间在7点至7点半之间或者接近7点,此时,可以根据该统计结果和当前时间生成语音引导文件“新闻联播正在/即将播出,您需要观看吗”,这种引导式的语音是基于用户的个人操作行为数据形成的,能给用户更贴心、人性化的帮助和服务,而且这种引导式的帮助语音是主动发生的,可以提高用户的语音交互频率,让用户轻松而自然的获得帮助,提升用户体验。
此外,在本申请实施例中,在上述步骤S20之后,该语音交互方法还可以包括:
若第一预设时间内接收到用户的语音指令,则对所述语音指令进行识别,并根据识别结果执行对应的操作。
当智能语音电视在第一预设时间内接收到用户的语音指令,则对该语音指令进行识别,并根据识别结果执行对应的操作。其中,具体的识别技术可参照现有技术,此处不做赘述。例如,在第一预设时间内接收到用户的语音指令“播放电视剧A”,则可以对采集到的用户语音进行识别,然后根据识别结果控制电视界面跳转到电视剧A的剧集选择界面;再例如,在第一预设时间内接收到用户的语音指令“播放电影D”,则在识别后直接控制电视界面开始播放电影D。
本申请提供一种语音交互方法,通过当接收到用户触发的语音交互功能唤醒指令时,根据所述语音交互功能唤醒指令启动语音交互功能,并播放预设语音回应文件进行回应;监测第一预设时间内是否接收到用户的语音指令;若第一预设时间内未接收到用户的语音指令,则获取用户的个人操作行为数据,并根据所述个人操作行为数据生成对应的语音引导文件并播放,以引导用户。通过上述方式,本申请终端在接收到用户触发的语音交互功能唤醒指令时,根据所述语音交互功能唤醒指令启动语音交互功能,并播放预设语音回应文件进行回应,本申请在启动语音交互功能时,可模拟人与人的互动方式及时主动地进行回应,可提高用户体验效果;然后监测第一预设时间内是否接收到用户的语音指令,若第一预设时间内未接收到用户的语音指令,即用户犹豫进行什么操作时,此时可获取用户的个人操作行为数据,并根据个人操作行为数据生成对应的语音引导文件并播放,以引导用户,相比于现有技术中,终端只有在接收到语音指令时才给予反馈回应,本申请可根据用户的个人操作行为数据进行统计分析,进而提供更人性化的语音帮助,主动对用户进行引导,因此,本申请可提高语音交互的频率,提升语音交互过程中的用户体验感。
请参阅图4,图4为本申请语音交互方法第二实施例的流程示意图。
基于图2所示的第一实施例,在步骤S30之后,该语音交互方法还包括:
步骤S40,监测第二预设时间内是否接收到用户的语音指令;
步骤S51,若第二预设时间内未接收到用户的语音指令,则获取线上各用户的操作行为数据,并根据所述操作行为数据生成对应的语音帮助文件并播放,以帮助用户;
在本申请实施例中,终端在对用户进行引导之后,继续监测第二预设时间内是否接收到用户的语音指令,若第二预设时间内未接收到用户的语音指令,则该引导语音可能未带给用户真正的帮助,或用户可能没有目的性,此时,则获取线上各用户的操作行为数据,并根据该操作行为数据生成对应的语音帮助文件并播放,以帮助用户。其中,第二预设时间是基于心理学并通过实验获得的,可以设为2s-3s,操作行为数据可以包括一定时间范围内(如近一个月内)的视频观看记录、网络搜索记录和网页浏览记录等,然后对线上各用户的操作行为数据进行统计分析,进而基于线上各用户操作行为数据的统计结果提供语音帮助,从而给用户提供参考意见,提升用户体验。
具体的,请参阅图5,图5为本申请实施例中获取线上各用户的操作行为数据,并根据所述操作行为数据生成对应的语音帮助文件并播放,以帮助用户的细化流程示意图。步骤S51可以包括:
步骤S511,获取线上各用户的操作行为数据,并对所述操作行为数据进行统计分析;
步骤S512,根据统计结果生成对应的语音帮助文件并播放,以帮助用户。
若第二预设时间内未接收到用户的语音指令,则说明用户可能没有目的性,此时,智能语音电视终端会先获取线上各用户的操作行为数据,线上各用户的操作行为数据可以包括一定时间范围内(如近一个月内)的视频观看记录、网络搜索记录和网页浏览记录等,然后对这些操作行为数据进行统计分析,根据统计结果生成对应的语音帮助文件并播放,从而给用户提供参考意见,以帮助用户进行选择。例如,对线上各用户的操作行为数据进行统计分析发现大多数用户最近都在观看某一电影X或搜索该电影X的相关信息,此时,可以根据该统计结果生成语音帮助文件“最近很热门的电影X,您想观看吗”,或者,可以根据统计结果生成一个选择性的语音帮助文件“最近的新片有a、b、c,有没有您感兴趣的呢”,这种帮助提供参考、选择的语音服务,可以给没有目的性的用户提供更人性化的帮助,而且这种帮助式的语音是主动发生的,同样也可以提高用户的语音交互频率,给用户提供参考选择,可提升用户体验。
步骤S52,若第二预设时间内接收到用户的语音指令,则对所述语音指令进行识别,并根据识别结果执行对应的操作。
当智能语音电视在第二预设时间内接收到用户的语音指令,则对该语音指令进行识别,并根据识别结果执行对应的操作。其中,具体的识别技术可参照现有技术,此处不做赘述。
本申请还提供一种语音交互装置,该语音交互装置包括存储器、处理器及存储在所述存储器上并可在所述处理器上运行的语音交互程序,所述语音交互程序被所述处理器执行时实现如以上任一项实施例所述的语音交互方法的步骤。
本申请语音交互装置的具体实施例与上述语音交互方法各实施例基本相同,在此不作赘述。
本申请还提供一种计算机可读存储介质,该计算机可读存储介质上存储有语音交互程序,所述语音交互程序被处理器执行时实现如以上任一项实施例所述的语音交互方法的步骤。
本申请计算机可读存储介质的具体实施例与上述语音交互方法各实施例基本相同,在此不作赘述。
需要说明的是,在本文中,术语“包括”、“包含”或者其任何其他变体意在涵盖非排他性的包含,从而使得包括一系列要素的过程、方法、物品或者系统不仅包括那些要素,而且还包括没有明确列出的其他要素,或者是还包括为这种过程、方法、物品或者系统所固有的要素。在没有更多限制的情况下,由语句“包括一个……”限定的要素,并不排除在包括该要素的过程、方法、物品或者系统中还存在另外的相同要素。
上述本申请实施例序号仅仅为了描述,不代表实施例的优劣。
通过以上的实施方式的描述,本领域的技术人员可以清楚地了解到上述实施例方法可借助软件加必需的通用硬件平台的方式来实现,当然也可以通过硬件,但很多情况下前者是更佳的实施方式。基于这样的理解,本申请的技术方案本质上或者说对现有技术做出贡献的部分可以以软件产品的形式体现出来,该计算机软件产品存储在如上所述的一个存储介质(如ROM/RAM、磁碟、光盘)中,包括若干指令用以使得一台终端设备(可以是手机,计算机,服务器,空调器,或者网络设备等)执行本申请各个实施例所述的方法。
以上仅为本申请的优选实施例,并非因此限制本申请的专利范围,凡是利用本申请说明书及附图内容所作的等效结构或等效流程变换,或直接或间接运用在其他相关的技术领域,均同理包括在本申请的专利保护范围内。

Claims (15)

  1. 一种语音交互方法,其中,所述语音交互方法包括以下步骤:
    当接收到用户触发的语音交互功能唤醒指令时,根据所述语音交互功能唤醒指令启动语音交互功能,并播放预设语音回应文件进行回应;
    监测第一预设时间内是否接收到用户的语音指令;
    若第一预设时间内未接收到用户的语音指令,则获取用户的个人操作行为数据,并根据所述个人操作行为数据生成对应的语音引导文件并播放,以引导用户。
  2. 如权利要求1所述的语音交互方法,其中,所述语音交互方法还包括:
    监测第二预设时间内是否接收到用户的语音指令;
    若第二预设时间内未接收到用户的语音指令,则获取线上各用户的操作行为数据,并根据所述操作行为数据生成对应的语音帮助文件并播放,以帮助用户;
    若第二预设时间内接收到用户的语音指令,则对所述语音指令进行识别,并根据识别结果执行对应的操作。
  3. 如权利要求2所述的语音交互方法,其中,所述获取线上各用户的操作行为数据,并根据所述操作行为数据生成对应的语音帮助文件并播放,以帮助用户的步骤,包括:
    获取线上各用户的操作行为数据,并对所述操作行为数据进行统计分析;
    根据统计结果生成对应的语音帮助文件并播放,以帮助用户。
  4. 如权利要求1所述的语音交互方法,其中,所述获取用户的个人操作行为数据,并根据所述个人操作行为数据生成对应的语音引导文件并播放,以引导用户的步骤,包括:
    获取用户的个人操作行为数据,并对所述个人操作行为数据进行统计分析;
    根据统计结果生成对应的语音引导文件并播放,以引导用户。
  5. 如权利要求1所述的语音交互方法,其中,所述监测第一预设时间内是否接收到用户的语音指令的步骤之后,包括:
    若第一预设时间内接收到用户的语音指令,则对所述语音指令进行识别,并根据识别结果执行对应的操作。
  6. 一种语音交互装置,其中,所述语音交互装置包括:存储器、处理器及存储在所述存储器上并可在所述处理器上运行的语音交互程序,所述语音交互程序被所述处理器执行时实现以下步骤:
    当接收到用户触发的语音交互功能唤醒指令时,根据所述语音交互功能唤醒指令启动语音交互功能,并播放预设语音回应文件进行回应;
    监测第一预设时间内是否接收到用户的语音指令;
    若第一预设时间内未接收到用户的语音指令,则获取用户的个人操作行为数据,并根据所述个人操作行为数据生成对应的语音引导文件并播放,以引导用户。
  7. 如权利要求6所述的语音交互装置,其中,所述语音交互程序被所述处理器执行时还实现以下步骤:
    监测第二预设时间内是否接收到用户的语音指令;
    若第二预设时间内未接收到用户的语音指令,则获取线上各用户的操作行为数据,并根据所述操作行为数据生成对应的语音帮助文件并播放,以帮助用户;
    若第二预设时间内接收到用户的语音指令,则对所述语音指令进行识别,并根据识别结果执行对应的操作。
  8. 如权利要求7所述的语音交互装置,其中,所述语音交互程序被所述处理器执行时还实现以下步骤:
    获取线上各用户的操作行为数据,并对所述操作行为数据进行统计分析;
    根据统计结果生成对应的语音帮助文件并播放,以帮助用户。
  9. 如权利要求6所述的语音交互装置,其中,所述语音交互程序被所述处理器执行时还实现以下步骤:
    获取用户的个人操作行为数据,并对所述个人操作行为数据进行统计分析;
    根据统计结果生成对应的语音引导文件并播放,以引导用户。
  10. 如权利要求6所述的语音交互装置,其中,所述语音交互程序被所述处理器执行时还实现以下步骤:
    若第一预设时间内接收到用户的语音指令,则对所述语音指令进行识别,并根据识别结果执行对应的操作。
  11. 一种计算机可读存储介质,其中,所述计算机可读存储介质上存储有语音交互程序,所述语音交互程序被处理器执行时实现以下步骤:
    当接收到用户触发的语音交互功能唤醒指令时,根据所述语音交互功能唤醒指令启动语音交互功能,并播放预设语音回应文件进行回应;
    监测第一预设时间内是否接收到用户的语音指令;
    若第一预设时间内未接收到用户的语音指令,则获取用户的个人操作行为数据,并根据所述个人操作行为数据生成对应的语音引导文件并播放,以引导用户。
  12. 如权利要求11所述的计算机可读存储介质,其中,所述语音交互程序被处理器执行时还实现以下步骤:
    监测第二预设时间内是否接收到用户的语音指令;
    若第二预设时间内未接收到用户的语音指令,则获取线上各用户的操作行为数据,并根据所述操作行为数据生成对应的语音帮助文件并播放,以帮助用户;
    若第二预设时间内接收到用户的语音指令,则对所述语音指令进行识别,并根据识别结果执行对应的操作。
  13. 如权利要求12所述的计算机可读存储介质,其中,所述语音交互程序被处理器执行时还实现以下步骤:
    获取线上各用户的操作行为数据,并对所述操作行为数据进行统计分析;
    根据统计结果生成对应的语音帮助文件并播放,以帮助用户。
  14. 如权利要求11所述的计算机可读存储介质,其中,所述语音交互程序被处理器执行时还实现以下步骤:
    获取用户的个人操作行为数据,并对所述个人操作行为数据进行统计分析;
    根据统计结果生成对应的语音引导文件并播放,以引导用户。
  15. 如权利要求11所述的计算机可读存储介质,其中,所述语音交互程序被处理器执行时还实现以下步骤:
    若第一预设时间内接收到用户的语音指令,则对所述语音指令进行识别,并根据识别结果执行对应的操作。
PCT/CN2018/093750 2017-12-20 2018-06-29 语音交互方法、装置及计算机可读存储介质 WO2019119771A1 (zh)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201711382608.5A CN108132805B (zh) 2017-12-20 2017-12-20 语音交互方法、装置及计算机可读存储介质
CN201711382608.5 2017-12-20

Publications (1)

Publication Number Publication Date
WO2019119771A1 true WO2019119771A1 (zh) 2019-06-27

Family

ID=62390875

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2018/093750 WO2019119771A1 (zh) 2017-12-20 2018-06-29 语音交互方法、装置及计算机可读存储介质

Country Status (2)

Country Link
CN (1) CN108132805B (zh)
WO (1) WO2019119771A1 (zh)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110333840A (zh) * 2019-06-28 2019-10-15 百度在线网络技术(北京)有限公司 推荐方法、装置、电子设备和存储介质
CN111552794A (zh) * 2020-05-13 2020-08-18 海信电子科技(武汉)有限公司 提示语生成方法、装置、设备和存储介质

Families Citing this family (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108132805B (zh) * 2017-12-20 2022-01-04 深圳Tcl新技术有限公司 语音交互方法、装置及计算机可读存储介质
CN109119076B (zh) * 2018-08-02 2022-09-30 重庆柚瓣家科技有限公司 一种老人用户交流习惯的收集系统及方法
CN109240640B (zh) * 2018-08-30 2021-04-20 百度在线网络技术(北京)有限公司 广告语音交互方法、装置及存储介质
CN109218843B (zh) * 2018-09-27 2020-10-23 四川长虹电器股份有限公司 基于电视设备的个性化智能语音提示方法
CN109360570B (zh) * 2018-10-19 2022-06-21 歌尔科技有限公司 语音设备的语音识别方法、语音设备及可读存储介质
CN109545207A (zh) * 2018-11-16 2019-03-29 广东小天才科技有限公司 一种语音唤醒方法及装置
CN109243462A (zh) * 2018-11-20 2019-01-18 广东小天才科技有限公司 一种语音唤醒方法及装置
CN109410944B (zh) * 2018-12-12 2020-06-09 百度在线网络技术(北京)有限公司 语音交互方法、装置和终端
CN111385595B (zh) * 2018-12-29 2022-05-31 阿里巴巴集团控股有限公司 网络直播方法、直播中补货处理方法、装置、直播服务器及终端设备
CN109903760A (zh) * 2019-01-02 2019-06-18 百度在线网络技术(北京)有限公司 语音交互方法、装置和存储介质
CN109584878A (zh) * 2019-01-14 2019-04-05 广东小天才科技有限公司 一种语音唤醒方法及系统
CN110120222A (zh) * 2019-05-23 2019-08-13 九牧厨卫股份有限公司 一种带语音播报功能的智能镜柜的语音播报方法
CN111026276A (zh) * 2019-12-12 2020-04-17 Oppo(重庆)智能科技有限公司 视觉辅助方法及相关产品
CN111988637A (zh) * 2020-08-21 2020-11-24 广州欢网科技有限责任公司 基于电视直播中用户迷茫时刻的节目推荐方法及装置
CN113301417A (zh) * 2021-04-30 2021-08-24 当趣网络科技(杭州)有限公司 用于智能电视的语音控制方法
CN113301394B (zh) * 2021-04-30 2023-07-11 当趣网络科技(杭州)有限公司 结合用户等级的语音控制方法

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1920946A (zh) * 2005-07-01 2007-02-28 伯斯有限公司 汽车接口
CN101415257A (zh) * 2007-10-16 2009-04-22 康佳集团股份有限公司 一种人机对话聊天方法
JP2012027785A (ja) * 2010-07-26 2012-02-09 Panasonic Corp 携帯端末及び表示方法
CN106233246A (zh) * 2014-04-22 2016-12-14 三菱电机株式会社 用户界面系统、用户界面控制装置、用户界面控制方法和用户界面控制程序
CN108132805A (zh) * 2017-12-20 2018-06-08 深圳Tcl新技术有限公司 语音交互方法、装置及计算机可读存储介质

Family Cites Families (24)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20060123220A1 (en) * 2004-12-02 2006-06-08 International Business Machines Corporation Speech recognition in BIOS
CN103376868A (zh) * 2012-04-19 2013-10-30 鸿富锦精密工业(武汉)有限公司 电源适配器识别系统及方法
CN102880649B (zh) * 2012-08-27 2016-03-02 北京搜狗信息服务有限公司 一种个性化信息处理方法和系统
CN104063043B (zh) * 2013-03-22 2017-07-25 联想(北京)有限公司 一种控制方法及装置
CN103391281B (zh) * 2013-06-25 2016-04-20 福建星网锐捷通讯股份有限公司 基于综合接入设备iad实现协议三合一的融合方法
CN104599616A (zh) * 2013-10-30 2015-05-06 西安景行数创信息科技有限公司 一种智能语音交互引导系统
JP2016224599A (ja) * 2015-05-28 2016-12-28 株式会社島津製作所 ガイドファイル作成プログラム
CN105117008B (zh) * 2015-08-20 2018-07-20 小米科技有限责任公司 操作引导方法及装置、电子设备
CN106970907A (zh) * 2016-01-14 2017-07-21 芋头科技(杭州)有限公司 一种语义识别方法
CN105786575B (zh) * 2016-03-17 2019-06-04 北京奇虎科技有限公司 移动终端及其用于重写系统时数据免清除的方法
CN107305769B (zh) * 2016-04-20 2020-06-23 斑马网络技术有限公司 语音交互处理方法、装置、设备及操作系统
CN106027485A (zh) * 2016-04-28 2016-10-12 乐视控股(北京)有限公司 基于语音交互的富媒体展示方法及系统
CN106205612B (zh) * 2016-07-08 2019-12-24 北京光年无限科技有限公司 面向智能机器人的信息处理方法及系统
CN116844543A (zh) * 2016-08-26 2023-10-03 王峥嵘 一种基于语音交互的控制方法及系统
CN106228975A (zh) * 2016-09-08 2016-12-14 康佳集团股份有限公司 一种移动终端的语音识别系统及方法
CN106653016B (zh) * 2016-10-28 2020-07-28 上海智臻智能网络科技股份有限公司 智能交互方法和装置
CN107357416A (zh) * 2016-12-30 2017-11-17 长春市睿鑫博冠科技发展有限公司 一种人机交互装置及交互方法
CN106598431A (zh) * 2016-11-30 2017-04-26 中国航空工业集团公司沈阳飞机设计研究所 一种基于有人机的快速引导多无人机指令发送装置
CN106531165A (zh) * 2016-12-15 2017-03-22 北京塞宾科技有限公司 一种便携式智能家居语音控制系统及控制方法
CN106910500B (zh) * 2016-12-23 2020-04-17 北京小鸟听听科技有限公司 对带麦克风阵列的设备进行语音控制的方法及设备
CN106648394A (zh) * 2016-12-31 2017-05-10 珠海市魅族科技有限公司 一种语音操控的方法及系统
CN106782606A (zh) * 2017-01-17 2017-05-31 山东南工机器人科技有限公司 用于导讲机器人的语言交流与互动系统及其工作方法
CN107066343A (zh) * 2017-01-24 2017-08-18 广东欧珀移动通信有限公司 移动终端中分区表的修复方法、装置及移动终端
CN107452378A (zh) * 2017-08-15 2017-12-08 北京百度网讯科技有限公司 基于人工智能的语音交互方法和装置

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1920946A (zh) * 2005-07-01 2007-02-28 伯斯有限公司 汽车接口
CN101415257A (zh) * 2007-10-16 2009-04-22 康佳集团股份有限公司 一种人机对话聊天方法
JP2012027785A (ja) * 2010-07-26 2012-02-09 Panasonic Corp 携帯端末及び表示方法
CN106233246A (zh) * 2014-04-22 2016-12-14 三菱电机株式会社 用户界面系统、用户界面控制装置、用户界面控制方法和用户界面控制程序
CN108132805A (zh) * 2017-12-20 2018-06-08 深圳Tcl新技术有限公司 语音交互方法、装置及计算机可读存储介质

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110333840A (zh) * 2019-06-28 2019-10-15 百度在线网络技术(北京)有限公司 推荐方法、装置、电子设备和存储介质
CN111552794A (zh) * 2020-05-13 2020-08-18 海信电子科技(武汉)有限公司 提示语生成方法、装置、设备和存储介质
CN111552794B (zh) * 2020-05-13 2023-09-19 海信电子科技(武汉)有限公司 提示语生成方法、装置、设备和存储介质

Also Published As

Publication number Publication date
CN108132805A (zh) 2018-06-08
CN108132805B (zh) 2022-01-04

Similar Documents

Publication Publication Date Title
WO2019119771A1 (zh) 语音交互方法、装置及计算机可读存储介质
WO2019080406A1 (zh) 电视机语音交互方法、语音交互控制装置及存储介质
WO2017028601A1 (zh) 智能终端的语音控制方法、装置及电视机系统
WO2014003283A1 (en) Display apparatus, method for controlling display apparatus, and interactive system
WO2017043857A1 (ko) 어플리케이션 제공 방법 및 이를 위한 전자 기기
WO2015178692A1 (en) Display apparatus, remote control apparatus, system and controlling method thereof
WO2018018681A1 (zh) 视频节目预览方法及装置
WO2015127870A1 (en) Method and apparatus for recommending application
WO2015144052A1 (en) Method and apparatus for collecting statistics on network information
WO2018113187A1 (zh) 一种显示控制方法及显示装置
WO2014000635A1 (zh) 阅读器提供服务的方法、阅读器、终端及计算机存储介质
WO2019051902A1 (zh) 终端控制方法、空调器及计算机可读存储介质
WO2017063368A1 (zh) 视频广告的插播方法及装置
WO2017063366A1 (zh) 应用启动方法和系统
WO2017036203A1 (zh) 媒体应用的播放控制方法、遥控装置及电视系统
WO2018124842A1 (ko) 컨텐트에 관한 정보를 제공하는 방법 및 디바이스
WO2019085543A1 (zh) 电视机系统及电视机控制方法
WO2021132922A1 (en) Computing device and operation method thereof
WO2017036208A1 (zh) 显示界面中的信息提取方法及系统
WO2021118225A1 (en) Display device and operating method thereof
WO2017084305A1 (zh) 终端控制方法及系统
WO2019100594A1 (zh) 电视时移控制方法、系统及计算机可读存储介质
WO2015069082A1 (en) Display apparatus and method of controlling the same
WO2015169177A1 (en) Web page display method and apparatus
WO2015072717A1 (en) Display apparatus and controlling method thereof

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 18891924

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 18891924

Country of ref document: EP

Kind code of ref document: A1