WO2017181901A1 - 语音唤起方法、装置和设备 - Google Patents

语音唤起方法、装置和设备 Download PDF

Info

Publication number
WO2017181901A1
WO2017181901A1 PCT/CN2017/080387 CN2017080387W WO2017181901A1 WO 2017181901 A1 WO2017181901 A1 WO 2017181901A1 CN 2017080387 W CN2017080387 W CN 2017080387W WO 2017181901 A1 WO2017181901 A1 WO 2017181901A1
Authority
WO
WIPO (PCT)
Prior art keywords
user
voice
vehicle
information
input
Prior art date
Application number
PCT/CN2017/080387
Other languages
English (en)
French (fr)
Inventor
郭云云
蔡丽娟
Original Assignee
斑马网络技术有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 斑马网络技术有限公司 filed Critical 斑马网络技术有限公司
Publication of WO2017181901A1 publication Critical patent/WO2017181901A1/zh

Links

Images

Classifications

    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60RVEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
    • B60R16/00Electric or fluid circuits specially adapted for vehicles and not otherwise provided for; Arrangement of elements of electric or fluid circuits specially adapted for vehicles and not otherwise provided for
    • B60R16/02Electric or fluid circuits specially adapted for vehicles and not otherwise provided for; Arrangement of elements of electric or fluid circuits specially adapted for vehicles and not otherwise provided for electric constitutive elements
    • B60R16/023Electric or fluid circuits specially adapted for vehicles and not otherwise provided for; Arrangement of elements of electric or fluid circuits specially adapted for vehicles and not otherwise provided for electric constitutive elements for transmission of signals between vehicle parts or subsystems
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W50/00Details of control systems for road vehicle drive control not related to the control of a particular sub-unit, e.g. process diagnostic or vehicle driver interfaces
    • B60W50/08Interaction between the driver and the control system
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60RVEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
    • B60R16/00Electric or fluid circuits specially adapted for vehicles and not otherwise provided for; Arrangement of elements of electric or fluid circuits specially adapted for vehicles and not otherwise provided for
    • B60R16/02Electric or fluid circuits specially adapted for vehicles and not otherwise provided for; Arrangement of elements of electric or fluid circuits specially adapted for vehicles and not otherwise provided for electric constitutive elements
    • B60R16/023Electric or fluid circuits specially adapted for vehicles and not otherwise provided for; Arrangement of elements of electric or fluid circuits specially adapted for vehicles and not otherwise provided for electric constitutive elements for transmission of signals between vehicle parts or subsystems
    • B60R16/0231Circuits relating to the driving or the functioning of the vehicle
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W50/00Details of control systems for road vehicle drive control not related to the control of a particular sub-unit, e.g. process diagnostic or vehicle driver interfaces
    • B60W50/08Interaction between the driver and the control system
    • B60W50/14Means for informing the driver, warning the driver or prompting a driver intervention

Definitions

  • the present application relates to Internet technologies, and in particular, to a voice arousing method, apparatus and device for use in a vehicle.
  • voice arousal in a vehicle is a common interaction behavior between a vehicle and a user. For example, when a user has an incoming call, the user can answer the call by evoking the in-vehicle voice function without the driver holding the mobile phone, thereby avoiding problems such as dangerous driving.
  • the current voice arousal requires the user to manually evoke, and the targeted object is relatively simple, and the applicability is not high, so that the human-computer interaction is not intelligent enough.
  • the present invention provides a voice arousing method, device and device to solve the technical problem that the prior art voice arousal requires the user to manually evoke, and the object to be targeted is relatively simple, the applicability is not high, and the human-computer interaction is not intelligent enough.
  • the application provides a voice arousing method, including:
  • the voice arousing method obtained by the above, obtains a voice interaction scene, and automatically evokes a voice function of the vehicle through the voice interaction scene to broadcast a first voice corresponding to the current voice interaction scene to prompt the user to communicate with the vehicle or the vehicle.
  • Devices interact so that the vehicle's system receives the user's basis After the operation information of the first voice input, the corresponding operation is performed according to the operation information, so as to ensure the interaction behavior with the user and satisfy the interaction requirement of the user.
  • the method automatically evokes the voice for the user when acquiring the voice interaction scene, provides the user with rich voice content, facilitates the user to interact with the vehicle, does not require the user to manually trigger the voice interaction function of the vehicle, and improves the human-computer interaction.
  • the voice interaction scene acquired by the in-vehicle system may correspond to the driving state of the vehicle. Since the driving state of the vehicle is relatively rich, the voice interaction scene corresponding to the driving state of the vehicle is also rich, and thus the application is made.
  • the voice function of the vehicle can be actively evoked for different vehicle driving states, and the applicability of the voice arousal of the vehicle is improved.
  • the receiving user performs the corresponding operation according to the operation information of the first voice input and according to the operation information, and specifically includes:
  • the method prompts the user to perform the user operation performed by the user by using the first voice, and then, after the user performs the user operation according to the instruction of the first voice, determines whether the user operation is the user indicated by the first voice according to the user operation instruction input by the user.
  • the user operation is performed, and when the system of the vehicle detects that the operation performed by the user is a user operation performed by the user indicated by the first voice, the first voice is stopped, so that the user can intuitively know the needs of the vehicle, and the user is timely
  • the abnormality of the vehicle or the improper operation of the user is found, which further improves the user experience.
  • the receiving user performs the corresponding operation according to the operation information of the first voice input and according to the operation information, and specifically includes:
  • the first voice is used to prompt the user to perform a selection operation on the content prompted by the first voice, and the second voice is a user selection result;
  • a corresponding operation is performed according to the second voice.
  • the first voice indicates that the user performs a selection operation on the content prompted by the first voice, thereby obtaining a second voice input by the user according to the first voice, and performing a corresponding operation according to the second voice, so that the human-computer interaction is more Intelligently; and the method provided by the embodiment of the present invention can know the potential triggering voice requirement of the user in advance, actively evoke the voice for the user, and prompt the user to wait for an event, thereby avoiding the user forgetting the to-do event, and providing the user with convenience. Further improve the user experience.
  • the acquiring a voice interaction scenario includes:
  • parameter information related to driving of the vehicle includes at least one of external driving environment information, vehicle state information, and user behavior information;
  • the outside driving environment information includes road condition information and/or weather information
  • the vehicle state information includes vehicle own status information and/or vehicle warning sound information.
  • the vehicle warning tone information includes a type of the vehicle warning tone or a number of broadcasts of the vehicle warning tone.
  • the method collects parameter information related to driving of the vehicle, and determines a voice interaction scenario corresponding to the current driving state of the vehicle according to the parameter information, thereby actively arousing the voice function of the vehicle according to the voice interaction scenario, and interacting with the user, that is, Through the diversity of parameter information, the diversity of voice interaction scenes is ensured, which enriches the way of actively arousing the voice function of the vehicle and improves the applicability of the voice arousal of the vehicle.
  • the parameter information includes user behavior information
  • the acquiring parameter information related to driving of the vehicle includes:
  • the method transmits, by a system of a vehicle (for example, an in-vehicle system), an acquisition instruction carrying an authorization code preset by a user or an identifier of the vehicle to a device storing user behavior information, so that the device storing the user behavior information is verifying the authorization code or After the identification of the vehicle is legal, the user behavior information is sent to the vehicle's system to ensure the privacy and security of the user behavior information.
  • a system of a vehicle for example, an in-vehicle system
  • an acquisition instruction carrying an authorization code preset by a user or an identifier of the vehicle to a device storing user behavior information, so that the device storing the user behavior information is verifying the authorization code or After the identification of the vehicle is legal, the user behavior information is sent to the vehicle's system to ensure the privacy and security of the user behavior information.
  • the parameter information includes user behavior information
  • the acquiring parameter information related to driving of the vehicle includes:
  • the user behavior information includes a user to-do event.
  • the user to-be-scheduled event includes the moment when the user is to be held, and the user broadcasts the first voice to the user, which specifically includes:
  • the first voice is broadcast to the user.
  • the method broadcasts the first voice to the user when the time arrives by determining the time at which the first voice is broadcasted, which may Knowing the user's potential triggering voice requirements in advance, actively evokes the voice for the user, and prompts the user to wait for the event, avoiding the user forgetting the to-do event, providing convenience to the user, and further improving the user experience.
  • the present application provides a voice arousing device, including:
  • the obtaining module is configured to obtain a voice interaction scenario
  • a determining module configured to determine a corresponding first voice according to the voice interaction scenario acquired by the acquiring module, where the first voice is a voice prompting a user to interact with the vehicle or a device on the vehicle;
  • a voice broadcast module configured to broadcast, to the user, the first voice determined by the determining module
  • a receiving module configured to receive operation information of the first voice input determined by the user according to the determining module
  • a processing module configured to perform a corresponding operation according to the operation information received by the receiving module.
  • the receiving module is specifically configured to receive a user operation instruction input by the user according to the first voice
  • the processing module is configured to determine, according to the user operation instruction, whether the operation of the user is a user operation performed by a user indicated by the first voice, and determine that the operation of the user is the first voice
  • the voice broadcast module is instructed to stop broadcasting the first voice.
  • the receiving module is configured to receive a second voice input by the user according to the content suggested by the first voice, where the first voice is used to prompt the user to target the first voice
  • the content of the prompt performs a selection operation, and the second voice is a selection result of the user
  • the processing module is specifically configured to perform a corresponding operation according to the second voice.
  • the acquiring module includes:
  • Obtaining a sub-module configured to acquire parameter information related to driving of the vehicle; wherein the parameter information includes at least one of external driving environment information, vehicle state information, and user behavior information;
  • the determining submodule is configured to determine, according to the parameter information, a voice interaction scenario corresponding to a current driving state of the vehicle.
  • the outside driving environment information includes road condition information and/or weather information
  • the vehicle state information includes vehicle own condition information and/or vehicle warning sound information.
  • the vehicle warning sound information includes a type of the vehicle warning sound or a number of broadcasts of the vehicle warning sound.
  • the parameter information includes user behavior information
  • the acquiring sub-module includes:
  • a sending unit configured to send an acquisition instruction to a device that stores the user behavior information, where the obtaining instruction carries The authorization code preset by the user or the identifier of the vehicle;
  • a receiving unit configured to receive user behavior information that is sent by the device after verifying the authorization code or the identifier of the vehicle is legal.
  • the parameter information includes user behavior information
  • the acquiring sub-module is specifically configured to receive user behavior information input by the user.
  • the user behavior information includes a user to-do event.
  • the user to-be-scheduled event includes a moment when the user's pending event occurs;
  • the voice broadcast module includes:
  • a determining unit configured to determine, according to the moment of occurrence of the user to-be-scheduled event, a time to broadcast the first voice to the user
  • a broadcast unit configured to broadcast the first voice to the user when the time determined by the determining unit arrives.
  • the beneficial effects of the voice evoking device can be referred to the advantages of the voice evoke method in the above-mentioned various implementation manners, and details are not described herein again.
  • the present application provides a voice arousing device, including:
  • a processor configured to acquire a voice interaction scenario, and determine a corresponding first voice according to the voice interaction scenario
  • An output device coupled to the processor for broadcasting the first voice to a user, the first voice being a voice prompting a user to interact with the vehicle or a device on the vehicle;
  • An input device coupled to the processor, configured to receive operation information of the user according to the first voice input
  • the processor is further configured to perform a corresponding operation according to the operation information obtained by the input device.
  • the input device is specifically configured to receive a user operation instruction input by a user according to the first voice
  • the processor is specifically configured to determine, according to the user operation instruction, whether the operation of the user is a user operation performed by a user indicated by the first voice, and determine that the operation of the user is the first voice
  • the output device is instructed to stop broadcasting the first voice.
  • the input device is specifically configured to receive a second voice input by the user according to the content suggested by the first voice, where the first voice is used to prompt the user for the first voice
  • the content of the prompt performs a selection operation, and the second voice is a selection result of the user;
  • the processor is specifically configured to perform a corresponding operation according to the second voice.
  • the input device is further configured to acquire a parameter letter related to driving of the vehicle.
  • the parameter information includes at least one of external driving environment information, vehicle state information, and user behavior information;
  • the processor is specifically configured to determine, according to the parameter information, a voice interaction scenario corresponding to a current driving state of the vehicle.
  • the outside driving environment information includes road condition information and/or weather information
  • the vehicle state information includes vehicle own condition information and/or vehicle warning sound information.
  • the vehicle warning sound information includes a type of the vehicle warning sound or a number of broadcasts of the vehicle warning sound.
  • the parameter information includes user behavior information
  • the output device is further configured to send an acquisition instruction to the device that stores the user behavior information, where the acquisition instruction carries an authorization code preset by the user or an identifier of the vehicle;
  • the input device is specifically configured to receive user behavior information that is sent by the device after verifying the authorization code or the identifier of the vehicle is legal.
  • the parameter information includes user behavior information
  • the input device is specifically configured to receive user behavior information input by the user.
  • the user behavior information includes a user to-do event.
  • the user to-do event includes a moment when the user's pending event occurs
  • the processor is further configured to determine, according to an occurrence moment of the user to-be-scheduled event, a time to broadcast the first voice to the user;
  • the output device is specifically configured to broadcast the first voice to the user when the time arrives.
  • the present application provides a voice arousing device for a vehicle, including: an onboard processor, an onboard output device, and an onboard input device;
  • the airborne processor is configured to obtain a voice interaction scenario, and determine a corresponding first voice according to the voice interaction scenario;
  • the onboard output device is coupled to the onboard processor for broadcasting the first voice to a user, the first voice being to prompt a user to interact with the vehicle or a device on the vehicle Voice
  • the onboard input device is coupled to the onboard processor for receiving operation information of the user according to the first voice input;
  • the onboard processor is further configured to perform a corresponding operation according to the operation information obtained by the onboard input device.
  • the application provides an in-vehicle internet operating system, including:
  • the voice control unit determines a corresponding first voice according to the obtained voice interaction scenario and broadcasts the message to the user;
  • the operation control unit controls the voice arousing system to perform a corresponding operation according to the operation information acquired by the vehicle input device; wherein the operation information is input to the vehicle input device by the user according to the first voice.
  • the user by acquiring the voice interaction scenario corresponding to the current driving state of the vehicle, and automatically arousing the voice function of the vehicle through the voice interaction scenario, the user broadcasts the first voice corresponding to the current voice interaction scenario to prompt the user and the traffic.
  • the tool or the device on the vehicle interacts, so that after receiving the operation information input by the user according to the first voice, the system of the vehicle performs a corresponding operation according to the operation information to ensure interaction with the user and satisfy User interaction requirements.
  • the application automatically evokes the voice for the user when the voice interaction scene is obtained, provides the user with rich voice content, facilitates the user to interact with the vehicle, does not require the user to manually trigger the voice interaction function of the vehicle, and improves the human-computer interaction.
  • the obtained voice interaction scene can correspond to the driving state of the vehicle, and the vehicle interaction state is rich, so that the voice interaction scene corresponding to the driving state of the vehicle is also rich, thereby enabling the present application to
  • the different driving states of the vehicle actively evoke the voice function of the vehicle, which improves the applicability of the voice arousal of the vehicle.
  • FIG. 1 is a schematic diagram of an optional networking manner of the present application
  • FIG. 2 is a schematic flowchart of a voice arousing method according to an embodiment of the present application
  • FIG. 3 is a schematic flowchart diagram of a voice arousing method according to an embodiment of the present application
  • FIG. 4 is a schematic flowchart of a voice arousing method according to an embodiment of the present application.
  • FIG. 5 is a schematic flowchart diagram of a voice arousing method according to an embodiment of the present application.
  • FIG. 6 is a schematic flowchart diagram of a voice arousing method according to an embodiment of the present application.
  • FIG. 7 is a schematic flowchart diagram of a voice arousing method according to an embodiment of the present application.
  • FIG. 8 is a schematic flowchart of a voice arousing method according to an embodiment of the present disclosure.
  • FIG. 9 is a schematic structural diagram of a voice arousing device according to an embodiment of the present application.
  • FIG. 10 is a schematic structural diagram of a voice arousing device according to an embodiment of the present application.
  • FIG. 11 is a schematic structural diagram of a voice arousing device according to an embodiment of the present application.
  • FIG. 12 is a schematic structural diagram of a hardware of a voice arousing device according to an embodiment of the present disclosure.
  • FIG. 13 is a schematic structural diagram of an in-vehicle system according to an embodiment of the present application.
  • FIG. 14 is a schematic structural diagram of an in-vehicle Internet operating system according to an embodiment of the present application.
  • the vehicle involved in the embodiments of the present application includes, but is not limited to, an automobile or a motorcycle, an electric car or a motorcycle, an electric bicycle, an electric balance vehicle, a remote control vehicle, a small aircraft (for example, an unmanned aerial vehicle, a person) Small aircraft, remotely piloted aircraft, and various variants.
  • the vehicle involved herein may be a single oil road vehicle, a single steam road vehicle, a fuel-air combined vehicle, or a power-assisted electric vehicle.
  • the embodiment of the present application does not do the type of the vehicle. By definition, the vehicle has a corresponding onboard system. The following embodiments are described by taking a vehicle as an example of a vehicle.
  • the voice arousing method, device and device may be applied to the networking mode shown in FIG. 1.
  • the network architecture in FIG. 1 may include an in-vehicle system of a vehicle and a wireless network.
  • the user operation interface may be set in the in-vehicle system.
  • the user operation interface may be a voice interface for user input, or an interface for receiving an operation instruction manually triggered by the user, such as a USB disk interface, a USB interface, and a seat belt. Insert the mouth and so on.
  • the vehicle can be connected to a wireless network.
  • the wireless network can be a 2G network, a 3G network, a 4G network, or a 5G network, a Wireless Fidelity (WIFI) network, etc.
  • the wireless network It can also be the Internet of Things or the Internet of Vehicles.
  • the specific type or specific form of the network is not limited as long as it can provide an interface for the vehicle to access and acquire data.
  • the in-vehicle system can access different network servers, such as a mailbox server, a short message server, a cloud server, etc., and only three servers are shown in FIG. 1, but are not limited thereto.
  • the method involved in the embodiment of the present application may be an in-vehicle system, and optionally, the vehicle
  • the carrier system may be a system integrated with a vehicle on the vehicle, such as a car navigation system and/or an in-vehicle entertainment system, or may be a system including a vehicle and other devices of the vehicle, such as sensors, which may be carried out with the vehicle and the user.
  • the embodiment of the present application does not limit the specific content of the in-vehicle system, as long as it can actively evoke the voice function of the vehicle.
  • the execution subject is an in-vehicle system, but the execution subject of the embodiment of the present application is not limited thereto.
  • the voice arousing method, device and device are directed to solving the problem that the voice arousal in the prior art requires the user to manually evoke, and the targeted object is relatively simple, the applicability is not high, and the human-computer interaction is not intelligent enough. problem.
  • FIG. 2 is a schematic flowchart diagram of a voice arousing method according to an embodiment of the present application.
  • the embodiment is directed to the in-vehicle system actively reporting the first voice corresponding to the current voice interaction scenario to the user according to the determined voice interaction scenario, and executing the operation information input by the user according to the content prompted by the first voice.
  • Corresponding operations to ensure active interaction with the user providing users with a convenient and specific process.
  • the method may include:
  • the vehicle is a vehicle
  • the voice interaction scenario involved in the embodiment of the present application may be a scenario that can actively trigger the in-vehicle system to evoke the voice function to broadcast the voice to the user and interact with the user.
  • the voice interaction scenario may be a scenario in which the in-vehicle system can actively perform a corresponding operation to the user voice prompting, for example, actively presenting the vehicle abnormality information to the user by using a voice (eg, the trunk of the vehicle is not closed) to enable the user to eliminate the abnormality of the vehicle. Voice scene.
  • the voice interaction scenario may also be that the in-vehicle system actively searches for the potential voice interaction requirement of the user by learning or acquiring some parameter information related to the driving of the vehicle (for example, the vehicle often passes through a certain geographic location), and actively initiates a prompt to the user or
  • some parameter information related to the driving of the vehicle for example, the vehicle often passes through a certain geographic location
  • the in-vehicle system can learn the location information of the vehicle to the vehicle, when the user 8 When starting, you can ask the user whether to go to the convenience store to buy breakfast at 8:30, and then actively navigate the convenience store to the user according to the user's voice response or actively play the soothing music for the user.
  • the in-vehicle system can log in to the user's mailbox server to obtain some mails of the user after authorization by the user. If there is a user in a mail to drive to a certain place at a certain time, the vehicle system can take the initiative after the user gets on the vehicle. Initiating a voice inquiry to the user, that is, asking the user whether to go to the place to have a meeting, thereby performing a corresponding operation according to the obtained user's voice answer. That is, the technical solution related to the embodiment of the present application does not require the user to manually trigger to arouse the voice interaction function of the vehicle, but the voice interaction scenes that are fundamentally different are automatically used. The user evokes the voice and provides the user with rich voice content, which is convenient for the user to interact with the vehicle, thereby improving the intelligence of human-computer interaction.
  • the voice interaction scenario in the embodiment of the present application may be determined in any manner.
  • the voice interaction scenario may be determined by using a running state of the vehicle.
  • the driving state of the vehicle may be the surrounding environment information of the vehicle, the vehicle speed of the vehicle, etc., or may be whether there is a user operation inside the vehicle, whether the vehicle itself has an unsafe factor, or the like, and the type of the driving state of the vehicle in the embodiment of the present application is not Make a limit.
  • different driving states of the vehicle may correspond to different voice interaction scenarios, and regardless of the driving state of the vehicle, the vehicle system may actively evoke the vehicle as long as the in-vehicle system can acquire the voice interaction scenario corresponding to the driving state of the vehicle.
  • the voice function that is, the voice interaction scenario that automatically evokes the voice function of the vehicle in the embodiment of the present application is rich and has high applicability.
  • the determination of the voice interaction scenario by the driving state of the vehicle is only an example, and the embodiment of the present application is not limited thereto.
  • S102 Determine a corresponding first voice according to the voice interaction scenario, and broadcast the first voice to a user, where the first voice is a voice prompting a user to interact with a vehicle or a device on the vehicle.
  • the first voice corresponding to the voice interaction scenario may be determined.
  • the in-vehicle system may pre-establish a mapping relationship between different voice interaction scenarios and different broadcast voices.
  • the in-vehicle system actively broadcasts the first message to the user. voice.
  • the form of the voice interaction scenario may be an information form, so the in-vehicle system may perform keyword recognition on the voice interaction scenario, and compose the identified keyword into a first voice through a voice synthesis technology, and broadcast the message to the user.
  • the method for determining the first voice according to the obtained voice interaction scenario is not limited in the embodiment of the present application.
  • the voice content of the first voice is a voice content that prompts the user to interact with the vehicle or the device on the vehicle.
  • the “interaction” referred to herein may be a voice interaction between the user and the vehicle or the device on the vehicle. It may also be a manually operated interaction of the user with the vehicle or equipment on the vehicle, and may also be other ways of interaction of the user with the vehicle or equipment on the vehicle.
  • the device on the vehicle may be a car radio, a driving recorder or other in-vehicle device.
  • S103 Receive operation information of the user according to the first voice input, and perform a corresponding operation according to the operation information.
  • the corresponding user operation is performed according to the content of the first voice, and the user operation may be a voice input of the user, which may be a manual operation of the user.
  • the in-vehicle system receives the operation information input by the user, the corresponding operation is performed according to the operation information to ensure the interaction behavior with the user and meet the interaction requirement of the user.
  • the voice evoke method provided by the embodiment of the present application obtains a voice interaction scenario, and automatically evokes the voice function of the vehicle through the voice interaction scenario to broadcast the first voice corresponding to the current voice interaction scenario to prompt the user to communicate with the vehicle or the traffic.
  • the device on the tool interacts, so that after receiving the operation information input by the user according to the first voice, the system of the vehicle performs a corresponding operation according to the operation information to ensure interaction with the user and satisfy the interaction of the user.
  • the method provided by the embodiment of the present application automatically evokes a voice for a user when acquiring a voice interaction scenario, and provides a rich voice content for the user, so that the user can interact with the device on the vehicle or the vehicle without requiring the user to manually trigger the traffic.
  • the voice interaction function of the tool improves the intelligence of human-computer interaction; in addition, the voice interaction scene acquired above can correspond to the driving state of the vehicle, and the determined voice interaction scene is also made due to the rich driving state of the vehicle. It is richer, and thus enables the present application to actively evoke the voice function of the vehicle for different vehicle driving states, thereby improving the applicability of the voice arousal of the vehicle.
  • FIG. 3 is a schematic flowchart diagram of a voice arousing method according to an embodiment of the present application. This embodiment relates to a specific process for acquiring a voice interaction scenario.
  • the foregoing S101 may specifically include:
  • S201 Acquire parameter information related to driving of the vehicle; wherein the parameter information includes at least one of external driving environment information, vehicle state information, and user behavior information.
  • the external driving environment information may be external driving environment information
  • the vehicle status information may be vehicle status information
  • the involved vehicle status information may be vehicle condition information
  • the vehicle warning tone information may be vehicle warning tone information. Therefore, the in-vehicle system can acquire parameter information related to driving of the vehicle in real time, and the parameter information may include at least one of external driving environment information, vehicle state information, and user behavior information.
  • the external driving environment information includes road condition information and/or weather information
  • the vehicle status information includes vehicle condition information and/or vehicle warning sound information
  • the user behavior information may include a to-do event of the user, and may further include:
  • Some operation of the vehicle for example, the user inserts a USB flash drive into the USB interface of the vehicle
  • may also include a certain behavior habit of the user for example, the user is accustomed to doing the same thing at a certain time).
  • the road condition information may include information about whether the road is congested, or whether a traffic accident occurs on the road ahead or whether the road section has a surveillance camera
  • the vehicle condition information may include information about the vehicle's own facility conditions, such as whether the engine is faulty or the brake pad. Whether the information is good or not, the vehicle warning sound information may include volume information of a vehicle warning sound, timbre information of a vehicle warning sound, and the like.
  • the vehicle warning tone information may also be a vehicle alarm The type of announcement or the number of announcements of the vehicle warning tone.
  • the in-vehicle system may establish a communication connection with the traffic road monitoring server through the wireless network shown in FIG. 1 to obtain the road condition information, and obtain the road condition information through the radio broadcast; optionally,
  • the parameter information is weather information
  • the in-vehicle system can obtain weather information through the network server in the wireless network shown in FIG. 1;
  • the parameter information is the vehicle condition information
  • the in-vehicle system can detect by using different sensors or software of the vehicle itself.
  • the vehicle condition information is obtained; when the parameter information is the vehicle warning sound information, the vehicle system can obtain the vehicle warning sound information through cooperation between the vehicle audio and the processor.
  • the specific manner in which the in-vehicle system obtains the user behavior information may include two possible implementation manners, as follows:
  • the foregoing S201 may specifically include:
  • S301 Send an acquisition instruction to a device that stores the user behavior information, where the acquisition instruction carries an authorization code preset by the user or an identifier of the vehicle.
  • the identification of the vehicle may be the identity of the vehicle.
  • the device for storing user behavior information may be a user's terminal, and may be various servers in the network, such as a mailbox server, a short message server, a cloud server, and the like.
  • the user may pre-configure the vehicle with an authorization code, and authorize the vehicle to obtain the behavior information of the user from the device that stores the user behavior information.
  • the user may not need to configure an authorization code for the vehicle, and the user may register or reserve the identifier of the vehicle in the device that stores the user behavior information, and when the vehicle corresponding to the identifier accesses the device that stores the user behavior information,
  • the user's behavior information can be obtained from a device that stores user behavior information. Based on this, the in-vehicle system can request to acquire the user behavior information by sending an acquisition instruction to the device that stores the user behavior information.
  • S302 Receive user behavior information that is sent by the device after verifying the authorization code or the identifier of the vehicle is legal.
  • the device that stores the user behavior information receives the acquisition instruction of the vehicle, the device performs the corresponding legal judgment according to the content carried in the acquisition instruction, that is, determines whether the authorization code in the acquisition instruction is a pre-configured authorization code or a judgment of the user. Obtain whether the identification of the vehicle carried in the instruction is the identification of the vehicle authorized by the user.
  • the user behavior information can be sent to the in-vehicle system to ensure the privacy of the user behavior information.
  • the in-vehicle system can directly receive user behavior information input by the user.
  • the in-vehicle system may provide a user input interface to the user, and the user input interface may be a device access interface, a voice input interface, a handwriting interface of the display screen, or the like. Therefore, The user can insert a device that stores its own behavior information or a USB flash drive into the device access interface, so that the vehicle system can read the user behavior information, or obtain the voice input by the user through the voice input interface, and pass the corresponding voice.
  • the recognition technology obtains the user behavior information, and can also obtain the user behavior information input by the user through the handwriting interface.
  • the in-vehicle system can obtain user behavior information through the first possible implementation manner or the second possible implementation manner described above.
  • S202 Determine, according to the parameter information, a voice interaction scenario corresponding to a current driving state of the vehicle.
  • the in-vehicle system can determine the current vehicle running state, and further determine a voice interaction scenario corresponding to the current driving state of the vehicle.
  • the in-vehicle system may determine, by using a preset first mapping relationship, a voice interaction scenario corresponding to a current driving state of the vehicle, where the first mapping relationship may include different parameter information, different vehicle driving states, and different voice interaction scenarios. Correspondence between them.
  • the embodiment of the present application does not limit the manner of the voice interaction scenario corresponding to the current driving state of the vehicle according to the parameter information, as long as the voice interaction scenario corresponding to the current driving state of the vehicle can be determined according to the parameter information.
  • the voice arousing method provided by the embodiment of the present application collects parameter information related to driving of the vehicle, and determines a voice interaction scene corresponding to the current driving state of the vehicle according to the parameter information, thereby actively arousing the voice of the vehicle according to the voice interaction scene. Function to interact with the user.
  • the technical solution of the embodiment of the present application ensures the diversity of the voice interaction scene through the diversity of the parameter information, thereby enriching the manner of actively arousing the voice function of the vehicle, and improving the applicability of the voice arousal of the vehicle.
  • FIG. 5 is a schematic flowchart diagram of a voice arousing method according to an embodiment of the present application.
  • the embodiment relates to a specific process in which the user inputs a user operation instruction to the in-vehicle system according to the content indicated by the first voice, and the in-vehicle system determines whether to stop broadcasting the first voice according to the user operation instruction.
  • the foregoing S103 may specifically include:
  • S401 Receive a user operation instruction input by the user according to the first voice.
  • the first voice in the embodiment is used to instruct the user to perform a corresponding user operation. Therefore, after the user receives the first voice broadcast by the in-vehicle system, the user operation indicated by the first voice is performed according to the first voice.
  • the in-vehicle system after the user performs the user operation, the in-vehicle system can detect the operation instruction corresponding to the user operation through an interaction interface between the vehicle and the user, or an interaction interface between the device and the user on the vehicle. .
  • the inside of the vehicle will sound a beep sound (the beep
  • the sound is the vehicle warning sound), that is, the current vehicle driving state is the beep sound
  • the in-vehicle system determines that the voice interaction scene corresponding to the driving state of the vehicle is the seat belt prompting scene, and the in-vehicle system determines the corresponding scene according to the seat belt prompting scene.
  • the first voice is “Please fasten the seat belt” and broadcast to the user; after the user receives the first voice, the seat belt is fastened according to the content of the first voice, whereby the vehicle system can be provided by the vehicle.
  • the seat belt insertion port detects the user's operation, and then the operation instruction corresponding to the user operation is “inserted seat belt”.
  • the inside of the vehicle will also sound a beep sound (the beep sound may be different from the type of the beep sound in the above example), that is, the current vehicle running state is beep sound
  • the in-vehicle system determines that the voice interaction scenario corresponding to the driving state of the vehicle is the left door abnormality scenario, and the in-vehicle system determines that the first voice corresponding to the scenario is “Please close the left door” according to the door abnormal scenario, and broadcasts the message to the user.
  • the left door is closed according to the content of the first voice, whereby the vehicle system can detect the operation of the user through the door detection interface provided by the vehicle, thereby obtaining the user operation.
  • the corresponding operation instruction is "closed door”.
  • the in-vehicle system can determine the voice interaction scene corresponding to the current vehicle driving state, and then determine the first voice corresponding to the voice interaction scene, that is, the text.
  • the method of the text to speech (TTS) prompts the user to the clear voice content, which prevents the user from being able to know the meaning of the beep sound when the user sings the beep sound in the car.
  • TTS text to speech
  • the solution of the embodiment of the present application can make the user intuitive and effective. Know the meaning of the beep tone and improve the user experience.
  • the in-vehicle system can distinguish the voice interaction scene corresponding to the sounding of the warning sound of the vehicle by the type of the warning sound of the vehicle or the number of warning sounds of the vehicle, for example, when the warning sound of the vehicle is “drip”, the in-vehicle system determines The corresponding voice interaction scene should be a seat belt prompt scene. When the vehicle warning sound is “ ⁇ ”, the vehicle system determines that the corresponding voice interaction scene is a door abnormal scene.
  • the in-vehicle system broadcasts the intuitive first voice to the user as an example.
  • the vehicle system determines its corresponding voice interaction according to the current driving state.
  • the scenario is to play the first voice corresponding to the voice interaction scenario to the user, to prompt the user to perform the operation of the user indicated by the first voice, so that the user can intuitively and effectively know the needs of the vehicle, so that the user can find the abnormality of the vehicle in time. Or the user's improper operation further enhances the user's experience.
  • S402 Determine, according to the user operation instruction, whether the operation of the user is a user operation performed by a user indicated by the first voice. If yes, execute S403, and if no, execute S404.
  • S404 Play the first voice according to a preset period.
  • the in-vehicle system After the in-vehicle system detects the user operation instruction, it is determined, according to the user operation instruction, whether the operation currently performed by the user is a user operation performed by the user indicated by the first voice; when the in-vehicle system detects that the operation currently performed by the user is When the voice indicates the user operation performed by the user, the in-vehicle system stops broadcasting the first voice. When the in-vehicle system detects that the operation currently performed by the user is not the user operation performed by the user indicated by the first voice, the in-vehicle system plays the preset period.
  • the first voice may be the first voice played once every few seconds.
  • the device when the driving state of the vehicle for which the first voice is directed is a beep sound, when the in-vehicle system detects that the operation currently performed by the user is not the user operation performed by the user indicated by the first voice, the device may repeatedly repeat the sound. Beep sound.
  • the voice evoke method provided by the embodiment of the present application prompts the user to perform a user operation performed by the user through the first voice, and then, after the user performs the user operation according to the instruction of the first voice, determines the user operation according to the user operation instruction input by the user. Whether it is a user operation performed by the user indicated by the first voice, and stops playing the first voice when the system of the vehicle detects that the operation performed by the user is a user operation performed by the user indicated by the first voice.
  • the method provided by the embodiment of the present application can enable the user to intuitively know the needs of the vehicle, and facilitate the user to timely discover the abnormality of the vehicle or the improper operation of the user, thereby further improving the user experience.
  • FIG. 6 is a schematic flowchart diagram of a voice arousing method according to an embodiment of the present application.
  • the embodiment relates to a specific process in which the user inputs a second voice to the in-vehicle system according to the content suggested by the first voice, and the in-vehicle system interacts with the user according to the second voice.
  • the first voice is used to prompt the user to perform a selection operation on the content prompted by the first voice
  • the second voice is a selection result of the user.
  • the foregoing S103 may specifically include:
  • S501 Receive a second voice input by the user according to the content suggested by the first voice.
  • the first voice in the embodiment is used to prompt the user to perform a selection operation on the content prompted by the first voice. Therefore, after the user receives the first voice broadcasted by the in-vehicle system, the user selects according to the content suggested by the first voice, and inputs the selection result to the in-vehicle system by means of voice, and the selection result is the second voice. For example, when the in-vehicle system determines that the voice interaction scene corresponding to the current driving state is the driving mode prompting scene, the in-vehicle system broadcasts the first voice “whether the driving mode is switched to the economic mode” to the user, and the user may input the selection result for the first voice.
  • the user can type "yes” or "no, I want to choose comfort mode.”
  • the parameter information acquired by the in-vehicle system is a user pending event, and the event to be held by the user is “next to the airport at nine o'clock”, and the first voice determined by the in-vehicle system is “now open. The car goes to the airport, so that when the user starts the vehicle, the in-vehicle system broadcasts the first voice, prompting the user to do the event.
  • the in-vehicle system determines an operation corresponding to the second voice according to a matching internal mechanism, and then performs the operation to ensure interaction between the vehicle and the user.
  • the in-vehicle system can learn the content of the second voice through the voice recognition technology, and then according to the second voice. The content switches the driving mode to the comfort mode to meet the user's interaction with the vehicle.
  • the foregoing broadcasting the first voice to the user may include:
  • S601 Determine, according to the moment of occurrence of the user to-be-scheduled event, a moment to broadcast the first voice to the user.
  • S602 Broadcast the first voice to the user when the time arrives.
  • the in-vehicle system may determine, according to the occurrence time, the time at which the first voice is broadcast to the user, and when the time arrives, broadcast to the user. The first voice.
  • the in-vehicle system can transmit the first voice to the user's terminal, and broadcast the first voice to the user through the user's terminal, so as to prompt the user.
  • the parameter information acquired by the in-vehicle system is a user to-do event
  • the user to-do event includes
  • the user's pending event is "the user's small X has a meeting at the A position at 9 o'clock”
  • the on-board system determines that the first speech is broadcasted at 8:30, the first determined by the in-vehicle system.
  • the voice is "After half an hour there is a meeting in the A position, it is expected to arrive in 20 minutes, now whether to go.”
  • the method specifically includes:
  • S701 The in-vehicle system determines the parameter information as “the user small X has a meeting at the A position at 9 o'clock”.
  • S702 The in-vehicle system determines, according to the parameter information, that the voice interaction scenario corresponding to the current driving state is “9 points have a meeting at the A location”.
  • S703 The in-vehicle system determines, according to the voice interaction scenario, that the time for playing the first voice is 8:30.
  • S705 The user inputs the second voice “Yes, go now” to the in-vehicle system according to the first voice.
  • the vehicle system collects another parameter information in real time, and the parameter information is assumed to be road condition information, and the road condition information is “front road congestion”, and the vehicle system determines the voice corresponding to the current driving state.
  • the interaction scenario is a “congested scenario”, and the in-vehicle system determines that the first voice corresponding to the congestion scenario is “front congestion, may go late to the A location, and notify the conference initiator that the phone is small Y”, so that the user receives the first voice.
  • the second voice can be selected and input to the in-vehicle system. If the second voice is “Yes”, the in-vehicle system can initiate a call request to the conference initiator according to the second voice.
  • the first voice indicates that the user performs a selection operation on the content prompted by the first voice, thereby obtaining a second voice input by the user according to the first voice, and performing corresponding according to the second voice.
  • the operation makes the human-computer interaction more intelligent; and the method provided by the embodiment of the present application can know the potential triggering voice requirement of the user in advance, actively evokes the voice for the user, and prompts the user to wait for the event to prevent the user from forgetting to do things. It provides users with convenience and further enhances the user experience.
  • the voice arousing device can be implemented in the infrastructure of the vehicle or in an interactive system of the vehicle and the wireless network.
  • the voice arousing device can be constructed using commercially available hardware components configured by the steps taught by the present solution.
  • the processor component or processing module, processing unit, determining module, etc.
  • the processor component can be implemented using components such as a microcontroller, a microcontroller, a microprocessor, etc. from companies such as Texas Instruments, Intel Corporation, and ARM Corporation.
  • FIG. 9 is a schematic structural diagram of a voice arousing device according to an embodiment of the present disclosure.
  • the voice arousing device may be implemented by software, hardware, or a combination of the two.
  • the voice arousing device includes: an obtaining module 10, a determining module 11, a voice broadcast module 12, a receiving module 13, and a processing module 14.
  • the obtaining module 10 is configured to acquire a voice interaction scenario.
  • the determining module 11 is configured to determine, according to the voice interaction scenario acquired by the acquiring module 10, a corresponding first voice, where the first voice is to prompt the user to interact with the vehicle or the device on the vehicle. voice;
  • the voice broadcast module 12 is configured to broadcast to the user the first voice determined by the determining module 11;
  • the receiving module 13 is configured to receive operation information of the first voice input determined by the user according to the determining module 11;
  • the processing module 14 is configured to perform a corresponding operation according to the operation information received by the receiving module 13.
  • the voice arousal device provided by the embodiment of the present invention can perform the foregoing method embodiments, and the implementation principle and technical effects are similar, and details are not described herein again.
  • the receiving module 13 is specifically configured to receive a user operation instruction input by the user according to the first voice
  • the processing module 14 is configured to determine, according to the user operation instruction, whether the operation of the user is a user operation performed by the user indicated by the first voice, and determine that the operation of the user is the first When the voice indicates the user operation performed by the user, the voice broadcast module 12 is instructed to stop broadcasting the first voice.
  • the receiving module 13 is specifically configured to receive a second voice input by the user according to the content suggested by the first voice, where the first voice is used to prompt the user. Performing a selection operation on the content suggested by the first voice, where the second voice is a selection result of the user;
  • the processing module 14 is specifically configured to perform a corresponding operation according to the second voice.
  • FIG. 10 is a schematic structural diagram of a voice arousing device provided by an embodiment of the present application.
  • the obtaining module 10 includes:
  • the obtaining sub-module 101 is configured to acquire parameter information related to driving of the vehicle; wherein the parameter information includes at least one of external driving environment information, vehicle state information, and user behavior information;
  • the determining sub-module 102 is configured to determine, according to the parameter information, a voice interaction scenario corresponding to a current driving state of the vehicle.
  • the outside driving environment information includes road condition information and/or weather information
  • the vehicle status information includes vehicle own status information and/or vehicle warning sound information
  • the vehicle warning sound information includes a type of the vehicle warning sound or a number of broadcasts of the vehicle warning sound.
  • the parameter information includes user behavior information
  • the method further includes: the sending sub-module 101, the sending unit 1011, configured to send an acquiring instruction to the device that stores the user behavior information, where the acquiring The instruction carries an authorization code preset by the user or an identifier of the vehicle;
  • the receiving unit 1012 is configured to receive user behavior information that is sent by the device after verifying the authorization code or the identifier of the vehicle is legal.
  • the parameter information includes user behavior information
  • the obtaining sub-module 101 is specifically configured to receive user behavior information input by the user.
  • the user behavior information includes a user to-do event.
  • the foregoing user to-be-scheduled event includes a moment of occurrence of the user-to-do event
  • the structure of the voice-evoking device provided in an embodiment of the present application is provided in FIG.
  • the voice broadcast module 12 may include:
  • a determining unit 121 configured to determine, according to the moment of occurrence of the user to-be-scheduled event, a time to broadcast the first voice to the user;
  • the broadcast unit 122 is configured to broadcast the first voice to the user when the time determined by the determining unit 121 arrives.
  • the voice arousal device provided by the embodiment of the present invention can perform the foregoing method embodiments, and the implementation principle and technical effects are similar, and details are not described herein again.
  • FIG. 12 is a schematic structural diagram of a hardware of a voice arousing device according to an embodiment of the present application.
  • the voice arousing device may be integrated in the in-vehicle system in the above embodiment, or may be a stand-alone in-vehicle system.
  • the voice arousing device can include a processor 20, an output device 21, an input device 22, a memory 23, and at least one communication bus 24.
  • Communication bus 24 is used to implement a communication connection between the components.
  • Memory 23 may include high speed RAM memory, and may also include non-volatile memory NVM, such as at least one disk memory, in which various programs may be stored for performing various processing functions and implementing the method steps of the present embodiments.
  • the processor 20 can be, for example, a central processing unit (CPU), an application specific integrated circuit (ASIC), a digital signal processor (DSP), a digital signal processing device (DSPD), and a programmable logic.
  • CPU central processing unit
  • ASIC application specific integrated circuit
  • DSP digital signal processor
  • DSPD digital signal processing device
  • PLD device
  • FPGA field programmable gate array
  • controller controller
  • microcontroller microprocessor
  • microprocessor or other electronic component coupled to the input device 22 and the output device via an in-vehicle line or wireless connection twenty one.
  • the input device 22 may include multiple input devices, for example, at least one of a user-oriented user interface, a device-oriented device interface, and a transceiver.
  • the device-oriented device interface may be a wired interface for data transmission between the device and the device, or may be a hardware insertion interface (for example, a USB interface, for data or instruction transmission between the device and the device,
  • the user-oriented user interface can be, for example, a user-oriented control button, a voice input device for receiving voice input, and a touch perception of the user receiving a user's touch input.
  • the device for example, a touch screen with a touch sensing function, a touch panel, etc.
  • the transceiver may be a radio frequency transceiver chip with a communication function, a baseband processing chip, and a transceiver antenna.
  • the voice arousing device in the embodiment of the present application is a universal voice arousing device. It can be applied to any control system or control device or other type of device.
  • the output device 21 may be a corresponding output interface or a voice playback device or a transceiver with a communication function.
  • the voice arousing device may be a voice arousing device for a vehicle, for example, may be a voice arousing device for a vehicle, a voice arousing device for an aircraft, a voice arousing device for a waterway vehicle, and the like.
  • the voice-activated device for the vehicle the present application provides another embodiment for introduction. Please refer to the following embodiments, which will not be described in detail herein.
  • the processor 20 is configured to acquire a voice interaction scenario, and determine a corresponding first voice according to the voice interaction scenario.
  • An input device 22 coupled to the processor 20, for receiving operation information of a user according to the first voice input;
  • the processor 20 is further configured to perform a corresponding operation according to the operation information obtained by the input device 22.
  • the voice arousal device provided by the embodiment of the present application may perform the foregoing method embodiments, and the implementation principle and technical effects are similar, and details are not described herein again.
  • the input device 22 is specifically configured to receive a user operation instruction input by the user according to the first voice;
  • the processor 20 is configured to determine, according to the user operation instruction, whether the operation of the user is a user operation performed by the user indicated by the first voice, and determine that the operation of the user is the first When the voice indicates the user operation performed by the user, the output device 21 is instructed to stop broadcasting the first voice.
  • the input device 22 is configured to receive a second voice input by the user according to the content suggested by the first voice, where the first voice is used to prompt the user to prompt the content for the first voice. Performing a selection operation, where the second voice is a result of selection by the user;
  • the processor 20 is specifically configured to perform a corresponding operation according to the second voice.
  • the input device 22 is further configured to acquire parameter information related to driving of the vehicle; wherein the parameter information includes at least one of external driving environment information, vehicle state information, and user behavior information;
  • the processor 20 is specifically configured to determine, according to the parameter information, a voice interaction scenario corresponding to a current driving state of the vehicle.
  • the outside driving environment information includes road condition information and/or weather information
  • the vehicle status information includes vehicle own status information and/or vehicle warning sound information.
  • the vehicle warning tone The information includes the type of the vehicle warning tone or the number of times the vehicle warning tone is broadcast.
  • the parameter information includes user behavior information.
  • the output device 21 is further configured to send an acquisition instruction to the device that stores the user behavior information, where the acquisition instruction carries an authorization code preset by the user or an identifier of the vehicle;
  • the input device 22 is specifically configured to receive user behavior information that is sent by the device after verifying the authorization code or the identifier of the vehicle is legal.
  • the parameter information includes user behavior information
  • the input device 22 is specifically configured to receive user behavior information input by the user.
  • the user behavior information includes a user to-do event.
  • the user to-do event includes a moment when the user's pending event occurs
  • the processor 20 is further configured to determine, according to an occurrence moment of the user to-be-scheduled event, a time for broadcasting the first voice to the user;
  • the output device 21 is specifically configured to broadcast the first voice to the user when the time arrives.
  • the voice arousal device provided by the embodiment of the present application may perform the foregoing method embodiments, and the implementation principle and technical effects are similar, and details are not described herein again.
  • FIG. 13 is a block diagram of an in-vehicle system according to an embodiment of the present application.
  • the in-vehicle system 800 can be a device that integrates multiple functions.
  • the in-vehicle system can be an on-board computer, a car, etc., and the in-vehicle system can include the above-described voice arousing device.
  • the in-vehicle system 800 can include one or more of the following components: a processing component 802, a memory 804, a power component 806, a multimedia component 808, an audio component 810, an input/output (I/O) interface 812, and a sensor component 814. And communication component 816.
  • Processing component 802 typically controls the overall operation of in-vehicle system 800, such as operations associated with display, telephone calls, data communications, camera operations, and recording operations.
  • Processing component 802 can include one or more processors 820 to execute instructions to perform all or part of steps S101 through S706 in the voice arousing method described above.
  • processing component 802 can include one or more modules to facilitate interaction between component 802 and other components.
  • processing component 802 can include a multimedia module to facilitate interaction between multimedia component 808 and processing component 802.
  • Memory 804 is configured to store various types of data to support operation at onboard system 800. Examples of such data include instructions for any application or method operating on the in-vehicle system 800, contact data, phone book data, messages, pictures, videos, and the like. Memory 804 can be implemented by any type of volatile or non-volatile storage device, or a combination thereof, such as static random access memory (SRAM), electrically erasable programmable read-only memory Memory (EEPROM), Erasable Programmable Read Only Memory (EPROM), Programmable Read Only Memory (PROM), Read Only Memory (ROM), Magnetic Memory, Flash Memory, Disk or Optical Disk.
  • SRAM static random access memory
  • EEPROM electrically erasable programmable read-only memory Memory
  • EPROM Erasable Programmable Read Only Memory
  • PROM Programmable Read Only Memory
  • ROM Read Only Memory
  • Magnetic Memory Flash Memory
  • Disk Disk or Optical Disk.
  • Power component 806 provides power to various components of in-vehicle system 800.
  • Power component 806 can include a power management system, one or more power sources, and other components associated with generating, managing, and distributing power for in-vehicle system 800.
  • the multimedia component 808 includes a screen that provides an output interface between the in-vehicle system 800 and a user.
  • the screen can include a liquid crystal display (LCD) and a touch panel (TP). If the screen includes a touch panel, the screen can be implemented as a touch screen to receive input signals from the user.
  • the touch panel includes one or more touch sensors to sense touches, slides, and gestures on the touch panel. The touch sensor may sense not only the boundary of the touch or sliding action, but also the duration and pressure associated with the touch or slide operation.
  • the multimedia component 808 can also include a front camera.
  • the audio component 810 is configured to output and/or input an audio signal.
  • the audio component 810 includes a microphone (MIC) that is configured to receive an external audio signal when the in-vehicle system 800 is in an operational mode, such as a call mode, a recording mode, and a voice recognition mode.
  • the received audio signal may be further stored in memory 804 or transmitted via communication component 816.
  • the audio component 810 also includes a speaker for outputting an audio signal.
  • the I/O interface 812 provides an interface between the processing component 802 and the peripheral interface module, which may be a click wheel, a button, or the like. These buttons may include, but are not limited to, a volume button, a start button, and a lock button.
  • Sensor assembly 814 includes one or more sensors for providing on-board system 800 with various aspects of status assessment.
  • the sensor assembly 814 can also include an acceleration sensor, a gyro sensor, a magnetic sensor, a pressure sensor, or a temperature sensor.
  • Communication component 816 is configured to facilitate wired or wireless communication between in-vehicle system 800 and other devices.
  • the in-vehicle system 800 can access a wireless network based on communication standards, such as WiFi, 2G or 3G, or a combination thereof.
  • communication component 816 receives broadcast signals or broadcast associated information from an external broadcast management system via a broadcast channel.
  • the communication component 816 also includes a near field communication (NFC) module to facilitate short range communication.
  • the NFC module can be implemented based on radio frequency identification (RFID) technology, infrared data association (IrDA) technology, ultra-wideband (UWB) technology, Bluetooth (BT) technology, and other technologies.
  • RFID radio frequency identification
  • IrDA infrared data association
  • UWB ultra-wideband
  • Bluetooth Bluetooth
  • the in-vehicle system 800 can be implemented by one or more application specific integrated circuits (ASICs), digital signal processors (DSPs), digital signal processing devices (DSPDs), programmable logic devices (PLDs), Programming gate array (FPGA), controller, microcontroller, microprocessor or other electronic component implementation for implementation The above voice arousal method is performed.
  • ASICs application specific integrated circuits
  • DSPs digital signal processors
  • DSPDs digital signal processing devices
  • PLDs programmable logic devices
  • FPGA Programming gate array
  • the present application further provides another embodiment, which specifically discloses a voice arousing device for a vehicle.
  • the voice arousing device may be integrated in the central control system of the vehicle, for example, may be integrated in the in-vehicle system involved in the above embodiments.
  • the in-vehicle system may be a system integrated with a vehicle on a vehicle, such as a car navigation system and/or an in-vehicle entertainment system, or may be a system including a car machine and other devices of the vehicle such as sensors.
  • the voice arousing device for the vehicle includes but is not limited to: a vehicle device, an additional control device after the vehicle leaves the factory, and the like.
  • the voice arousing device for a vehicle may include: an onboard input device, an onboard processor, an onboard output device, and other additional devices.
  • the airborne in the "airborne input device”, “airborne output device”, and “airborne processor” may be an "vehicle input device” carried on a vehicle, “Vehicle output device” and “vehicle processor” may also be “onboard input device”, “onboard output device”, “onboard processor” carried on the aircraft, or may be carried on other types of vehicles.
  • the above apparatus does not limit the meaning of "airborne” in the embodiment of the present application.
  • the onboard input device may be an in-vehicle input device
  • the onboard processor may be an onboard processor
  • the onboard output device may be an onboard output device.
  • the onboard processor can use various application specific integrated circuits (ASICs), digital signal processors (DSPs), digital signal processing devices (DSPDs), programmable logic devices (depending on the type of vehicle installed). PLD), field programmable gate array (FPGA), central processing unit (CPU), controller, microcontroller, microprocessor or other electronic component implementation and used to perform the above methods.
  • ASICs application specific integrated circuits
  • DSPs digital signal processors
  • DSPDs digital signal processing devices
  • FPGA field programmable gate array
  • CPU central processing unit
  • controller microcontroller, microprocessor or other electronic component implementation and used to perform the above methods.
  • the onboard processor is coupled to the onboard input device and the onboard output device via an in-vehicle line or wireless connection. According to the method in the embodiment corresponding to the foregoing FIG. 2 to FIG. 8 , the onboard processor is configured to acquire a voice interaction scenario, and determine a corresponding first voice according to the voice interaction scenario.
  • the above-mentioned onboard output device may be an interface capable of interacting with a user (eg, a voice broadcast device, a speaker, a headset, etc.) depending on the type of vehicle installed, or may be established with a user's handheld device or the like.
  • a wirelessly transmitted transceiver that can be coupled to the onboard input device and the onboard processor via an in-vehicle line or wirelessly.
  • the onboard output device is configured to broadcast the first voice to the user, the first voice being a prompting user and the vehicle or the vehicle The voice that the device interacts with.
  • the above-described onboard input device may include a variety of input devices depending on the type of vehicle installed. For example, at least one of a user-oriented in-vehicle user interface, a device-oriented in-vehicle device interface, and a transceiver may be included.
  • the device-oriented device interface may be a wired interface for data transmission between the device and the device (for example, a connection interface with a driving recorder on a center console of the vehicle, and a center console of the vehicle).
  • the line interface between the doors, the hardware interface between the vehicle's center console and the vehicle air conditioner, or a hardware insertion interface (such as a USB interface, a serial port, etc.) for data transmission between the device and the device It may also be a seat belt socket of a vehicle, an interface between a hardware device such as a vehicle engine and other control devices, etc.; alternatively, the user-oriented vehicle user interface may be, for example, a steering wheel control button for a vehicle, for a large vehicle.
  • a central control button for a small vehicle a voice input device for receiving voice input (eg, a microphone placed on a steering wheel or steering rudder, a central sound collection device, etc.), and a touch sensing device that the user receives a user's touch input (such as a touch screen with touch sensing function, a touchpad, etc.); optionally, the above transceiver may have communication in the vehicle Energy RF transceiver chip, baseband chip and a transceiver antenna.
  • the onboard input device is configured to receive operation information of the user according to the first voice input
  • the onboard processor is further configured to perform according to the onboard input. The operation information obtained by the device performs a corresponding operation.
  • the above-mentioned on-board processor may also be used in all or part of the steps in the embodiment of the present invention in the method of pushing the above-mentioned message.
  • a computer/processor readable storage medium having stored therein program instructions for causing the computer/processor to execute:
  • the receiving user performs the corresponding operation according to the operation information of the first voice input, and performs the corresponding operation according to the operation information, specifically:
  • the receiving user performs the corresponding operation according to the operation information of the first voice input, and performs the corresponding operation according to the operation information, specifically:
  • Receiving a second voice input by the user according to the content suggested by the first voice, where the first voice is used for prompting The user performs a selection operation on the content suggested by the first voice, where the second voice is a selection result of the user;
  • a corresponding operation is performed according to the second voice.
  • the obtaining the voice interaction scenario corresponding to the current driving state of the vehicle includes:
  • parameter information related to driving of the vehicle includes at least one of external driving environment information, vehicle state information, and user behavior information;
  • the outside driving environment information includes road condition information and/or weather information
  • the vehicle state information includes vehicle own status information and/or vehicle warning sound information
  • the vehicle warning sound information includes a type of the vehicle warning sound or a number of broadcasts of the vehicle warning sound.
  • the parameter information includes user behavior information
  • the acquiring parameter information related to driving of the vehicle includes:
  • the parameter information includes user behavior information
  • the acquiring parameter information related to driving of the vehicle includes:
  • the user behavior information includes a user to-do event.
  • the user to-be-scheduled event includes the moment when the user's pending event occurs, and the transmitting the first voice to the user, specifically:
  • the first voice is broadcast to the user.
  • the readable storage medium described above can be implemented by any type of volatile or non-volatile storage device, or a combination thereof, such as static random access memory (SRAM), electrically erasable programmable read only memory (EEPROM), Erasable Programmable Read Only Memory (EPROM), Programmable Read Only Memory (PROM), Read Only Memory (ROM), Magnetic Memory, Flash Memory, Disk or Optical Disk.
  • SRAM static random access memory
  • EEPROM electrically erasable programmable read only memory
  • EPROM Erasable Programmable Read Only Memory
  • PROM Programmable Read Only Memory
  • ROM Read Only Memory
  • Magnetic Memory Flash Memory
  • Disk Disk or Optical Disk.
  • the present application further provides an in-vehicle internet operating system.
  • the in-vehicle Internet operating system can manage and control the hardware of the voice arousing device shown in FIG. 12 or FIG. 13 or the hardware of the in-vehicle system or the voice arousing device for the vehicle according to the present application.
  • Hardware And the computer program of the software resource referred to in the present application is software directly running on the voice arousing device or the voice arousing device for the vehicle or the above-described vehicle system according to FIG.
  • the operating system may be an interface between the user and the voice arousing device or the voice arousing device for the vehicle, or may be an interface between the hardware and other software.
  • the in-vehicle Internet operating system provided by the present application can interact with other modules or functional devices on the vehicle to control the functions of the corresponding modules or functional devices.
  • the vehicle is no longer independent of communication.
  • vehicles can be connected to a server or a network server to form a network, thereby forming an in-vehicle Internet.
  • the in-vehicle Internet system can provide voice communication services, location services, navigation services, mobile internet access, vehicle emergency rescue, vehicle data and management services, in-vehicle entertainment services, and the like.
  • FIG. 14 is a schematic structural diagram of an in-vehicle Internet operating system according to an embodiment of the present application. As shown in FIG. 14, the operating system provided by the present application includes:
  • the voice control unit 31 determines a corresponding first voice according to the acquired voice interaction scenario and broadcasts the message to the user;
  • the operation control unit 32 controls the voice arousing system to perform a corresponding operation according to the operation information acquired by the vehicle input device; wherein the operation information is input to the vehicle input device by the user according to the first voice.
  • the voice evoke system in this embodiment may include part of hardware of the voice evoking device in the above embodiment, and may include, for example, the processor and the output device in the above embodiment.
  • the voice evoking system can also be integrated in the above-mentioned vehicle Internet operating system, and can also be used as a system for assisting the vehicle Internet operating system to perform corresponding functional operations.
  • the voice control unit 31 can control the voice arousing system to determine a voice interaction scenario corresponding to the current driving state according to at least one type of parameter information of the collected outside driving environment information, the vehicle state information, and the user behavior information.
  • the voice interaction scenario may be acquired by the voice control unit 31, or may be controlled by the voice control unit 31 by the voice arousing system.
  • the in-vehicle input device in this embodiment may include the input device in the above embodiment, that is, after the voice control unit 31 controls the voice arousing system to broadcast the first voice to the user, the user inputs the operation information to the in-vehicle input device according to the first voice.
  • the operation control unit 32 can control the voice evoking system to perform a corresponding operation based on the operation information.
  • the in-vehicle Internet operating system may control the corresponding components to perform the above-mentioned FIG. 2 to FIG. 8 through the above-mentioned voice control unit 31 and the operation control unit 32, or on the basis of the above two units, in combination with other units. Methods.

Landscapes

  • Engineering & Computer Science (AREA)
  • Automation & Control Theory (AREA)
  • Mechanical Engineering (AREA)
  • Human Computer Interaction (AREA)
  • Transportation (AREA)
  • User Interface Of Digital Computer (AREA)
  • Traffic Control Systems (AREA)
  • Navigation (AREA)

Abstract

一种语音唤起方法、装置和设备。该方法包括:获取语音交互场景;根据所述语音交互场景确定对应的第一语音;接收用户根据所述第一语音输入操作信息,并根据所述操作信息执行相应的操作。该装置包括:获取模块(10);确定模块(11);语音播报模块(12);接收模块(13);处理模块(14)。该设备包括:处理器(20);输出设备(21);输入设备(22)。还公开了一种车载互联网操作系统,包括语音控制单元和操作控制单元。该系统能够保证交通工具与用户之间的交互行为,为用户提供了丰富的语音内容,提高了人机交互的智能性和交通工具的语音唤起的适用性。

Description

语音唤起方法、装置和设备
本申请要求2016年04月20日递交的申请号为201610246576.5、发明名称为“语音唤起方法、装置和设备”的中国专利申请的优先权,其全部内容通过引用结合在本申请中。
技术领域
本申请涉及互联网技术,尤其涉及一种应用于交通工具的语音唤起方法、装置和设备。
背景技术
随着社会经济的不断发展,有车一族越来越多,人们对车辆的要求也越来越高。为了满足用户的需求,各大车辆生产商在车辆的智能化上投入了很大的研发成本,以使得车辆能够更加便捷于人们的生活。
目前,车辆中的语音唤起是常见的车辆与用户的交互行为,例如,当用户有来电时,用户可以通过唤起车载语音功能来接听电话,无需驾驶员手持手机,从而避免了危险驾驶等问题。
但是,目前的语音唤起需要用户手动进行唤起,并且所针对的对象较为单一,适用性不高,使得人机交互不够智能化。
发明内容
本申请提供一种语音唤起方法、装置和设备,以解决现有技术的语音唤起需要用户手动唤起,且针对的对象较为单一,适用性不高,人机交互不够智能化的技术问题。
一个方面,本申请提供一种语音唤起方法,包括:
获取语音交互场景;
根据所述语音交互场景确定对应的第一语音,并向用户播报所述第一语音,所述第一语音为提示用户与所述交通工具或者所述交通工具上的设备进行交互的语音;
接收用户根据所述第一语音输入的操作信息,并根据所述操作信息执行相应的操作。
上述所提供的语音唤起方法,通过获取语音交互场景,并通过该语音交互场景自动唤起交通工具的语音功能为用户播报当前语音交互场景对应的第一语音,以提示用户与交通工具或者交通工具上的设备进行交互,从而使得交通工具的系统在接收到用户根据 第一语音输入的操作信息后,根据该操作信息执行相应的操作,以保证与用户之间的交互行为,满足用户的交互要求。该方法在获取到语音交互场景时自动为用户唤起语音,为用户提供丰富的语音内容,便于用户与交通工具进行交互,无需用户手动触发来唤起交通工具的语音交互功能,并且提高了人机交互的智能性;另外,车载系统所获取的语音交互场景可以与交通工具行驶状态对应,由于交通工具行驶状态较为丰富,因此使得与交通工具行驶状态对应的语音交互场景也比较丰富,进而使得本申请能够针对不同的交通工具行驶状态主动唤起交通工具的语音功能,提高了交通工具的语音唤起的适用性。
作为一种可实现的方式,所述接收用户根据所述第一语音输入的操作信息,并根据所述操作信息执行相应的操作,具体包括:
接收用户根据所述第一语音所输入的用户操作指令;
根据所述用户操作指令,判断所述用户的操作是否为所述第一语音所指示用户执行的用户操作;
若是,则停止播报所述第一语音。
该方式通过第一语音向用户提示用户所需执行的用户操作,进而在用户根据第一语音的指示执行用户操作后,根据用户输入的用户操作指令判断用户的操作是否为第一语音所指示用户执行的用户操作,并在交通工具的系统检测到用户执行的操作是第一语音所指示用户执行的用户操作时,停止播放第一语音,使得用户能够直观的获知交通工具的需求,便于用户及时发现交通工具的异常或者用户的不当操作,进一步提高了用户的体验。
作为一种可实现的方式,所述接收用户根据所述第一语音输入的操作信息,并根据所述操作信息执行相应的操作,具体包括:
接收用户根据所述第一语音所提示的内容输入的第二语音,所述第一语音用于提示用户针对所述第一语音所提示的内容进行选择操作,所述第二语音为用户的选择结果;
根据所述第二语音执行相应的操作。
该方式通过第一语音指示用户针对第一语音所提示的内容进行选择操作,从而获得用户根据第一语音输入的第二语音,并根据该第二语音执行相应的操作,使得人机交互更加的智能化;并且,本发明实施例提供的方法,可以预先获知用户潜在的触发语音的需求,主动为用户唤起语音,并提示用户待办事件,避免用户遗忘待办事件,给用户提供了便捷,进一步提高了用户的体验。
作为一种可实现的方式,所述获取语音交互场景,具体包括:
获取与所述交通工具行驶相关的参数信息;其中,所述参数信息包括外界行驶环境信息、交通工具状态信息和用户行为信息中的至少一类;
根据所述参数信息确定交通工具当前行驶状态对应的语音交互场景。
可选的,所述外界行驶环境信息包括路况信息和/或天气信息,所述交通工具状态信息包括交通工具自身状况信息和/或交通工具警告音信息。可选的,作为一种可实现的方式,所述交通工具警告音信息包括所述交通工具警告音的类型或者所述交通工具警告音的播报次数。
该方式通过采集与交通工具行驶相关的参数信息,并根据该参数信息确定交通工具当前行驶状态对应的语音交互场景,从而根据该语音交互场景主动唤起交通工具的语音功能,与用户进行交互,即通过参数信息的多样性,保证了语音交互场景的多样性,进而丰富了主动唤起交通工具语音功能的方式,提高了交通工具的语音唤起的适用性。
作为一种可实现的方式,所述参数信息包括用户行为信息,所述获取与交通工具行驶相关的参数信息,具体包括:
向存储所述用户行为信息的设备发送获取指令,所述获取指令携带所述用户预设的授权码或者所述交通工具的标识;
接收所述设备在验证所述授权码或者所述交通工具的标识合法后发送的用户行为信息。
该方式通过交通工具的系统(例如车载系统)向存储用户行为信息的设备发送携带用户预设的授权码或者所述交通工具的标识的获取指令,使得存储用户行为信息的设备在验证授权码或者交通工具的标识合法后,将用户行为信息发送给交通工具的系统,保证了用户行为信息的隐私安全。
作为一种可实现的方式,所述参数信息包括用户行为信息,所述获取与交通工具行驶相关的参数信息,具体包括:
接收用户输入的用户行为信息。
可选的,所述用户行为信息包括用户待办事件。
作为一种可实现的方式,所述用户待办事件包括所述用户待办事件的发生时刻,则所述向用户播报所述第一语音,具体包括:
根据所述用户待办事件的发生时刻,确定向所述用户播报所述第一语音的时刻;
在所述时刻到达时,向所述用户播报所述第一语音。
该方式通过确定播报第一语音的时刻,在时刻到达时向用户播报第一语音,其可以 预先获知用户潜在的触发语音的需求,主动为用户唤起语音,并提示用户待办事件,避免用户遗忘待办事件,给用户提供了便捷,进一步提高了用户的体验。
另一方面,本申请提供一种语音唤起装置,包括:
获取模块,用于获取语音交互场景;
确定模块,用于根据所述获取模块获取的所述语音交互场景确定对应的第一语音,所述第一语音为提示用户与所述交通工具或者所述交通工具上的设备进行交互的语音;
语音播报模块,用于向用户播报所述确定模块确定的所述第一语音;
接收模块,用于接收用户根据所述确定模块确定的所述第一语音输入的操作信息;
处理模块,用于根据所述接收模块接收到的所述操作信息执行相应的操作。
作为一种可实现的方式,所述接收模块,具体用于接收用户根据所述第一语音所输入的用户操作指令;
所述处理模块,具体用于根据所述用户操作指令,判断所述用户的操作是否为所述第一语音所指示用户执行的用户操作,并在判断所述用户的操作为所述第一语音所指示用户执行的用户操作时,指示所述语音播报模块停止播报所述第一语音。
作为一种可实现的方式,所述接收模块,具体用于接收用户根据所述第一语音所提示的内容输入的第二语音,所述第一语音用于提示用户针对所述第一语音所提示的内容进行选择操作,所述第二语音为用户的选择结果;
所述处理模块,具体用于根据所述第二语音执行相应的操作。
作为一种可实现的方式,所述获取模块,包括:
获取子模块,用于获取与交通工具行驶相关的参数信息;其中,所述参数信息包括外界行驶环境信息、交通工具状态信息和用户行为信息中的至少一类;
确定子模块,用于根据所述参数信息确定交通工具当前行驶状态对应的语音交互场景。
作为一种可实现的方式,所述外界行驶环境信息包括路况信息和/或天气信息,所述交通工具状态信息包括交通工具自身状况信息和/或交通工具警告音信息。
作为一种可实现的方式,所述交通工具警告音信息包括所述交通工具警告音的类型或者所述交通工具警告音的播报次数。
作为一种可实现的方式,所述参数信息包括用户行为信息,所述获取子模块,具体包括:
发送单元,用于向存储所述用户行为信息的设备发送获取指令,所述获取指令携带 所述用户预设的授权码或者所述交通工具的标识;
接收单元,用于接收所述设备在验证所述授权码或者所述交通工具的标识合法后发送的用户行为信息。
作为一种可实现的方式,所述参数信息包括用户行为信息,所述获取子模块,具体用于接收用户输入的用户行为信息。
作为一种可实现的方式,所述用户行为信息包括用户待办事件。
作为一种可实现的方式,所述用户待办事件包括所述用户待办事件的发生时刻;所述语音播报模块,包括:
确定单元,用于根据所述用户待办事件的发生时刻,确定向所述用户播报所述第一语音的时刻;
播报单元,用于在所述确定单元确定的所述时刻到达时,向所述用户播报所述第一语音。
上述各可实现方式所提供的语音唤起装置,其有益效果可以参照上述各可实现方式中的语音唤起方法所带来的有益效果,在此不再赘述。
另一方面,本申请提供一种语音唤起设备,包括:
处理器,用于获取语音交互场景,并根据所述语音交互场景确定对应的第一语音;
输出设备,耦合至所述处理器,用于向用户播报所述第一语音,所述第一语音为提示用户与所述交通工具或者所述交通工具上的设备进行交互的语音;
输入设备,耦合至所述处理器,用于接收用户根据所述第一语音输入的操作信息;
所述处理器,还用于根据所述输入设备获得的所述操作信息执行相应的操作。
作为一种可实现的方式,所述输入设备,具体用于接收用户根据所述第一语音所输入的用户操作指令;
所述处理器,具体用于根据所述用户操作指令,判断所述用户的操作是否为所述第一语音所指示用户执行的用户操作,并在判断所述用户的操作为所述第一语音所指示用户执行的用户操作时,指示所述输出设备停止播报所述第一语音。
作为一种可实现的方式,所述输入设备,具体用于接收用户根据所述第一语音所提示的内容输入的第二语音,所述第一语音用于提示用户针对所述第一语音所提示的内容进行选择操作,所述第二语音为用户的选择结果;
所述处理器,具体用于根据所述第二语音执行相应的操作。
作为一种可实现的方式,所述输入设备,还用于获取与交通工具行驶相关的参数信 息;其中,所述参数信息包括外界行驶环境信息、交通工具状态信息和用户行为信息中的至少一类;
所述处理器,具体用于根据所述参数信息确定交通工具当前行驶状态对应的语音交互场景。
作为一种可实现的方式,所述外界行驶环境信息包括路况信息和/或天气信息,所述交通工具状态信息包括交通工具自身状况信息和/或交通工具警告音信息。
作为一种可实现的方式,所述交通工具警告音信息包括所述交通工具警告音的类型或者所述交通工具警告音的播报次数。
作为一种可实现的方式,所述参数信息包括用户行为信息;
所述输出设备,还用于向存储所述用户行为信息的设备发送获取指令,所述获取指令携带所述用户预设的授权码或者所述交通工具的标识;
所述输入设备,具体用于接收所述设备在验证所述授权码或者所述交通工具的标识合法后发送的用户行为信息。
作为一种可实现的方式,所述参数信息包括用户行为信息,所述输入设备,具体用于接收用户输入的用户行为信息。
作为一种可实现的方式,所述用户行为信息包括用户待办事件。
作为一种可实现的方式,所述用户待办事件包括所述用户待办事件的发生时刻;
所述处理器,还用于根据所述用户待办事件的发生时刻,确定向所述用户播报所述第一语音的时刻;
所述输出设备,具体用于在所述时刻到达时,向所述用户播报所述第一语音。
上述各可实现方式所提供的语音唤起设备,其有益效果可以参照上述各可实现方式中的语音唤起方法所带来的有益效果,在此不再赘述。
另一方面,本申请提供一种用于交通工具的语音唤起设备,包括:机载处理器、机载输出设备和机载输入设备;
所述机载处理器,用于获取语音交互场景,并根据所述语音交互场景确定对应的第一语音;
所述机载输出设备,耦合至所述机载处理器,用于向用户播报所述第一语音,所述第一语音为提示用户与所述交通工具或者所述交通工具上的设备进行交互的语音;
所述机载输入设备,耦合至所述机载处理器,用于接收用户根据所述第一语音输入的操作信息;
所述机载处理器,还用于根据所述机载输入设备获得的所述操作信息执行相应的操作。
另一方面,本申请提供一种车载互联网操作系统,包括:
语音控制单元,根据所获取的语音交互场景确定对应的第一语音并向用户进行播报;
操作控制单元,根据车载输入设备获取的操作信息控制语音唤起系统执行相应的操作;其中,所述操作信息为用户根据所述第一语音输入给所述车载输入设备的。
在本申请中,通过获取交通工具当前行驶状态所对应的语音交互场景,并通过该语音交互场景自动唤起交通工具的语音功能为用户播报当前语音交互场景对应的第一语音,以提示用户与交通工具或者交通工具上的设备进行交互,从而使得交通工具的系统在接收到用户根据第一语音输入的操作信息后,根据该操作信息执行相应的操作,以保证与用户之间的交互行为,满足用户的交互要求。该申请在获取到语音交互场景时自动为用户唤起语音,为用户提供丰富的语音内容,便于用户与交通工具进行交互,无需用户手动触发来唤起交通工具的语音交互功能,并且提高了人机交互的智能性;另外,所获取的语音交互场景可以与交通工具行驶状态对应,由于交通工具行驶状态较为丰富,因此使得与交通工具行驶状态对应的语音交互场景也比较丰富,进而使得本申请能够针对不同的交通工具行驶状态主动唤起交通工具的语音功能,提高了交通工具的语音唤起的适用性。
附图说明
为了更清楚地说明本申请实施例或现有技术中的技术方案,下面将对实施例或现有技术描述中所需要使用的附图作一简单地介绍,显而易见地,下面描述中的附图是本发明的一些实施例,对于本领域普通技术人员来讲,在不付出创造性劳动性的前提下,还可以根据这些附图获得其他的附图。
图1为本申请的一种可选的组网方式的示意图;
图2为本申请一实施例提供的语音唤起方法的流程示意图;
图3为本申请一实施例提供的语音唤起方法的流程示意图;
图4为本申请一实施例提供的语音唤起方法的流程示意图;
图5为本申请一实施例提供的语音唤起方法的流程示意图;
图6为本申请一实施例提供的语音唤起方法的流程示意图;
图7为本申请一实施例提供的语音唤起方法的流程示意图;
图8为本申请一实施例提供的语音唤起方法的流程示意图;
图9为本申请一实施例提供的语音唤起装置结构示意图;
图10为本申请一实施例提供的语音唤起装置结构示意图;
图11为本申请一实施例提供的语音唤起装置结构示意图;
图12为本申请一实施例提供的语音唤起设备硬件结构示意图;
图13为本申请一实施例提供的车载系统结构示意图;
图14为本申请一实施例提供的车载互联网操作系统的结构示意图。
具体实施方式
这里将详细地对示例性实施例进行说明,其示例表示在附图中。下面的描述涉及附图时,除非另有表示,不同附图中的相同数字表示相同或相似的要素。以下示例性实施例中所描述的实施方式并不代表与本发明相一致的所有实施方式。相反,它们仅是与如所附权利要求书中所详述的、本发明的一些方面相一致的装置和方法的例子。
本申请实施例所涉及的交通工具,包括但不限于,内燃机汽车或摩托车、电动汽车或摩托车、电动助力车、电动平衡车、遥控车辆等车辆、小型飞行器(例如,无人驾驶飞行器、有人小型飞行器、遥控飞行器)、以及各种变形。这里所涉及的车辆可以为单一的油路车辆、还可以是单一的汽路车辆、还可以是油汽结合的车辆、还可以是助力的电动车辆,本申请实施例对车辆的类型并不做限定,该车辆具有相应的车载系统。下述实施例均以交通工具是车辆为例来进行说明。
本申请实施例涉及的语音唤起方法、装置和设备,可以适用于图1所示的组网方式,该图1中的网络架构可以包括车辆的车载系统和无线网络。该车载系统中可以设置用户操作接口,可选的,该用户操作接口可以是供用户输入的语音接口、还可以是接收用户手动触发的操作指令的接口,例如U盘接口、USB接口、安全带插入口等等。该车辆可以与无线网络建立连接,可选的,该无线网络可以是2G网络、3G网络、4G网络或者5G网络、无线保真(Wireless Fidelity,简称WIFI)网络等,可选的,该无线网络还可以是物联网或者车联网,本申请实施例对网络的具体类型或者具体形式并不做限定,只要其能够提供供车辆访问并获取数据的接口即可。车载系统可以该无线网络访问不同的网络服务器,例如邮箱服务器、短信服务器、云服务器等等,图1中仅示出了三种服务器,但并不以此为限。
可选的,本申请实施例所涉及的方法,其执行主体可以是车载系统,可选的,该车 载系统可以是车辆上的车机所集成的系统,例如车载导航系统和/或车载娱乐系统,还可以是包含车机和车辆其他设备例如传感器等的系统,该车载系统可以与车辆和用户进行交互。本申请实施例对车载系统的具体内容并不做限制,只要其能够主动唤起车辆的语音功能即可。下述实施例中均以执行主体是车载系统为例进行说明,但本申请实施例的执行主体并不以此为限。
本申请实施例所涉及的语音唤起方法、装置和设备,旨在解决现有技术中的语音唤起需要用户手动唤起,且针对的对象较为单一,适用性不高,人机交互不够智能化的技术问题。
下面以具体地实施例对本申请的技术方案进行详细说明。下面这几个具体的实施例可以相互结合,对于相同或相似的概念或过程可能在某些实施例中不再赘述。
图2为本申请一实施例提供的语音唤起方法的流程示意图。本实施例涉及的是车载系统根据所确定的语音交互场景主动向用户播报与当前语音交互场景对应的第一语音,并在接收到用户根据该第一语音所提示的内容输入的操作信息后执行相应的操作,以确保和用户的主动交互,为用户提供便捷的具体过程。如图2所示,该方法可以包括:
S101:获取语音交互场景。
具体的,以交通工具是车辆为例,本申请实施例所涉及的语音交互场景可以是能够主动触发车载系统唤起语音功能为用户播报语音,与用户进行交互的场景。可选的,该语音交互场景可以为车载系统能够主动向用户语音提示执行相应操作的场景,例如,主动通过语音向用户提示车辆异常信息(例如车辆后备箱未关闭)以使用户消除车辆的异常的语音场景。可选的,该语音交互场景还可以是车载系统通过学习或者获取一些与车辆行驶相关的参数信息(例如车辆经常经过某个地理位置)来发掘用户潜在的语音交互需求,主动向用户发起提示或者询问并和用户进行语音交互的场景,例如,用户一般是8点从家里出发,8点半沿途会去便利店买早点,车载系统可以学习该车辆到该车辆行驶相关的位置信息,当用户8点出发的时候,可以主动询问用户是否8点半要沿途去该便利店买早点,进而根据用户的语音回答主动为用户导航去便利店或者还可以主动为用户播放舒缓的音乐等。再例如,车载系统可以在经过用户授权的情况下登录用户的邮箱服务器获取用户的一些邮件,假设邮件中存在用户在某个时刻开车去某一个地方,则在用户上车后,车载系统可以主动发起对用户的语音询问,即询问用户是否要去该地方开会,从而根据所得到的用户的语音回答来执行相应的操作。即本申请实施例涉及的技术方案,无需用户手动触发来唤起车辆的语音交互功能,而是根本不同的语音交互场景自动为用 户唤起语音,为用户提供丰富的语音内容,便于用户与车辆进行交互,从而提高人机交互的智能性。
可选的,本申请实施例中的语音交互场景可以通过任意的方式确定,例如,该语音交互场景可以通过车辆的行驶状态确定。该车辆的行驶状态可以是车辆的周围环境信息、车辆的车速等,还可以是车辆内部是否存在用户操作、车辆自身是否存在不安全因素等,本申请实施例对车辆的行驶状态的类型并不做限定。可选的,不同的车辆行驶状态可以对应不同的语音交互场景,而无论是哪一种车辆的行驶状态,只要车载系统能够获取到与该车辆行驶状态对应的语音交互场景,均可以主动唤起车辆的语音功能,即本申请实施例中的自动唤起车辆语音功能的语音交互场景比较丰富,适用性高。当然,这里通过车辆的行驶状态确定语音交互场景只是一种示例,本申请实施例并不以此为限。
S102:根据所述语音交互场景确定对应的第一语音,并向用户播报所述第一语音,所述第一语音为提示用户与交通工具或者所述交通工具上的设备进行交互的语音。
具体的,当车载系统获取到上述语音交互场景之后,就可以确定出该语音交互场景对应的第一语音。可选的,车载系统可以预先建立不同的语音交互场景和不同的播报语音之间的映射关系,一旦车载系统确定了当前行驶状态所对应的语音交互场景之后,就会主动为用户播报该第一语音。可选的,上述语音交互场景的形式可以为信息形式,因此车载系统也可以对该语音交互场景进行关键字识别,并将所识别出的关键字通过语音合成技术组成第一语音,向用户播报,本申请实施例对车载系统根据所获取的语音交互场景确定第一语音的方式并不做限定。
需要说明的是,上述第一语音的语音内容为提示用户与车辆或者车辆上的设备进行交互的语音内容,这里所说的“交互”,可以是用户与车辆或者车辆上的设备进行的语音交互,还可以是用户与车辆或者车辆上的设备的手动操作式的交互,还可以是用户与车辆或者车辆上的设备的其他方式的交互。可选的,该车辆上的设备可以是车载收音机、行车记录仪或者其他的车载设备等。
S103:接收用户根据所述第一语音输入的操作信息,并根据所述操作信息执行相应的操作。
具体的,当用户接收到车载系统播报的第一语音之后,根据第一语音的内容执行相应的用户操作,该用户操作可以是用户的语音输入、可以是用户的手动操作。当车载系统接收到用户输入的操作信息之后,根据该操作信息执行相应的操作,以保证与用户之间的交互行为,满足用户的交互要求。
本申请实施例提供的语音唤起方法,通过获取语音交互场景,并通过该语音交互场景自动唤起交通工具的语音功能为用户播报当前语音交互场景对应的第一语音,以提示用户与交通工具或者交通工具上的设备进行交互,从而使得交通工具的系统在接收到用户根据第一语音输入的操作信息后,根据该操作信息执行相应的操作,以保证与用户之间的交互行为,满足用户的交互要求。本申请实施例提供的方法,在获取到语音交互场景时自动为用户唤起语音,为用户提供丰富的语音内容,便于用户与交通工具或者交通工具上的设备进行交互,无需用户手动触发来唤起交通工具的语音交互功能,并且提高了人机交互的智能性;另外,上述所获取的语音交互场景可以与交通工具行驶状态对应,由于交通工具行驶状态较为丰富,因此使得所确定的语音交互场景也比较丰富,进而使得本申请能够针对不同的交通工具行驶状态主动唤起交通工具的语音功能,提高了交通工具的语音唤起的适用性。
图3为本申请一实施例提供的语音唤起方法的流程示意图。本实施例涉及的是获取语音交互场景的具体过程。在上述图2所示实施例的基础上,进一步地,如图3所示,上述S101具体可以包括:
S201:获取与交通工具行驶相关的参数信息;其中,所述参数信息包括外界行驶环境信息、交通工具状态信息和用户行为信息中的至少一类。
具体的,继续以交通工具是车辆为例,上述外界行驶环境信息可以是外界行车环境信息、交通工具状态信息可以是车辆状态信息,所涉及的交通工具的自身状况信息可以是车况信息,所涉及的交通工具警告音信息可以是车辆警告音信息。因此,车载系统可以实时的获取与车辆行驶相关的参数信息,该参数信息可以包括外界行车环境信息、车辆状态信息和用户行为信息中的至少一类。可选的,该外界行车环境信息包括路况信息和/或天气信息,该车辆状态信息包括车况信息和/或车辆警告音信息,该用户行为信息可以包括用户的待办事件、还可以包括用户针对车辆的某项操作(例如用户在车辆的USB接口中插入了U盘)、还可以包括用户的某项行为习惯(例如用户习惯在某一个时刻做同样的事情)等。
可选的,上述路况信息可以包括道路是否拥堵、或者前方道路是否发生交通事故或者、该路段是否有监控摄像头等信息,上述车况信息可以包括车辆的自身设施条件信息,例如发动机是否故障、刹车片是否良好等信息,上述车辆警告音信息可以包括车辆警告音的音量信息、车辆警告音的音色信息等。可选的,该车辆警告音信息还可以是车辆警 告音的类型或者车辆警告音的播报次数。
可选的,当上述参数信息为路况信息时,车载系统可以通过图1所示的无线网络与交通道路监控服务器建立通信连接,获取路况信息,还可以通过电台广播获取路况信息;可选的,当上述参数信息为天气信息时,车载系统可以通过图1所示的无线网络中的网络服务器获取天气信息;当上述参数信息为车况信息时,车载系统可以通过车辆自身的不同的传感器或者软件检测获得车况信息;当上述参数信息为车辆警告音信息时,车载系统可以通过车载音响和处理器的配合获取该车辆警告音信息。
可选的,当上述参数信息包括用户行为信息时,车载系统获取该用户行为信息的具体方式可以包括两种可能的实施方式,具体如下:
第一种可能的实施方式:参见图4所示,上述S201具体可以包括:
S301:向存储所述用户行为信息的设备发送获取指令,所述获取指令携带所述用户预设的授权码或者所述交通工具的标识。可选的,该交通工具的标识可以是车辆的标识。
具体的,该存储用户行为信息的设备可以是用户的终端、可以是网络中的各种服务器,例如邮箱服务器、短信服务器、云端服务器等。可选的,用户可以预先给车辆配置一授权码,授权该车辆可以从存储用户行为信息的设备中获取用户的行为信息。可选的,用户也可以不用为车辆配置授权码,用户可以在存储用户行为信息的设备中注册或者预留自己的车辆的标识,当该标识对应的车辆访问存储用户行为信息的设备时,就可以从存储用户行为信息的设备中获取用户的行为信息。基于此,车载系统就可以通过向存储用户行为信息的设备发送获取指令,来请求获取用户行为信息。
S302:接收所述设备在验证所述授权码或者所述交通工具的标识合法后发送的用户行为信息。
具体的,当存储用户行为信息的设备接收到车辆的获取指令后,结合获取指令中所携带的内容进行相应的合法判断,即判断获取指令中的授权码是否为用户预先配置的授权码或者判断获取指令中携带的车辆的标识是否为用户预先授权的车辆的标识。当存储用户行为信息的设备确定获取指令中的授权码或者车辆的标识合法后,就可以将用户行为信息发送给车载系统,保证了用户行为信息的隐私安全。
第二种可能的实施方式:车载系统可以直接接收用户输入的用户行为信息。
在该可能的实施方式中,车载系统可以向用户提供一用户输入接口,该用户输入接口可以是设备访问接口、还可以是语音输入接口、还可以是显示屏的手写接口等。因此, 用户可以通过在设备访问接口中插入存储自己行为信息的终端或者U盘等设备,使得车载系统可以读取到用户行为信息,还可以是通过语音输入接口获取到用户输入的语音,通过相应的语音识别技术获取到用户行为信息,还可以通过手写接口获取用户手写输入的用户行为信息。
至此,通过上述第一种可能的实施方式或者第二种可能的实施方式,车载系统就可以获得用户行为信息。
S202:根据所述参数信息确定交通工具当前行驶状态对应的语音交互场景。
具体的,当车载系统获得上述参数信息之后,车载系统就可以确定当前车辆行驶状态,进而确定车辆当前行驶状态对应的语音交互场景。可选的,车载系统可以通过预设的第一映射关系确定车辆当前行驶状态对应的语音交互场景,该第一映射关系中可以包括不同的参数信息与不同的车辆行驶状态以及不同的语音交互场景之间的对应关系。本申请实施例对根据参数信息如何确定的车辆当前行驶状态对应的语音交互场景的方式并不做限定,只要能够根据该参数信息确定出车辆当前行驶状态对应的语音交互场景即可。
本申请实施例提供的语音唤起方法,通过采集与交通工具行驶相关的参数信息,并根据该参数信息确定交通工具当前行驶状态对应的语音交互场景,从而根据该语音交互场景主动唤起交通工具的语音功能,与用户进行交互。本申请实施例的技术方案,通过参数信息的多样性,保证了语音交互场景的多样性,进而丰富了主动唤起交通工具语音功能的方式,提高了交通工具的语音唤起的适用性。
图5为本申请一实施例提供的语音唤起方法的流程示意图。本实施例涉及的是用户根据第一语音所提示的内容向车载系统输入用户操作指令,车载系统根据该用户操作指令确定是否停止播报第一语音的具体过程。在上述实施例的基础上,进一步地,如图5所示,上述S103具体可以包括:
S401:接收用户根据所述第一语音所输入的用户操作指令。
具体的,继续以交通工具是车辆为例,本实施例中的第一语音用于指示用户执行对应的用户操作。故,当用户接收到车载系统播报的第一语音之后,根据该第一语音执行该第一语音所指示的用户操作。对于车载系统而言,当用户执行了该用户操作之后,车载系统通过车辆与用户之间的交互接口、或者车辆上的设备与用户之间的交互接口就可以检测到该用户操作对应的操作指令。
例如,当车辆行驶过程中用户没有系安全带,则车辆内部就会响起beep音(该beep 音为车辆警告音),即当前车辆行驶状态为beep音响起,则车载系统确定该车辆行驶状态对应的语音交互场景为安全带提示场景,进而车载系统根据该安全带提示场景确定该场景对应的第一语音为“请系好安全带”,并向用户播报;当用户接收到该第一语音之后,按照该第一语音的内容系好安全带,由此,车载系统就可以通过车辆提供的安全带插入口检测到用户的操作,进而获取到该用户操作对应的操作指令为“已插入安全带”。
再例如,当车辆行驶过程中用户没有关好左侧车门,则车辆内部也会响起beep音(该beep音与上述例子中的beep音类型可以不同),即当前车辆行驶状态为beep音响起,则车载系统确定该车辆行驶状态对应的语音交互场景为左侧车门异常场景,进而车载系统根据该车门异常场景确定该场景对应的第一语音为“请关好左侧车门”,并向用户播报;当用户接收到该第一语音之后,按照该第一语音的内容关闭好左侧车门,由此,车载系统就可以通过车辆提供的车门检测接口检测到用户的操作,进而获取到该用户操作对应的操作指令为“已关闭好车门”。
无论是上述哪一种方式,在当前车辆行驶状态为beep音响起时,车载系统均可以确定该当前车辆行驶状态对应的语音交互场景,进而确定该语音交互场景对应的第一语音,即通过文本到语音(Text To Speech,简称TTS)的方式向用户提示明确的语音内容,避免了用户在车内响起beep音时无法获知该beep音的含义,本申请实施例的方案可以使得用户直观有效的获知beep音的含义,提高了用户的体验。可选的,车载系统可以通过车辆警告音的类型或者车辆警告音的次数来区分所响起的车辆警告音应该对应的语音交互场景,例如,当车辆警告音为“滴滴”时,车载系统确定其对应的语音交互场景应该是安全带提示场景,当车辆警告音为“哒哒”时,车载系统确定其对应的语音交互场景为车门异常场景。
需要说明的是,对于上述beep响起,车载系统向用户播报直观的第一语音只是一种举例。可选的,当车辆行驶状态为车辆发生其他异常,例如,车辆的机油量低、车辆的发动机过热、车辆的玻璃水耗尽等等,车载系统也会根据当前行驶状态确定其对应的语音交互场景,从而向用户播放与该语音交互场景对应的第一语音,以提示用户执行该第一语音所指示的用户的操作,使得用户可以直观有效的获知车辆的需求,便于用户及时发现车辆的异常或者用户的不当操作,进一步提高了用户的体验。
S402:根据所述用户操作指令,判断所述用户的操作是否为所述第一语音所指示用户执行的用户操作。若是,则执行S403,若否,则执行S404。
S403:停止播报所述第一语音。
S404:按照预设周期播放第一语音。
具体的,当车载系统检测到用户操作指令后,根据该用户操作指令判断用户当前所执行的操作是否为第一语音所指示用户执行的用户操作;当车载系统检测到用户当前执行的操作为第一语音所指示用户执行的用户操作,则车载系统停止播报第一语音,当车载系统检测到用户当前执行的操作不是第一语音所指示用户执行的用户操作,则车载系统按照预设周期播放该第一语音,该预设周期可以是间隔数秒播放一次第一语音。可选的,当上述第一语音针对的车辆行驶状态是beep音响起,则在车载系统检测到用户当前执行的操作不是第一语音所指示用户执行的用户操作时,也可以间隔数秒重复响起该beep音。
本申请实施例提供的语音唤起方法,通过第一语音向用户提示用户所需执行的用户操作,进而在用户根据第一语音的指示执行用户操作后,根据用户输入的用户操作指令判断用户的操作是否为第一语音所指示用户执行的用户操作,并在交通工具的系统检测到用户执行的操作是第一语音所指示用户执行的用户操作时,停止播放第一语音。本申请实施例提供的方法,可以使得用户能够直观的获知交通工具的需求,便于用户及时发现交通工具的异常或者用户的不当操作,进一步提高了用户的体验。
图6为本申请一实施例提供的语音唤起方法的流程示意图。本实施例涉及的是用户根据第一语音所提示的内容向车载系统输入第二语音,车载系统根据该第二语音与用户进行交互的具体过程。本实施例中,第一语音用于提示用户针对所述第一语音所提示的内容进行选择操作,第二语音为用户的选择结果。在上述实施例的基础上,进一步地,如图6所示,上述S103具体可以包括:
S501:接收用户根据所述第一语音所提示的内容输入的第二语音。
具体的,继续以交通工具是车辆为例,本实施例中的第一语音用于提示用户针对第一语音所提示的内容进行选择操作。故,当用户接收到车载系统播报的第一语音之后,根据该第一语音所提示的内容进行选择,并将选择结果通过语音的方式输入给车载系统,该选择结果即就是上述第二语音。例如,当车载系统确定当前行驶状态对应的语音交互场景为驾驶模式提示场景,则车载系统向用户播报第一语音“是否切换驾驶模式为经济模式”,则用户针对该第一语音可以输入选择结果,例如,用户可以是输入“是”或者“不,我要选择舒适模式”。再例如,假设车载系统获取到的参数信息为用户待办事件,设该用户待办事件为“九点开车去机场”,车载系统所确定的第一语音为“现在是否开 车去机场”,这样当用户一旦启动车辆之后,车载系统就会播报该第一语音,提示用户该待办事件。
S502:根据所述第二语音执行相应的操作。
具体的,车载系统在接收到用户输入的第二语音之后,根据结合内部的匹配机制,确定该第二语音对应的操作,然后执行该操作,以保证车辆与用户之间的交互。继续以S501中所举例的例子为例,假设用户输入的第二语音为“不,我要选择舒适模式”,则车载系统可以通过语音识别技术获知第二语音的内容,然后根据该第二语音的内容将驾驶模式切换至舒适模式,从而满足用户与车辆的交互需求。
可选的,当车载系统获取的参数信息为用户待办事件,且该用户待办事件中包括用户待办事件的发生时刻,则在车载系统确定该参数信息对应的第一语音之后,如图7所示,上述向用户播报所述第一语音,具体可以包括:
S601:根据所述用户待办事件的发生时刻,确定向所述用户播报所述第一语音的时刻。
S602:在所述时刻到达时,向所述用户播报所述第一语音。
具体的,当车载系统获取的用户待办事件中包括用户待办事件的发生时刻时,车载系统可以根据该发生时刻确定向用户播报第一语音的时刻,并在该时刻到达时,向用户播报第一语音。可选的,若在车辆启动时播报第一语音的时刻还未到达,则不需要向用户播报第一语音,在该时刻到达时开始播播报第一语音,或者,若播报第一语音的时刻已经到达车辆还未启动,则车载系统可以通过向用户的终端发送该第一语音,通过用户的终端向用户播报该第一语音,以起到提示用户的作用。
为了更清楚的介绍本实施例的方案,下面举一个具体的例子,参见图8所示的示例图,图8中,假设车载系统获取的参数信息为用户待办事件,该用户待办事件包括该待办事件的发生时刻,设该用户待办事件为“用户小X在9点在A位置有会议”,车载系统确定播报第一语音的时刻为8点30分,车载系统确定的第一语音为“半小时后在A位置有会议,预计20分钟到达,现在是否前往”。如图8所示,该方法具体包括:
S701:车载系统确定参数信息为“用户小X在9点在A位置有会议”。
S702:车载系统根据该参数信息确定当前行驶状态对应的语音交互场景为“9点在A位置有会议”。
S703:车载系统根据该语音交互场景确定播放第一语音的时刻为8点30分。
S704:车载系统在8点30分到达时,向用户播报第一语音“半小时后在A位置有 会议,预计20分钟到达,现在是否前往”。
S705:用户根据第一语音向车载系统输入第二语音“是的,现在去”。
S706:车载系统向用户播报“好的,现在为您导航去A位置”。
可选的,当用户行车途中,假设车载系统又实时采集到另一参数信息,该参数信息假设为路况信息,设该路况信息为“前方道路拥堵”,则车载系统确定当前行驶状态对应的语音交互场景为“拥堵场景”,进而车载系统确定该拥堵场景对应的第一语音为“前方拥堵,去A位置可能迟到,是否电话通知会议发起人小Y”,这样用户在接收到第一语音后,可以进行选择并向车载系统输入第二语音,假设该第二语音为“是”,则车载系统就可以根据该第二语音,主动向会议发起人发起呼叫请求。
本申请实施例提供的语音唤起方法,通过第一语音指示用户针对第一语音所提示的内容进行选择操作,从而获得用户根据第一语音输入的第二语音,并根据该第二语音执行相应的操作,使得人机交互更加的智能化;并且,本申请实施例提供的方法,可以预先获知用户潜在的触发语音的需求,主动为用户唤起语音,并提示用户待办事件,避免用户遗忘待办事件,给用户提供了便捷,进一步提高了用户的体验。
以下将详细描述根据本申请的一个或多个实施例的语音唤起装置。该语音唤起装置可以被实现在交通工具的基础架构中,也可以被实现在交通工具和无线网络的交互系统中。本领域技术人员可以理解,该语音唤起装置均可使用市售的硬件组件通过本方案所教导的步骤进行配置来构成。例如,处理器组件(或处理模块、处理单元、确定模块等)可以使用来自德州仪器公司、英特尔公司、ARM公司等企业的单片机、微控制器、微处理器等组件实现。
下述为本申请装置实施例,可以用于执行本申请方法实施例。对于本申请装置实施例中未披露的细节,请参照本申请方法实施例。
图9为本申请一实施例提供的是语音唤起装置结构示意图,该语音唤起装置,可以通过软件、硬件或者两者的结合实现。如图9所示,该语音唤起装置,包括:获取模块10、确定模块11、语音播报模块12、接收模块13和处理模块14。
其中,获取模块10,用于获取语音交互场景;
确定模块11,用于根据所述获取模块10获取的所述语音交互场景确定对应的第一语音,所述第一语音为提示用户与所述交通工具或者所述交通工具上的设备进行交互的语音;
语音播报模块12,用于向用户播报所述确定模块11确定的所述第一语音;
接收模块13,用于接收用户根据所述确定模块11确定的所述第一语音输入的操作信息;
处理模块14,用于根据所述接收模块13接收到的所述操作信息执行相应的操作。
本申请实施例提供的语音唤起装置,可以执行上述方法实施例,其实现原理和技术效果类似,在此不再赘述。
在本申请实施例一种可能的实施方式中,所述接收模块13,具体用于接收用户根据所述第一语音所输入的用户操作指令;
所述处理模块14,具体用于根据所述用户操作指令,判断所述用户的操作是否为所述第一语音所指示用户执行的用户操作,并在判断所述用户的操作为所述第一语音所指示用户执行的用户操作时,指示所述语音播报模块12停止播报所述第一语音。
在本申请实施例另一种可能的实施方式中,所述接收模块13,具体用于接收用户根据所述第一语音所提示的内容输入的第二语音,所述第一语音用于提示用户针对所述第一语音所提示的内容进行选择操作,所述第二语音为用户的选择结果;
所述处理模块14,具体用于根据所述第二语音执行相应的操作。
进一步地,在上述图9所示实施例的基础上,参见图10提供的本申请一实施例提供的语音唤起装置结构示意图。在图10中,所述获取模块10,包括:
获取子模块101,用于获取与交通工具行驶相关的参数信息;其中,所述参数信息包括外界行驶环境信息、交通工具状态信息和用户行为信息中的至少一类;
确定子模块102,用于根据所述参数信息确定交通工具当前行驶状态对应的语音交互场景。
可选的,所述外界行车环境信息包括路况信息和/或天气信息,所述交通工具状态信息包括交通工具自身状况信息和/或交通工具警告音信息。可选的,所述交通工具警告音信息包括所述交通工具警告音的类型或者所述交通工具警告音的播报次数。
可选的,所述参数信息包括用户行为信息,则继续参照图10,上述获取子模块101,具体包括:发送单元1011,用于向存储所述用户行为信息的设备发送获取指令,所述获取指令携带所述用户预设的授权码或者所述交通工具的标识;
接收单元1012,用于接收所述设备在验证所述授权码或者所述交通工具的标识合法后发送的用户行为信息。
可选的,所述参数信息包括用户行为信息,上述获取子模块101,具体用于接收用户输入的用户行为信息。
可选的,所述用户行为信息包括用户待办事件。
可选的,上述用户待办事件包括所述用户待办事件的发生时刻,则参见图11提供的本申请一实施例提供的语音唤起装置结构示意图。在图11中,上述语音播报模块12,可以包括:
确定单元121,用于根据所述用户待办事件的发生时刻,确定向所述用户播报所述第一语音的时刻;
播报单元122,用于在所述确定单元121确定的所述时刻到达时,向所述用户播报所述第一语音。
本申请实施例提供的语音唤起装置,可以执行上述方法实施例,其实现原理和技术效果类似,在此不再赘述。
图12为本申请一实施例提供的是语音唤起设备硬件结构示意图。该语音唤起设备可以集成在上述实施例中的车载系统中,还可以是独立的车载系统。如图12所示,该语音唤起设备可以包括处理器20、输出设备21、输入设备22、存储器23和和至少一个通信总线24。通信总线24用于实现元件之间的通信连接。存储器23可能包含高速RAM存储器,也可能还包括非易失性存储NVM,例如至少一个磁盘存储器,存储器中可以存储各种程序,用于完成各种处理功能以及实现本实施例的方法步骤。
可选的,上述处理器20例如可以为中央处理器(Central Processing Unit,简称CPU)、应用专用集成电路(ASIC)、数字信号处理器(DSP)、数字信号处理设备(DSPD)、可编程逻辑器件(PLD)、现场可编程门阵列(FPGA)、控制器、微控制器、微处理器或其他电子元件实现,该处理器20通过车内线路或无线连接耦合到上述输入设备22和输出设备21。
可选的,上述输入设备22可以包括多种输入设备,例如可以包括面向用户的用户接口、面向设备的设备接口、收发信机中的至少一个。可选的,该面向设备的设备接口可以是用于设备与设备之间进行数据传输的有线接口、还可以是用于设备与设备之间进行数据或者指令传输的硬件插入接口(例如USB接口、串口、车体硬件设施之间的接口等);可选的,该面向用户的用户接口例如可以是面向用户的控制按键、用于接收语音输入的语音输入设备以及用户接收用户触摸输入的触摸感知设备(例如具有触摸感应功能的触摸屏、触控板等);可选的,上述收发信机可以是具有通信功能的射频收发芯片、基带处理芯片以及收发天线等。本申请实施例中的语音唤起设备为一通用的语音唤起设备, 其可以适用于任一的控制系统或者控制设备或者其他类型的设备。可选的,上述输出设备21可以为相应的具有通信功能的输出接口或者语音播放设备或者收发信机。
可选的,该语音唤起设备可以是用于交通工具的语音唤起设备,例如,可以是用于车辆的语音唤起设备、用于飞行器的语音唤起设备、用于水路运输工具的语音唤起设备等。关于用于交通工具的消语音唤起设备的具体内容,本申请提供了另一实施例来进行介绍,请参见后面的实施例,在此不再详述。
在本申请实施例中,处理器20,用于获取语音交互场景,并根据所述语音交互场景确定对应的第一语音;
输出设备21,耦合至所述处理器20,用于向用户播报所述第一语音,所述第一语音为提示用户与所述交通工具或者所述交通工具上的设备进行交互的语音;
输入设备22,耦合至所述处理器20,用于接收用户根据所述第一语音输入的操作信息;
所述处理器20,还用于根据所述输入设备22获得的所述操作信息执行相应的操作。
本申请实施例提供的语音唤起设备,可以执行上述方法实施例,其实现原理和技术效果类似,在此不再赘述。
可选的,所述输入设备22,具体用于接收用户根据所述第一语音所输入的用户操作指令;
所述处理器20,具体用于根据所述用户操作指令,判断所述用户的操作是否为所述第一语音所指示用户执行的用户操作,并在判断所述用户的操作为所述第一语音所指示用户执行的用户操作时,指示所述输出设备21停止播报所述第一语音。
可选的,所述输入设备22,具体用于接收用户根据所述第一语音所提示的内容输入的第二语音,所述第一语音用于提示用户针对所述第一语音所提示的内容进行选择操作,所述第二语音为用户的选择结果;
所述处理器20,具体用于根据所述第二语音执行相应的操作。
可选的,所述输入设备22,还用于获取与交通工具行驶相关的参数信息;其中,所述参数信息包括外界行驶环境信息、交通工具状态信息和用户行为信息中的至少一类;
所述处理器20,具体用于根据所述参数信息确定交通工具当前行驶状态对应的语音交互场景。
可选的,所述外界行车环境信息包括路况信息和/或天气信息,所述交通工具状态信息包括交通工具自身状况信息和/或交通工具警告音信息。可选的,所述交通工具警告音 信息包括所述交通工具警告音的类型或者所述交通工具警告音的播报次数。
可选的,所述参数信息包括用户行为信息;
所述输出设备21,还用于向存储所述用户行为信息的设备发送获取指令,所述获取指令携带所述用户预设的授权码或者所述交通工具的标识;
所述输入设备22,具体用于接收所述设备在验证所述授权码或者所述交通工具的标识合法后发送的用户行为信息。
可选的,所述参数信息包括用户行为信息,所述输入设备22,具体用于接收用户输入的用户行为信息。
可选的,所述用户行为信息包括用户待办事件。
可选的,所述用户待办事件包括所述用户待办事件的发生时刻;
所述处理器20,还用于根据所述用户待办事件的发生时刻,确定向所述用户播报所述第一语音的时刻;
所述输出设备21,具体用于在所述时刻到达时,向所述用户播报所述第一语音。
本申请实施例提供的语音唤起设备,可以执行上述方法实施例,其实现原理和技术效果类似,在此不再赘述。
图13是本申请一实施例提供的一种车载系统的框图。该车载系统800可以是一集成了多种功能的设备,例如,该车载系统可以是车载电脑、车机等,该车载系统可以包括上述的语音唤起设备。
参照图13,车载系统800可以包括以下一个或多个组件:处理组件802,存储器804,电源组件806,多媒体组件808,音频组件810,输入/输出(I/O)的接口812,传感器组件814,以及通信组件816。
处理组件802通常控制车载系统800的整体操作,诸如与显示,电话呼叫,数据通信,相机操作和记录操作相关联的操作。处理组件802可以包括一个或多个处理器820来执行指令,以完成上述语音唤起方法中S101至S706的全部或部分步骤。此外,处理组件802可以包括一个或多个模块,便于处理组件802和其他组件之间的交互。例如,处理组件802可以包括多媒体模块,以方便多媒体组件808和处理组件802之间的交互。
存储器804被配置为存储各种类型的数据以支持在车载系统800的操作。这些数据的示例包括用于在车载系统800上操作的任何应用程序或方法的指令,联系人数据,电话簿数据,消息,图片,视频等。存储器804可以由任何类型的易失性或非易失性存储设备或者它们的组合实现,如静态随机存取存储器(SRAM),电可擦除可编程只读存 储器(EEPROM),可擦除可编程只读存储器(EPROM),可编程只读存储器(PROM),只读存储器(ROM),磁存储器,快闪存储器,磁盘或光盘。
电源组件806为车载系统800的各种组件提供电力。电源组件806可以包括电源管理系统,一个或多个电源,及其他与为车载系统800生成、管理和分配电力相关联的组件。
多媒体组件808包括在所述车载系统800和用户之间的提供一个输出接口的屏幕。在一些实施例中,屏幕可以包括液晶显示器(LCD)和触摸面板(TP)。如果屏幕包括触摸面板,屏幕可以被实现为触摸屏,以接收来自用户的输入信号。触摸面板包括一个或多个触摸传感器以感测触摸、滑动和触摸面板上的手势。所述触摸传感器可以不仅感测触摸或滑动动作的边界,而且还检测与所述触摸或滑动操作相关的持续时间和压力。在一些实施例中,多媒体组件808还可以包括前置摄像头。
音频组件810被配置为输出和/或输入音频信号。例如,音频组件810包括一个麦克风(MIC),当车载系统800处于操作模式,如呼叫模式、记录模式和语音识别模式时,麦克风被配置为接收外部音频信号。所接收的音频信号可以被进一步存储在存储器804或经由通信组件816发送。在一些实施例中,音频组件810还包括一个扬声器,用于输出音频信号。
I/O接口812为处理组件802和外围接口模块之间提供接口,上述外围接口模块可以是点击轮、按钮等。这些按钮可包括但不限于:音量按钮、启动按钮和锁定按钮。
传感器组件814包括一个或多个传感器,用于为车载系统800提供各个方面的状态评估。在一些实施例中,该传感器组件814还可以包括加速度传感器,陀螺仪传感器,磁传感器,压力传感器或温度传感器。
通信组件816被配置为便于车载系统800和其他设备之间有线或无线方式的通信。车载系统800可以接入基于通信标准的无线网络,如WiFi,2G或3G,或它们的组合。在一个示例性实施例中,通信组件816经由广播信道接收来自外部广播管理系统的广播信号或广播相关信息。在一个示例性实施例中,所述通信组件816还包括近场通信(NFC)模块,以促进短程通信。例如,在NFC模块可基于射频识别(RFID)技术,红外数据协会(IrDA)技术,超宽带(UWB)技术,蓝牙(BT)技术和其他技术来实现。
在示例性实施例中,车载系统800可以被一个或多个应用专用集成电路(ASIC)、数字信号处理器(DSP)、数字信号处理设备(DSPD)、可编程逻辑器件(PLD)、现场可编程门阵列(FPGA)、控制器、微控制器、微处理器或其他电子元件实现,用于执 行上述语音唤起方法。
在上述图12中关于通用的语音唤起设备的描述的基础上,本申请还提供了另一实施例,本实施例具体公开了一种用于交通工具的语音唤起设备。可选的,该语音唤起设备可以被整合在交通工具的中央控制系统中,例如可以被整合在上述实施例所涉及的车载系统中。可选的,该车载系统可以是车辆上的车机所集成的系统,例如车载导航系统和/或车载娱乐系统,还可以是包含车机和车辆其他设备例如传感器等的系统。可选的,该用于交通工具的语音唤起设备包括但不限于:车机设备、交通工具出厂后附加的控制设备等等。
具体的,该用于交通工具的语音唤起设备可以包括;机载输入设备、机载处理器、机载输出设备以及其他附加设备。需要说明的是,本申请实施例所涉及的“机载输入设备”、“机载输出设备”、“机载处理器”中的机载,可以是承载于车辆上的“车载输入设备”、“车载输出设备”以及“车载处理器”,还可以是承载于飞行器上的“机载输入设备”、“机载输出设备”、“机载处理器”,还可以是承载于其他类型交通工具上的设备,本申请实施例对“机载”的含义并不做限定。以交通工具是车辆为例,该机载输入设备可以是车载输入设备、机载处理器可以是车载处理器、机载输出设备可以是车载输出设备。
取决于所安装的交通工具的类型的不同,上述机载处理器可以使用各种应用专用集成电路(ASIC)、数字信号处理器(DSP)、数字信号处理设备(DSPD)、可编程逻辑器件(PLD)、现场可编程门阵列(FPGA)、中央处理器(CPU)、控制器、微控制器、微处理器或其他电子元件实现,并用于执行上述方法。上述机载处理器通过车内线路或无线连接耦合到上述机载输入设备和机载输出设备。按照上述图2至8对应的实施例中的方法,机载处理器用于用于获取语音交互场景,并根据所述语音交互场景确定对应的第一语音。
取决于所安装的交通工具的类型的不同,上述机载输出设备可以是能够与用户进行交互的接口(例如语音播报设备、扬声器、耳机等),或者,还可以是与用户的手持设备等建立无线传输的收发信机,该机载输出设备可以通过车内线路或者无线方式耦合至上述机载输入设备和机载处理器。按照上述图2至8对应的实施例中的方法,机载输出设备,用于向用户播报所述第一语音,所述第一语音为提示用户与所述交通工具或者所述交通工具上的设备进行交互的语音。
取决于所安装的交通工具的类型的不同,上述机载输入设备可以包括多种输入设备, 例如可以包括面向用户的车载用户接口、面向设备的车载设备接口、收发信机中的至少一个。可选的,该面向设备的设备接口可以是用于设备与设备之间进行数据传输的有线接口(例如车辆的中控台上的与行车记录仪的连接接口、车辆的中控台上的与车门之间的线路接口、车辆的中控台上的与车载空调之间的硬件接口)、还可以是用于设备与设备之间进行数据传输的硬件插入接口(例如USB接口、串口等)、还可以是车辆的安全带插口、车辆发动机等硬件设施与其他控制设备之间的接口等;可选的,该面向用户的车载用户接口例如可以是用于车辆的方向盘控制按键、用于大型车辆或小型车辆的中控控制按键、用于接收语音输入的语音输入设备(例如,安置在方向盘或操作舵上的麦克风、中央声音采集设备、等等)、以及用户接收用户触摸输入的触摸感知设备(例如具有触摸感应功能的触摸屏、触控板等);可选的,上述收发信机可以是车辆中具有通信功能的射频收发芯片、基带处理芯片以及收发天线等。按照上述图2至8对应的实施例中的方法,该机载输入设备用于接收用户根据所述第一语音输入的操作信息,则上述机载处理器,还用于根据所述机载输入设备获得的所述操作信息执行相应的操作。
进一步地,上述机载处理器还可以用于上述消息的推送方法中图3至图8对应的实施例中的全部或部分步骤,本申请实施例在此不再赘述。
一种计算机/处理器可读存储介质,所述存储介质中存储有程序指令,所述程序指令用于使所述计算机/处理器执行:
获取语音交互场景;
根据所述语音交互场景确定对应的第一语音,并向用户播报所述第一语音,所述第一语音为提示用户与所述交通工具或者所述交通工具上的设备进行交互的语音;
接收用户根据所述第一语音输入的操作信息,并根据所述操作信息执行相应的操作。
可选的,所述接收用户根据所述第一语音输入的操作信息,并根据所述操作信息执行相应的操作,具体包括:
接收用户根据所述第一语音所输入的用户操作指令;
根据所述用户操作指令,判断所述用户的操作是否为所述第一语音所指示用户执行的用户操作;
若是,则停止播报所述第一语音。
可选的,所述接收用户根据所述第一语音输入的操作信息,并根据所述操作信息执行相应的操作,具体包括:
接收用户根据所述第一语音所提示的内容输入的第二语音,所述第一语音用于提示 用户针对所述第一语音所提示的内容进行选择操作,所述第二语音为用户的选择结果;
根据所述第二语音执行相应的操作。
可选的,所述获取交通工具当前行驶状态所对应的语音交互场景,具体包括:
获取与交通工具行驶相关的参数信息;其中,所述参数信息包括外界行驶环境信息、交通工具状态信息和用户行为信息中的至少一类;
根据所述参数信息确定交通工具当前行驶状态对应的语音交互场景。
可选的,所述外界行驶环境信息包括路况信息和/或天气信息,所述交通工具状态信息包括交通工具自身状况信息和/或交通工具警告音信息。可选的,所述交通工具警告音信息包括所述交通工具警告音的类型或者所述交通工具警告音的播报次数。
可选的,所述参数信息包括用户行为信息,所述获取与交通工具行驶相关的参数信息,具体包括:
向存储所述用户行为信息的设备发送获取指令,所述获取指令携带所述用户预设的授权码或者所述交通工具的标识;
接收所述设备在验证所述授权码或者所述交通工具的标识合法后发送的用户行为信息。
可选的,所述参数信息包括用户行为信息,所述获取与交通工具行驶相关的参数信息,具体包括:
接收用户输入的用户行为信息。
可选的,所述用户行为信息包括用户待办事件。
可选的,所述用户待办事件包括所述用户待办事件的发生时刻,则所述向用户播报所述第一语音,具体包括:
根据所述用户待办事件的发生时刻,确定向所述用户播报所述第一语音的时刻;
在所述时刻到达时,向所述用户播报所述第一语音。
上述可读存储介质可以是由任何类型的易失性或非易失性存储设备或者它们的组合实现,如静态随机存取存储器(SRAM),电可擦除可编程只读存储器(EEPROM),可擦除可编程只读存储器(EPROM),可编程只读存储器(PROM),只读存储器(ROM),磁存储器,快闪存储器,磁盘或光盘。
在上述实施例的基础上,本申请还提供一种车载互联网操作系统。本领域技术人员可以理解,该车载互联网操作系统可以管理和控制上述图12或图13所示的语音唤起设备的硬件或者车载系统的硬件或者本申请所涉及的用于交通工具的语音唤起设备的硬件 以及本申请所涉及的软件资源的计算机程序,是直接运行在上述语音唤起设备或用于交通工具的语音唤起设备或者上述图13所涉及的车载系统上的软件。该操作系统可以是用户与上述语音唤起设备或者用于交通工具的语音唤起设备的接口,也可以是硬件与其它软件的接口。
本申请提供的车载互联网操作系统,可以与车辆上的其他模块或功能设备进行交互,以控制相应模块或功能设备的功能。
具体地,以上述实施例中的交通工具为车辆、该语音唤起设备为车辆上的车机为例,基于本申请提供的车载互联网操作系统以及车辆通信技术的发展,使得车辆不再独立于通信网络以外,车辆可以与服务端或者网络服务器互相连接起来组成网络,从而形成车载互联网。该车载互联网系统可以提供语音通信服务、定位服务、导航服务、移动互联网接入、车辆紧急救援、车辆数据和管理服务、车载娱乐服务等。
下面详细说明本申请提供的车载互联网操作系统的结构示意图。图14为本申请一实施例提供的车载互联网操作系统的结构示意图。如图14所示,本申请提供的操作系统包括:
语音控制单元31,根据所获取的语音交互场景确定对应的第一语音并向用户进行播报;
操作控制单元32,根据车载输入设备获取的操作信息控制语音唤起系统执行相应的操作;其中,所述操作信息为用户根据所述第一语音输入给所述车载输入设备的。
具体地,本实施例中的语音唤起系统可以包括上述实施例中的语音唤起设备的部分硬件,例如可以包括上述实施例中的处理器和输出设备。该语音唤起系统还可以集成在上述车载互联网操作系统,还可以作为辅助车载互联网操作系统执行相应功能操作的系统。
语音控制单元31可以控制语音唤起系统根据所采集的外界行驶环境信息、交通工具状态信息和用户行为信息中的至少一类参数信息来确定当前行驶状态对应的语音交互场景。可选的,该语音交互场景可以是语音控制单元31获取的,还可以是语音控制单元31控制语音唤起系统获取的。
另外,本实施例中的车载输入设备可以包括上述实施例中的输入设备,即在语音控制单元31控制语音唤起系统向用户播报第一语音之后,用户根据第一语音向车载输入设备输入操作信息,从而操作控制单元32可以根据该操作信息控制语音唤起系统执行相应的操作。
进一步地,该车载互联网操作系统可以通过上述的语音控制单元31以及操作控制单元32,或者在上述两种单元的基础上,结合其它单元,控制相应的组件以执行上述图2至图8所述的方法。
最后应说明的是:以上各实施例仅用以说明本申请的技术方案,而非对其限制;尽管参照前述各实施例对本申请进行了详细的说明,本领域的普通技术人员应当理解:其依然可以对前述各实施例所记载的技术方案进行修改,或者对其中部分或者全部技术特征进行等同替换;而这些修改或者替换,并不使相应技术方案的本质脱离本申请各实施例技术方案的范围。

Claims (35)

  1. 一种语音唤起方法,其特征在于,包括:
    获取语音交互场景;
    根据所述语音交互场景确定对应的第一语音,并向用户播报所述第一语音,所述第一语音为提示用户与交通工具或者所述交通工具上的设备进行交互的语音;
    接收用户根据所述第一语音输入的操作信息,并根据所述操作信息执行相应的操作。
  2. 根据权利要求1所述的方法,其特征在于,所述接收用户根据所述第一语音输入的操作信息,并根据所述操作信息执行相应的操作,具体包括:
    接收用户根据所述第一语音所输入的用户操作指令;
    根据所述用户操作指令,判断所述用户的操作是否为所述第一语音所指示用户执行的用户操作;
    若是,则停止播报所述第一语音。
  3. 根据权利要求1所述的方法,其特征在于,所述接收用户根据所述第一语音输入的操作信息,并根据所述操作信息执行相应的操作,具体包括:
    接收用户根据所述第一语音所提示的内容输入的第二语音,所述第一语音用于提示用户针对所述第一语音所提示的内容进行选择操作,所述第二语音为用户的选择结果;
    根据所述第二语音执行相应的操作。
  4. 根据权利要求1-3任一项所述的方法,其特征在于,所述获取语音交互场景,具体包括:
    获取与所述交通工具行驶相关的参数信息;其中,所述参数信息包括外界行驶环境信息、交通工具状态信息和用户行为信息中的至少一类;
    根据所述参数信息确定交通工具当前行驶状态对应的语音交互场景。
  5. 根据权利要求4所述的方法,其特征在于,所述外界行驶环境信息包括路况信息和/或天气信息,所述交通工具状态信息包括交通工具自身状况信息和/或交通工具警告音信息。
  6. 根据权利要求5所述的方法,其特征在于,所述交通工具警告音信息包括所述交通工具警告音的类型或者所述交通工具警告音的播报次数。
  7. 根据权利要求4所述的方法,其特征在于,所述参数信息包括用户行为信息,所述获取与交通工具行驶相关的参数信息,具体包括:
    向存储所述用户行为信息的设备发送获取指令,所述获取指令携带所述用户预设的授权码或者所述交通工具的标识;
    接收所述设备在验证所述授权码或者所述交通工具的标识合法后发送的用户行为信息。
  8. 根据权利要求4所述的方法,其特征在于,所述参数信息包括用户行为信息,所述获取与交通工具行驶相关的参数信息,具体包括:
    接收用户输入的用户行为信息。
  9. 根据权利要求7或8所述的方法,其特征在于,所述用户行为信息包括用户待办事件。
  10. 根据权利要求9所述的方法,其特征在于,所述用户待办事件包括所述用户待办事件的发生时刻,则所述向用户播报所述第一语音,具体包括:
    根据所述用户待办事件的发生时刻,确定向所述用户播报所述第一语音的时刻;
    在所述时刻到达时,向所述用户播报所述第一语音。
  11. 一种语音唤起装置,其特征在于,包括:
    获取模块,用于获取语音交互场景;
    确定模块,用于根据所述获取模块获取的所述语音交互场景确定对应的第一语音,所述第一语音为提示用户与交通工具或者所述交通工具上的设备进行交互的语音;
    语音播报模块,用于向用户播报所述确定模块确定的所述第一语音;
    接收模块,用于接收用户根据所述确定模块确定的所述第一语音输入的操作信息;
    处理模块,用于根据所述接收模块接收到的所述操作信息执行相应的操作。
  12. 根据权利要求11所述的装置,其特征在于,所述接收模块,具体用于接收用户根据所述第一语音所输入的用户操作指令;
    所述处理模块,具体用于根据所述用户操作指令,判断所述用户的操作是否为所述第一语音所指示用户执行的用户操作,并在判断所述用户的操作为所述第一语音所指示用户执行的用户操作时,指示所述语音播报模块停止播报所述第一语音。
  13. 根据权利要求11所述的装置,其特征在于,所述接收模块,具体用于接收用户根据所述第一语音所提示的内容输入的第二语音,所述第一语音用于提示用户针对所述第一语音所提示的内容进行选择操作,所述第二语音为用户的选择结果;
    所述处理模块,具体用于根据所述第二语音执行相应的操作。
  14. 根据权利要求11-13任一项所述的装置,其特征在于,所述获取模块,包括:
    获取子模块,用于获取与交通工具行驶相关的参数信息;其中,所述参数信息包括外界行驶环境信息、交通工具状态信息和用户行为信息中的至少一类;
    确定子模块,用于根据所述参数信息确定交通工具当前行驶状态对应的语音交互场景。
  15. 根据权利要求14所述的装置,其特征在于,所述外界行驶环境信息包括路况信息和/或天气信息,所述交通工具状态信息包括交通工具自身状况信息和/或交通工具警告音信息。
  16. 根据权利要求15所述的装置,其特征在于,所述交通工具警告音信息包括所述交通工具警告音的类型或者所述交通工具警告音的播报次数。
  17. 根据权利要求14所述的装置,其特征在于,所述参数信息包括用户行为信息,所述获取子模块,具体包括:
    发送单元,用于向存储所述用户行为信息的设备发送获取指令,所述获取指令携带所述用户预设的授权码或者所述交通工具的标识;
    接收单元,用于接收所述设备在验证所述授权码或者所述交通工具的标识合法后发送的用户行为信息。
  18. 根据权利要求14所述的装置,其特征在于,所述参数信息包括用户行为信息,所述获取子模块,具体用于接收用户输入的用户行为信息。
  19. 根据权利要求17或18所述的装置,其特征在于,所述用户行为信息包括用户待办事件。
  20. 根据权利要求19所述的装置,其特征在于,所述用户待办事件包括所述用户待办事件的发生时刻;所述语音播报模块,包括:
    确定单元,用于根据所述用户待办事件的发生时刻,确定向所述用户播报所述第一语音的时刻;
    播报单元,用于在所述确定单元确定的所述时刻到达时,向所述用户播报所述第一语音。
  21. 一种语音唤起设备,其特征在于,包括:
    处理器,用于获取语音交互场景,并根据所述语音交互场景确定对应的第一语音;
    输出设备,耦合至所述处理器,用于向用户播报所述第一语音,所述第一语音为提示用户与交通工具或者所述交通工具上的设备进行交互的语音;
    输入设备,耦合至所述处理器,用于接收用户根据所述第一语音输入的操作信息;
    所述处理器,还用于根据所述输入设备获得的所述操作信息执行相应的操作。
  22. 根据权利要求21所述的设备,其特征在于,所述输入设备,具体用于接收用户根据所述第一语音所输入的用户操作指令;
    所述处理器,具体用于根据所述用户操作指令,判断所述用户的操作是否为所述第一语音所指示用户执行的用户操作,并在判断所述用户的操作为所述第一语音所指示用户执行的用户操作时,指示所述输出设备停止播报所述第一语音。
  23. 根据权利要求21所述的设备,其特征在于,所述输入设备,具体用于接收用户根据所述第一语音所提示的内容输入的第二语音,所述第一语音用于提示用户针对所述第一语音所提示的内容进行选择操作,所述第二语音为用户的选择结果;
    所述处理器,具体用于根据所述第二语音执行相应的操作。
  24. 根据权利要求21-23任一项所述的设备,其特征在于,所述输入设备,还用于获取与交通工具行驶相关的参数信息;其中,所述参数信息包括外界行驶环境信息、交通工具状态信息和用户行为信息中的至少一类;
    所述处理器,具体用于根据所述参数信息确定交通工具当前行驶状态对应的语音交互场景。
  25. 根据权利要求24所述的设备,其特征在于,所述外界行车环境信息包括路况信息和/或天气信息,所述交通工具状态信息包括交通工具自身状况信息和/或交通工具警告音信息。
  26. 根据权利要求25所述的设备,其特征在于,所述交通工具警告音信息包括所述交通工具警告音的类型或者所述交通工具警告音的播报次数。
  27. 根据权利要求24所述的设备,其特征在于,所述参数信息包括用户行为信息;
    所述输出设备,还用于向存储所述用户行为信息的设备发送获取指令,所述获取指令携带所述用户预设的授权码或者所述交通工具的标识;
    所述输入设备,具体用于接收所述设备在验证所述授权码或者所述交通工具的标识合法后发送的用户行为信息。
  28. 根据权利要求24所述的设备,其特征在于,所述参数信息包括用户行为信息,所述输入设备,具体用于接收用户输入的用户行为信息。
  29. 根据权利要求27或28所述的设备,其特征在于,所述用户行为信息包括用户待办事件。
  30. 根据权利要求29所述的设备,其特征在于,所述用户待办事件包括所述用户 待办事件的发生时刻;
    所述处理器,还用于根据所述用户待办事件的发生时刻,确定向所述用户播报所述第一语音的时刻;
    所述输出设备,具体用于在所述时刻到达时,向所述用户播报所述第一语音。
  31. 一种用于交通工具的语音唤起设备,其特征在于,包括:机载处理器、机载输出设备和机载输入设备;
    所述机载处理器,用于获取语音交互场景,并根据所述语音交互场景确定对应的第一语音;
    所述机载输出设备,耦合至所述机载处理器,用于向用户播报所述第一语音,所述第一语音为提示用户与所述交通工具或者所述交通工具上的设备进行交互的语音;
    所述机载输入设备,耦合至所述机载处理器,用于接收用户根据所述第一语音输入的操作信息;
    所述机载处理器,还用于根据所述机载输入设备获得的所述操作信息执行相应的操作。
  32. 根据权利要求31所述的用于交通工具的语音唤起设备,其特征在于,所述机载输入设备包括面向用户的车载用户接口、面向设备的车载设备接口、收发信机中的至少一个。
  33. 根据权利要求32所述的用于交通工具的语音唤起设备,其特征在于,所述面向用户的车载用户接口包括以下一个或多个:
    中控台控制按键;
    方向盘控制按键;
    语音接收设备;
    触控感知设备。
  34. 根据权利要求31-33任一项所述的用于交通工具的语音唤起设备,其特征在于,所述机载处理器,还用于执行上述权利要求2-10任一项所述的方法。
  35. 一种车载互联网操作系统,其特征在于,包括:
    语音控制单元,根据所获取的语音交互场景确定对应的第一语音并向用户进行播报;
    操作控制单元,根据车载输入设备获取的操作信息控制语音唤起系统执行相应的操作;其中,所述操作信息为用户根据所述第一语音输入给所述车载输入设备的。
PCT/CN2017/080387 2016-04-20 2017-04-13 语音唤起方法、装置和设备 WO2017181901A1 (zh)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201610246576.5A CN107303909B (zh) 2016-04-20 2016-04-20 语音唤起方法、装置和设备
CN201610246576.5 2016-04-20

Publications (1)

Publication Number Publication Date
WO2017181901A1 true WO2017181901A1 (zh) 2017-10-26

Family

ID=60115613

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2017/080387 WO2017181901A1 (zh) 2016-04-20 2017-04-13 语音唤起方法、装置和设备

Country Status (3)

Country Link
CN (1) CN107303909B (zh)
TW (1) TW201742424A (zh)
WO (1) WO2017181901A1 (zh)

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109532725A (zh) * 2019-01-09 2019-03-29 北京梧桐车联科技有限责任公司 一种车载系统
CN111007938A (zh) * 2018-10-08 2020-04-14 阿里巴巴集团控股有限公司 交互设备及其处理方法、装置
CN111158477A (zh) * 2019-12-25 2020-05-15 北京金山安全软件有限公司 一种可穿戴设备的工作方法、控制方法、可穿戴设备及服务器
CN111724772A (zh) * 2019-03-20 2020-09-29 阿里巴巴集团控股有限公司 一种智能设备的交互方法、装置和智能设备
CN112109729A (zh) * 2019-06-19 2020-12-22 宝马股份公司 用于车载系统的人机交互方法、装置和系统
CN112201240A (zh) * 2020-09-27 2021-01-08 上汽通用五菱汽车股份有限公司 车辆控制方法、车载无屏装置、服务器及可读存储介质
CN112634551A (zh) * 2020-11-30 2021-04-09 中油国家油气钻井装备工程技术研究中心有限公司 一种司钻房内的语音报警控制方法
CN113990322A (zh) * 2021-11-04 2022-01-28 广州小鹏汽车科技有限公司 语音交互方法、服务器、语音交互系统和介质
CN114465837A (zh) * 2022-01-30 2022-05-10 云知声智能科技股份有限公司 一种智能语音设备协同唤醒处理方法及装置

Families Citing this family (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107878467B (zh) * 2017-11-10 2020-07-17 爱驰汽车(上海)有限公司 用于汽车的语音播报方法及系统
CN110096249A (zh) * 2018-01-31 2019-08-06 阿里巴巴集团控股有限公司 用于提示快捷唤醒词的方法、装置和系统
CN108520744B (zh) * 2018-03-15 2020-11-10 斑马网络技术有限公司 语音控制方法与装置,以及电子设备与存储介质
CN108923808A (zh) * 2018-06-05 2018-11-30 上海博泰悦臻网络技术服务有限公司 车辆及其车载终端和语音交互模式主动触发方法
CN110874202B (zh) * 2018-08-29 2024-04-19 斑马智行网络(香港)有限公司 交互方法、设备、介质和操作系统
CN109741740B (zh) * 2018-12-26 2021-04-16 苏州思必驰信息科技有限公司 基于外部触发的语音交互方法及装置
CN113763956A (zh) * 2019-03-12 2021-12-07 百度在线网络技术(北京)有限公司 应用于车辆的交互方法和装置
CN110209278A (zh) * 2019-05-30 2019-09-06 广州小鹏汽车科技有限公司 人车交互方法、装置、存储介质及控制终端
CN110203209A (zh) * 2019-06-05 2019-09-06 广州小鹏汽车科技有限公司 一种语音提示方法和装置
CN112447180A (zh) * 2019-08-30 2021-03-05 华为技术有限公司 语音唤醒方法和装置
CN111086511B (zh) * 2019-12-31 2021-05-07 浙江合众新能源汽车有限公司 一种主动语音开启汽车tja功能的方法及装置
CN111204339B (zh) * 2019-12-31 2022-02-08 浙江合众新能源汽车有限公司 一种主动语音开启lka功能的方法及装置
CN112092820B (zh) * 2020-09-03 2022-03-18 广州小鹏汽车科技有限公司 车辆的初始化设置方法、车辆和存储介质
CN113113015A (zh) * 2020-11-17 2021-07-13 广州小鹏汽车科技有限公司 交互方法、信息处理方法、车辆和服务器
CN113548062B (zh) * 2021-08-03 2022-12-30 奇瑞汽车股份有限公司 汽车的交互控制方法、装置及计算机存储介质
CN114559886B (zh) * 2022-02-26 2024-03-29 东莞市台铃车业有限公司 防误触控制方法、系统、装置、电子设备及存储介质

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101992779A (zh) * 2009-08-12 2011-03-30 福特全球技术公司 车辆中的智能音乐选择方法
CN102941852A (zh) * 2012-10-30 2013-02-27 青岛海信网络科技股份有限公司 智能车载终端
KR20130063091A (ko) * 2011-12-06 2013-06-14 현대자동차주식회사 차량용 음성인식 장치
CN103802761A (zh) * 2012-11-06 2014-05-21 罗伯特·博世有限公司 激活与车辆乘员语音交互的方法和车辆的语音交互系统
CN204736855U (zh) * 2015-06-07 2015-11-04 沈陆垚 模块化智能车语音控制及安全保障系统

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101257680B (zh) * 2008-03-26 2012-04-25 宇龙计算机通信科技(深圳)有限公司 一种具有导航功能的移动终端及导航方法
CN101951553B (zh) * 2010-08-17 2012-10-10 深圳市车音网科技有限公司 基于语音命令的导航方法及系统
US20120268294A1 (en) * 2011-04-20 2012-10-25 S1Nn Gmbh & Co. Kg Human machine interface unit for a communication device in a vehicle and i/o method using said human machine interface unit
CN104535074A (zh) * 2014-12-05 2015-04-22 惠州Tcl移动通信有限公司 基于蓝牙耳机的语音导航方法、系统和终端

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101992779A (zh) * 2009-08-12 2011-03-30 福特全球技术公司 车辆中的智能音乐选择方法
KR20130063091A (ko) * 2011-12-06 2013-06-14 현대자동차주식회사 차량용 음성인식 장치
CN102941852A (zh) * 2012-10-30 2013-02-27 青岛海信网络科技股份有限公司 智能车载终端
CN103802761A (zh) * 2012-11-06 2014-05-21 罗伯特·博世有限公司 激活与车辆乘员语音交互的方法和车辆的语音交互系统
CN204736855U (zh) * 2015-06-07 2015-11-04 沈陆垚 模块化智能车语音控制及安全保障系统

Cited By (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111007938A (zh) * 2018-10-08 2020-04-14 阿里巴巴集团控股有限公司 交互设备及其处理方法、装置
CN111007938B (zh) * 2018-10-08 2023-11-28 盒马(中国)有限公司 交互设备及其处理方法、装置
CN109532725A (zh) * 2019-01-09 2019-03-29 北京梧桐车联科技有限责任公司 一种车载系统
CN111724772A (zh) * 2019-03-20 2020-09-29 阿里巴巴集团控股有限公司 一种智能设备的交互方法、装置和智能设备
CN112109729A (zh) * 2019-06-19 2020-12-22 宝马股份公司 用于车载系统的人机交互方法、装置和系统
CN112109729B (zh) * 2019-06-19 2023-06-06 宝马股份公司 用于车载系统的人机交互方法、装置和系统
CN111158477A (zh) * 2019-12-25 2020-05-15 北京金山安全软件有限公司 一种可穿戴设备的工作方法、控制方法、可穿戴设备及服务器
CN111158477B (zh) * 2019-12-25 2024-05-17 北京金山安全软件有限公司 一种可穿戴设备的工作方法、控制方法、可穿戴设备及服务器
CN112201240B (zh) * 2020-09-27 2023-03-14 上汽通用五菱汽车股份有限公司 车辆控制方法、车载无屏装置、服务器及可读存储介质
CN112201240A (zh) * 2020-09-27 2021-01-08 上汽通用五菱汽车股份有限公司 车辆控制方法、车载无屏装置、服务器及可读存储介质
CN112634551A (zh) * 2020-11-30 2021-04-09 中油国家油气钻井装备工程技术研究中心有限公司 一种司钻房内的语音报警控制方法
CN113990322A (zh) * 2021-11-04 2022-01-28 广州小鹏汽车科技有限公司 语音交互方法、服务器、语音交互系统和介质
CN113990322B (zh) * 2021-11-04 2023-10-31 广州小鹏汽车科技有限公司 语音交互方法、服务器、语音交互系统和介质
CN114465837A (zh) * 2022-01-30 2022-05-10 云知声智能科技股份有限公司 一种智能语音设备协同唤醒处理方法及装置
CN114465837B (zh) * 2022-01-30 2024-03-08 云知声智能科技股份有限公司 一种智能语音设备协同唤醒处理方法及装置

Also Published As

Publication number Publication date
CN107303909A (zh) 2017-10-31
TW201742424A (zh) 2017-12-01
CN107303909B (zh) 2020-06-23

Similar Documents

Publication Publication Date Title
WO2017181901A1 (zh) 语音唤起方法、装置和设备
US10127804B2 (en) Trainable transceiver and cloud computing system architecture systems and methods
US9696542B2 (en) Input device disposed in handle and vehicle including the same
US10163273B2 (en) Method and system for operating mobile applications in a vehicle
WO2017181900A1 (zh) 消息的推送方法、装置和设备
US20140357248A1 (en) Apparatus and System for Interacting with a Vehicle and a Device in a Vehicle
CN106184000B (zh) 一种基于汽车智能后视镜的语音控制方法及系统
WO2018068646A1 (zh) 一种导航方法、装置、设备和操作系统
CN106161081A (zh) 与可穿戴装置进行通信的车辆系统
JP2018523342A (ja) 無線接続管理
TW200825443A (en) Portable navigation device
US10911589B2 (en) Vehicle control device
CN103986490B (zh) 仲裁通过移动应用进行流传输的音频源的系统和方法
CN105644501A (zh) 远程操控汽车的方法、装置及终端
US10429852B2 (en) Methods and devices for controlling self-balancing vehicle to park
US11830302B2 (en) Computer system for utilizing ultrasonic signals to implement operations for autonomous vehicles
TW201743281A (zh) 交通工具控制方法、控制裝置與控制系統
CN105610896A (zh) 用于启动应用的方法和系统
CN112153522A (zh) 车载音箱控制系统及车载音箱
WO2023072118A1 (zh) 一种音频流的处理方法、地址的处理方法以及相关设备
WO2023227014A1 (zh) 一种隐私保护方法及相关装置
CN112153521B (zh) 车载音箱控制系统及车载音箱
WO2023072207A1 (zh) 一种位置的获取方法以及相关设备
WO2023072120A1 (zh) 一种位置的获取方法以及相关设备
CN107578559A (zh) 共享汽车的控制方法、装置及计算机可读存储介质

Legal Events

Date Code Title Description
NENP Non-entry into the national phase

Ref country code: DE

121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 17785386

Country of ref document: EP

Kind code of ref document: A1

32PN Ep: public notification in the ep bulletin as address of the adressee cannot be established

Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 14/02/2019)

122 Ep: pct application non-entry in european phase

Ref document number: 17785386

Country of ref document: EP

Kind code of ref document: A1