CN113779300B - Voice input guiding method, device and car machine - Google Patents

Voice input guiding method, device and car machine Download PDF

Info

Publication number
CN113779300B
CN113779300B CN202010519922.9A CN202010519922A CN113779300B CN 113779300 B CN113779300 B CN 113779300B CN 202010519922 A CN202010519922 A CN 202010519922A CN 113779300 B CN113779300 B CN 113779300B
Authority
CN
China
Prior art keywords
voice
user
vehicle
information
voice guidance
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202010519922.9A
Other languages
Chinese (zh)
Other versions
CN113779300A (en
Inventor
赵伟
肖金富
刘柯
杨冬生
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
BYD Co Ltd
Original Assignee
BYD Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by BYD Co Ltd filed Critical BYD Co Ltd
Priority to CN202010519922.9A priority Critical patent/CN113779300B/en
Publication of CN113779300A publication Critical patent/CN113779300A/en
Application granted granted Critical
Publication of CN113779300B publication Critical patent/CN113779300B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/60Information retrieval; Database structures therefor; File system structures therefor of audio data
    • G06F16/63Querying
    • G06F16/635Filtering based on additional data, e.g. user or group profiles
    • G06F16/637Administration of user profiles, e.g. generation, initialization, adaptation or distribution
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W50/00Details of control systems for road vehicle drive control not related to the control of a particular sub-unit, e.g. process diagnostic or vehicle driver interfaces
    • B60W50/08Interaction between the driver and the control system
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/60Information retrieval; Database structures therefor; File system structures therefor of audio data
    • G06F16/68Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/687Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using geographical or spatial information, e.g. location

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Databases & Information Systems (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Multimedia (AREA)
  • Automation & Control Theory (AREA)
  • Library & Information Science (AREA)
  • Human Computer Interaction (AREA)
  • Transportation (AREA)
  • Mechanical Engineering (AREA)
  • Navigation (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

The embodiment of the invention provides a voice input guiding method, a voice input guiding device and a vehicle machine, wherein the voice input guiding method comprises the following steps: acquiring a voice guidance instruction set matched with a driving scene of a vehicle and a user portrait of a user; and executing a guiding operation on at least one voice guiding instruction in the voice guiding instruction set under the condition that the vehicle is not detected to have an application program running in the foreground. According to the technical scheme, the guiding requirement of voice input on the vehicle-mounted device when the user is in the driving state can be met, and the user operation is not required to be triggered, so that the input operation steps of the user can be simplified, and the use experience of the user can be improved.

Description

Voice input guiding method, device and car machine
Technical Field
The invention relates to the technical field of automobile machines, in particular to a voice input guiding method, a voice input guiding device and an automobile machine.
Background
With the development of voice recognition technology, a voice input guiding device is integrated on a mobile phone or a computer and other terminals, after operation information of a user is received, a voice input interface is displayed according to the operation information, and guiding words are displayed on the voice input interface so as to realize guiding of voice input, and the purpose of guiding the user to use voice search is achieved.
As can be seen from the above, the voice input guiding device in the prior art is not suitable for use on a vehicle, and the user is usually in a driving state when the vehicle is running, and the mode of triggering the guiding device based on the user operation information also affects the concentration of the user when driving.
Disclosure of Invention
The technical problem to be solved by the embodiment of the invention is to provide a voice input guiding method so as to meet the guiding requirement of voice input on a vehicle when a user is in a driving state.
Correspondingly, the embodiment of the invention also provides a voice input guiding device and a vehicle machine, which are used for guaranteeing the implementation and application of the method.
An embodiment of a first aspect of the present invention provides a voice input guiding method for a vehicle, including:
acquiring a voice guidance instruction set matched with a driving scene of a vehicle and a user portrait of a user;
and executing a guiding operation on at least one voice guiding instruction in the voice guiding instruction set under the condition that the vehicle is not detected to have an application program running in the foreground.
Optionally, before acquiring the voice guidance instruction set matched with the driving scene of the vehicle and the user portrait of the user, the method includes: configuring a driving scene of the vehicle and/or generating a user portrait of the user.
Optionally, the configuring the driving scene of the vehicle includes:
collecting at least one of speed information, position information, running map data and gear information of the vehicle as scene configuration information;
and configuring the driving scene according to the scene configuration information.
Optionally, the generating a user portrait of the user includes:
acquiring working condition information and historical voice records of the vehicle and the machine;
generating the user portrait according to the working condition information and the historical voice record,
The working condition information comprises touch information of the vehicle and/or an application program operated by the vehicle.
Optionally, the performing a guiding operation on at least one voice guiding instruction in the voice guiding instruction set in the case that the vehicle has no application program running in the foreground is detected, includes:
determining updating frequency and/or updating time of the voice guidance instruction set according to the driving scene;
updating the at least one voice guidance instruction for performing a guidance operation according to the update frequency and/or update time.
Optionally, the acquiring a voice guidance instruction set matched with the driving scene and the user portrait specifically includes:
Detecting whether the user portrait matches a pre-stored user portrait;
If the user portrait is matched with the pre-stored user portrait, extracting the voice guidance instruction set matched with the user portrait and the driving scene from a pre-stored voice guidance instruction library;
And if the user portrait is not matched with the pre-stored user portrait, configuring the voice guidance instruction set according to a cold start strategy.
Optionally, before acquiring the voice guidance instruction set matched with the driving scene and the user portrait, the method includes:
Collecting guide information, wherein the guide information comprises at least one of function instruction information, voice self-learning instruction information, popularization instruction information and user instruction information;
deleting the information which is not matched with the guiding rule in the guiding information so as to reserve guiding information to be processed;
dividing the information to be processed into a training set and a testing set;
executing machine learning according to the training set, the testing set, the pre-stored driving scenes and the pre-stored user portraits,
Generating a plurality of groups of pre-stored voice instruction sets and index information corresponding to each group of pre-stored voice instruction sets;
and generating the voice guidance instruction library according to a plurality of groups of pre-stored voice instruction sets and the index information.
Optionally, the configuring the voice guidance instruction set according to a cold start policy includes:
and configuring at least one group of pre-stored voice instruction sets with highest use frequency in the voice guidance instruction library as the voice guidance instruction set.
Optionally, extracting the voice guidance instruction set matching the user portrait and the driving scene from a pre-stored voice guidance instruction library includes:
and determining the index information with the highest matching degree with the user portrait and the driving scene, and determining the pre-stored voice instruction set corresponding to the index information as the voice guiding instruction set.
Optionally, the method further comprises:
And stopping executing the guiding operation under the condition that the vehicle machine is detected to have the application program running in the foreground.
An embodiment of a second aspect of the present invention provides a voice input guiding apparatus for a vehicle, including:
The acquisition unit is used for acquiring a voice guidance instruction set matched with the driving scene of the vehicle and the user portrait of the user;
And the execution unit is used for executing the guiding operation on at least one voice guiding instruction in the voice guiding instruction set under the condition that the application program running in the foreground of the vehicle is not detected.
An embodiment of a third aspect of the present invention provides a vehicle apparatus, including:
a voice input guiding device according to an embodiment of the second aspect of the present invention.
According to the embodiment of the invention, the voice guiding instruction set is determined according to the driving scene and the user image, so that the obtained voice guiding instruction set can simultaneously consider the driving state of the vehicle and the preference of the user, the voice input guiding device on the vehicle can guide the voice input of the user in the driving process of the vehicle, the requirement of the user for transmitting related voice instructions in the driving process is met, and the travel related information can be pushed to the user based on the voice feedback of the user, so that the travel purpose of the auxiliary user is achieved. Further, by detecting whether the vehicle has the application program running in the foreground or not, when the application program running in the foreground is not detected, the guiding operation of executing the voice input is triggered, on one hand, the voice input guiding device and other application programs can be ensured to run relatively independently, on the other hand, the user operation of triggering is not needed, and the input operation steps of the user can be simplified, so that the use experience of the user can be improved.
Drawings
FIG. 1 is a flow chart of steps of an embodiment of a voice input guidance method of the present invention;
FIG. 2 is a flow chart of steps of another embodiment of a voice input guidance method of the present invention;
FIG. 3 is a schematic flow chart diagram of a startup embodiment of a voice input guidance module of a vehicle of the present invention;
FIG. 4 is a schematic flow chart diagram of an embodiment of the generation of a voice guidance instruction library of the present invention;
FIG. 5 is a schematic flow chart diagram of an embodiment of the present invention for capturing a voice guidance instruction set that matches a driving scene and a user representation;
FIG. 6 is a flowchart illustrating steps of yet another embodiment of a voice input guidance method of the present invention;
FIG. 7 is a block diagram of an embodiment of a voice input guiding apparatus of the present invention;
Fig. 8 is a block diagram of a vehicle machine embodiment of the present invention.
Detailed Description
In order that the above-recited objects, features and advantages of the present invention will be more clearly understood, a more particular description of the invention will be rendered by reference to the appended drawings and appended detailed description. It should be noted that, without conflict, the embodiments of the present invention and features in the embodiments may be combined with each other.
In the following description, numerous specific details are set forth in order to provide a thorough understanding of the present invention, but the present invention may be practiced in other ways than those described herein, and therefore the scope of the present invention is not limited to the specific embodiments disclosed below.
The embodiment of the invention is applied to a vehicle machine, the vehicle machine refers to short for vehicle-mounted information entertainment products arranged in an automobile, the vehicle machine can be used for realizing information communication with the automobile and the outside (automobile-to-automobile) in function, the vehicle machine comprises a host, a display module, a microphone and a network module, a touch module can be integrated on the display module, and the vehicle machine can also be electrically connected with a loudspeaker module of the automobile.
Referring to fig. 1, a flowchart illustrating steps of an embodiment of a voice input guiding method of the present invention may specifically include the following steps:
Step S102, a voice guidance instruction set matched with a driving scene of a vehicle and a user portrait of a user is obtained.
The voice instruction guiding instruction set can simultaneously give consideration to the running state of the vehicle and the preference of the user by respectively generating the running scene of the vehicle and the user image of the user so as to meet the requirement of the user on sending related voice instructions in the running process.
Specifically, the operation of obtaining the voice guidance instruction set may be performed locally in the vehicle, or may be performed through communication interaction with an adapted server, for example: and sending the driving scene and the user portrait to an adaptive server so as to receive a voice guidance instruction set issued by the server according to the driving scene and the user portrait.
Step S104, in the case that the vehicle is not detected to have the application program running in the foreground, the guiding operation is executed on at least one voice guiding instruction in the voice guiding instruction set.
The guiding operation executed by the voice guiding instruction specifically guides the user to speak the displayed voice guiding information.
The method comprises the steps of detecting whether the vehicle has a foreground running application program or not, triggering and executing the guiding operation of voice input when the foreground running application program is not detected, and normally, if the foreground running application program is not available, displaying a main interface on a display screen of the vehicle, and executing the guiding operation under the working condition without affecting the running of other application programs.
Specifically, the guiding operation is performed on at least one voice guiding instruction in the voice guiding instruction set, and guiding sentences can be displayed on a display screen, guiding audio can be played, and the like.
In addition, the first embodiment is preferable in that the guidance operation is performed on at least one voice guidance command in the voice guidance command set, and only one voice guidance command may be displayed at a time, or a plurality of voice guidance commands may be displayed at the same time.
According to the voice input guiding method, the voice input guiding device on the vehicle can guide voice input of the user in the vehicle driving process by determining the voice guiding instruction set according to the driving scene and the user image, the requirement of the user for sending related voice instructions in the driving process is met, and further travel related information can be pushed to the user based on voice feedback of the user, so that the purpose of assisting the user in traveling is achieved.
Further, by detecting whether the vehicle has the application program running in the foreground or not, when the application program running in the foreground is not detected, the guiding operation of executing the voice input is triggered, on one hand, the voice input guiding device and other application programs can be ensured to run relatively independently, on the other hand, the user operation of triggering is not needed, and the input operation steps of the user can be simplified, so that the use experience of the user can be improved.
In some embodiments, one possible implementation of step S104 is: determining updating frequency and/or updating time of a voice guidance instruction set according to driving scenes; at least one voice guidance instruction for performing the guidance operation is updated according to the update frequency and/or the update time.
In one preferred implementation manner, the performing, on at least one voice guidance instruction in the voice guidance instruction set, the guiding operation, specifically, displaying only voice guidance information corresponding to one voice guidance instruction on a display screen of the vehicle, and updating, according to the update frequency, the at least one voice guidance instruction for performing the guiding operation, includes: by combining the current driving scene, the updating frequency and the updating time of the single piece of data in the voice guiding instruction set at the vehicle machine end are determined, so that when the driving scene is changed, the displayed voice guiding information is correspondingly adjusted or when the user does not feed back the current voice guiding information, the next voice guiding information is timely switched to be displayed.
At least one voice guidance instruction for performing a guidance operation is updated according to the update time, and specifically includes: the corresponding time for displaying the voice guidance information is also different in different driving scenes, for example, the corresponding voice guidance information is a switching driving route in driving scenes of traffic jam driving, and can be displayed on a display screen of a vehicle in real time at the moment, for example, the corresponding driving speeds are different in driving scenes such as normal driving, overspeed driving, low-speed driving, traffic light stopping and the like, and if the voice guidance instruction is a multimedia file playing, the corresponding different display time is needed.
Referring to fig. 2, a flowchart illustrating steps of another embodiment of a voice input guidance method of the present invention may specifically include the following steps:
step S202, configuring a driving scene of a vehicle and/or generating a user portrait of a user.
The driving scene is used for determining the current driving state, the user portrays the preference of the user, and the matched voice guidance instruction set can be determined based on the information by determining the information, so that the driving of the user is assisted.
Specifically, the driving scene may include normal driving, overspeed driving, low-speed driving, temporary stopping, traffic light stopping, high-speed driving, traffic jam driving, and the like.
Step S204, a voice guidance instruction set matched with the driving scene of the vehicle and the user portrait of the user is obtained.
Step S206, in the case that the vehicle is not detected to have the application program running in the foreground, the guiding operation is executed on at least one voice guiding instruction in the voice guiding instruction set.
In this embodiment, the specific implementation process of step S204 and step S206 may refer to the descriptions related to step 102 and step 104 in the embodiment shown in fig. 1, and will not be repeated here.
In some embodiments, one possible implementation of configuring the driving scenario of the vehicle described in step S202 is: collecting at least one of speed information, position information, running map data and gear information of a vehicle as scene configuration information; and configuring a driving scene according to the scene configuration information.
The information is collected, so that the current vehicle can be judged to be in any driving scene of normal driving, overspeed driving, low-speed driving, temporary stopping, traffic light stopping, high-speed driving, traffic jam driving and the like.
In some embodiments, one possible implementation of generating a user representation of the user described in step S202 is: acquiring working condition information and historical voice records of a vehicle; and generating a user image according to the working condition information and the historical voice record, wherein the working condition information comprises touch information of the vehicle and/or an application program operated by the vehicle and the machine.
The user's habit and preference of application programs in the car machine can be known through analysis of the working condition information of the car machine, the working condition information comprises touch information of the car machine, application programs with higher running frequency in the car machine and the like, the preference of the user can be known through analysis of the historical voice record, and further, the user portrait is constructed through information such as the habit, the preference of use and the preference of the user, so that the user portrait can be more accurately applied to voice guidance.
Specifically, as shown in fig. 3, the starting module of the voice input guiding module of the vehicle machine includes: a data collector 302, a content decider 304, a presentation decider 306, and a presentation module 308.
The data collector 302 is used for collecting speed information, position information, running map data, gear information, touch information of the vehicle, application programs running by the vehicle and historical voice records of the vehicle.
The content decider 304 is configured to configure a driving scene according to speed information, position information, driving map data, and gear information of the vehicle.
And configuring user images according to the touch information of the vehicle and the machine, the application programs operated by the vehicle and the history voice records.
The presentation decision-maker 306 is configured to determine a voice guidance command set, and a voice guidance command update frequency and/or update time according to the driving scene and the user image.
The display module 308 is configured to display a decision result, i.e. a voice guidance instruction.
Optionally, as shown in fig. 4, one possible implementation manner of step S102, that is, an implementation manner of acquiring a voice guidance instruction set matched with a driving scene and a user portrait, includes:
Step S402, detecting whether the user portrait matches with a pre-stored user portrait.
Step S404, if the user image is matched with the pre-stored user image, extracting a voice guidance instruction set matched with the user image and the driving scene from a pre-stored voice guidance instruction library.
Step S406, if the user portrait does not match with the pre-stored user portrait, the voice guidance instruction set is configured according to the cold start strategy.
It can be understood by those skilled in the art that the implementation process may be performed at the vehicle-mounted device end, or may be performed at a server end adapted to the vehicle-mounted device.
In some embodiments, one possible implementation of step S404 is: and determining index information with highest matching degree with the user image and the driving scene, and determining a pre-stored voice instruction set corresponding to the index information as a voice guiding instruction set.
In some embodiments, one possible implementation of step S406 is: at least one group of pre-stored voice instruction sets with highest use frequency in the voice guidance instruction library is configured as a voice guidance instruction set.
Specifically, if a user representation is detected to be not matched with a pre-stored user representation, the voice guidance instruction library is indicated to not have a voice guidance instruction set matched with the user representation, and in this case, the cold start strategy may be to provide a group of voice guidance instruction sets with highest use frequency to the vehicle.
Optionally, before step S202, a process of generating a voice guidance command library is further included, so as to query the voice guidance command library for an adapted voice guidance command set.
Specifically, as shown in fig. 5, the generation process of the voice guidance instruction library includes:
step S502, collecting guiding information, where the guiding information includes function instruction information, voice self-learning instruction information, promotion instruction information and user instruction information.
The function instruction information can be an operation instruction of a specified application program, the voice self-learning instruction information can be a voice instruction identified through a training model, the promotion instruction information can be voice information based on promotion requirements, and the user instruction information is instruction information commonly used by a user.
Step S504, detecting whether the guiding information is matched with the guiding rule, so as to carry out security examination and content examination on the guiding information.
Specifically, the main function of detecting whether the guiding information is matched with the guiding rule is to perform security examination and content examination on data provided by a plurality of guiding word input sources, wherein the security examination is used for detecting whether the guiding information comprises sensitive words such as yellow and black, and the content examination comprises two steps, and as a voice guiding instruction only supports Chinese and English, the first step is to detect whether special symbols, numbers and the like, and the second step is an intelligent voice semantic analysis flow to see whether the intelligent voice can accurately analyze the instruction so as to ensure the reliability of the voice guiding instruction.
Step S506, reserving the matched guide information to be processed.
Step S508, dividing the information to be processed into a training set and a testing set.
Step S510, machine learning is executed according to the training set, the testing set, the pre-stored driving scene and the pre-stored user portrait, so as to generate a plurality of groups of pre-stored voice instruction sets and index information corresponding to each group of pre-stored voice instruction sets.
The machine learning operation is executed, and based on the pre-stored driving scenes, the pre-stored user image pair training set and the guiding information in the test set, the guiding information in the test set is classified to obtain a plurality of groups of pre-stored voice instruction sets, and each group of pre-stored voice instruction sets is provided with corresponding index information to establish indexes.
The pre-stored voice instruction set can be understood as another name of the voice guidance instruction set in the voice guidance instruction library.
Step S512, a voice guidance command library is generated according to the plurality of groups of pre-stored voice command sets and the index information.
The voice guidance instruction library is specifically a database of the voice guidance instruction library, and preferably the database is stored at the server side.
It can be understood by those skilled in the art that the generation process of the voice guidance instruction library can be completed at the vehicle-mounted terminal or at a server terminal adapted to the vehicle-mounted terminal.
After the two steps, an index is established for the voice guidance instruction according to the pre-stored driving scene and the pre-stored user image so as to form a voice guidance instruction library, wherein the voice guidance instruction library comprises a plurality of groups of voice guidance instruction sets.
Referring to fig. 6, a flowchart of steps of another embodiment of a voice input guidance method of the present invention is shown, where the input guidance method is based on interaction between a vehicle and an adapted server, and may specifically include the following steps:
step S602, collecting at least one of vehicle speed information, position information, running map data, and gear information of the vehicle as scene configuration information.
Step S604, configuring a driving scene according to the scene configuration information.
Step S606, working condition information and historical voice records of the vehicle machine are obtained.
Step S608, a user image is generated according to the working condition information and the historical voice record.
Step S610, the driving scene and the user image are sent to an adaptive server.
Step S612, the receiving server sends a voice guidance instruction set according to the driving scene and the user portrait.
Step S614, determining the update frequency and update time of the voice guidance instruction set according to the driving scene.
In step S616, it is detected whether the vehicle has an application running in the foreground.
In step S618, if it is not detected that the vehicle has an application running in the foreground, the voice guidance command is displayed according to the update frequency and the update time.
In step S620, in the case where it is detected that the vehicle has an application running in the foreground, execution of the booting operation is stopped.
According to the voice input guiding method, the voice input guiding module is displayed on the desktop of the vehicle computer screen, voice instructions are actively displayed for the user according to the using habit and the current vehicle state of the user, and the user only needs to speak the instructions according to the instructions displayed by the voice skill guiding system.
Specifically, as long as the user does not use the vehicle screen, voice instructions can be directly recommended to the user, the recommendation is performed based on the voice use history of the user and the user portrait, the recommendation of new voice skills is more favored, and aiming at a vehicle-mounted scene, vehicle traffic information is intelligently pushed by combining vehicle condition road conditions, vehicle owner knowledge is widened, and the driving alertness of the user is improved.
It should be noted that, for simplicity of description, the method embodiments are shown as a series of acts, but it should be understood by those skilled in the art that the embodiments are not limited by the order of acts, as some steps may occur in other orders or concurrently in accordance with the embodiments. Further, those skilled in the art will appreciate that the embodiments described in the specification are presently preferred embodiments, and that the acts are not necessarily required by the embodiments of the invention.
Referring to fig. 7, there is shown a schematic block diagram of one embodiment of a voice input guiding apparatus 700 of the present invention, comprising:
an acquisition unit 702 for acquiring a voice guidance instruction set matching with a driving scene of a vehicle and a user portrait of a user;
And an execution unit 704, configured to execute a guidance operation on at least one voice guidance instruction in the voice guidance instruction set, in a case where it is not detected that the vehicle has an application running in the foreground.
The acquiring unit 702 may be integrated in a host of the vehicle, or may be a network module of the vehicle, and the executing unit 704 may be a display module and/or a speaker module.
According to the voice input guiding device provided by the embodiment, the voice guiding instruction set is determined through the acquisition unit 702, so that in the running process of a vehicle, the voice input guiding device on the vehicle-mounted device of the execution unit 704 guides voice input of a user, the requirement that the user sends related voice instructions in the running process is met, and further, travel related information can be pushed to the user based on voice feedback of the user, and the purpose of assisting the user in traveling is achieved.
Optionally, the voice input guiding device 700 further comprises: a configuration unit 706, configured to configure a driving scenario of the vehicle, and/or generate a user portrait of the user.
The configuration unit 706 may be integrated in the host of the vehicle.
Optionally, the configuration unit 706 includes: an acquisition subunit 7062 configured to acquire at least one of vehicle speed information, position information, travel map data, and shift position information of the vehicle as scene configuration information; the configuration unit 706 is further configured to: and configuring a driving scene according to the scene configuration information.
Wherein the acquisition subunit 7062 may be a sensor and/or a data acquisition interface.
Optionally, the configuration unit 706 includes: an acquisition subunit 7064, configured to acquire working condition information and a historical voice record of the vehicle; the generating subunit 7066 is configured to generate a user image according to the working condition information and the historical voice record, where the working condition information includes touch information of the vehicle and/or an application program running on the vehicle.
Wherein the acquisition subunit 7064 and the generation subunit 7066 may be integrated in a host of the vehicle.
Optionally, the execution unit 704 includes: a first determining subunit 7062, configured to determine an update frequency and/or update time of the voice guidance instruction set according to a driving scenario; the update sub-unit 7064 updates at least one voice guidance instruction for performing a guidance operation according to the update frequency and/or the update time.
Wherein the first determining subunit 7062 and the updating subunit 7064 may be integrated in a host of the vehicle. Alternatively, the acquisition unit 702 includes: a detection subunit 7022, configured to detect whether the user portrait matches a pre-stored user portrait; an extraction subunit 7024, configured to extract, from a pre-stored voice guidance instruction library, a voice guidance instruction set that matches the user portrait and the driving scene if the user portrait matches the pre-stored user portrait; a configuration subunit 7026 is configured to configure the voice guidance instruction set according to the cold start policy if the user profile does not match the pre-stored user profile.
The detection subunit 7022, the extraction subunit 7024, and the configuration subunit 7026 may be integrated on a server adapted to the vehicle.
Optionally, the apparatus 700 further comprises: a collecting unit 708 for collecting guidance information, the guidance information including at least one of function instruction information, voice self-learning instruction information, promotion instruction information, and user instruction information; a deleting unit 610, configured to delete information that is not matched with the guidance rule in the guidance information, so as to preserve the guidance information to be processed; a dividing unit 612, configured to divide the information to be processed into a training set and a testing set; the generating unit 614 is configured to perform machine learning according to the training set, the testing set, the pre-stored driving scenario, and the pre-stored user portrait, so as to generate multiple groups of pre-stored voice instruction sets and index information corresponding to each group of pre-stored voice instruction sets; the generating unit 614 is further configured to: and generating a voice guidance instruction library according to the multiple groups of prestored voice instruction sets and the index information.
The collecting unit 708, the deleting unit 710, the dividing unit 712, and the generating unit 714 may be integrated on a server adapted to the vehicle.
Optionally, the configuration subunit 7026 is further configured to: at least one group of pre-stored voice instruction sets with highest use frequency in the voice guidance instruction library is configured as a voice guidance instruction set.
Optionally, the extraction subunit 7024 includes: the second determining subunit 7024A is configured to determine index information with highest matching degree with the user portrait and the driving scene, so as to determine a corresponding voice guidance instruction set according to the index information.
Optionally, the execution unit 704 is further configured to: in the case where it is detected that the vehicle has an application running in the foreground, execution of the booting operation is stopped.
For the device embodiments, since they are substantially similar to the method embodiments, the description is relatively simple, and reference is made to the description of the method embodiments for relevant points.
Referring to fig. 8, there is shown a schematic block diagram of an embodiment of a vehicle 80 of the present invention, comprising: the voice input guiding apparatus 700 described in the above embodiment.
In this specification, each embodiment is described in a progressive manner, and each embodiment is mainly described by differences from other embodiments, and identical and similar parts between the embodiments are all enough to be referred to each other.
It will be apparent to those skilled in the art that embodiments of the present invention may be provided as a method, apparatus, or computer program product. Accordingly, embodiments of the present invention may take the form of an entirely hardware embodiment, an entirely software embodiment or an embodiment combining software and hardware aspects. Furthermore, embodiments of the invention may take the form of a computer program product on one or more computer-usable storage media (including, but not limited to, disk storage, CD-ROM, optical storage, etc.) having computer-usable program code embodied therein.
Embodiments of the present invention are described with reference to flowchart illustrations and/or block diagrams of methods, terminal devices (systems), and computer program products according to embodiments of the invention. It will be understood that each flow and/or block of the flowchart illustrations and/or block diagrams, and combinations of flows and/or blocks in the flowchart illustrations and/or block diagrams, can be implemented by computer program instructions. These computer program instructions may be provided to a processor of a general purpose computer, special purpose computer, embedded processor, or other programmable data processing terminal device to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing terminal device, create means for implementing the functions specified in the flowchart flow or flows and/or block diagram block or blocks.
These computer program instructions may also be stored in a computer-readable memory that can direct a computer or other programmable data processing apparatus to function in a particular manner, such that the instructions stored in the computer-readable memory produce an article of manufacture including instruction means which implement the function specified in the flowchart flow or flows and/or block diagram block or blocks.
These computer program instructions may also be loaded onto a computer or other programmable data processing apparatus to cause a series of operational steps to be performed on the computer or other programmable apparatus to produce a computer implemented process such that the instructions which execute on the computer or other programmable apparatus provide steps for implementing the functions specified in the flowchart flow or flows and/or block diagram block or blocks.
While preferred embodiments of the present invention have been described, additional variations and modifications in those embodiments may occur to those skilled in the art once they learn of the basic inventive concepts. It is therefore intended that the following claims be interpreted as including the preferred embodiment and all such alterations and modifications as fall within the scope of the embodiments of the invention.
Finally, it is further noted that relational terms such as first and second, and the like are used solely to distinguish one entity or action from another entity or action without necessarily requiring or implying any actual such relationship or order between such entities or actions. Moreover, the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or terminal that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or terminal. Without further limitation, an element defined by the phrase "comprising one … …" does not exclude the presence of other like elements in a process, method, article, or terminal device that comprises the element.
The foregoing has outlined a speech input guiding method and a speech input guiding apparatus according to the present invention, and specific examples have been applied to illustrate the principles and embodiments of the present invention, the above examples being only for aiding in the understanding of the method and core idea of the present invention; meanwhile, as those skilled in the art will have variations in the specific embodiments and application scope in accordance with the ideas of the present invention, the present description should not be construed as limiting the present invention in view of the above.

Claims (10)

1. A voice input guiding method for a vehicle, comprising:
Acquiring a voice guidance instruction set matched with a driving scene of a vehicle and a user portrait of a user; the driving scene comprises one of normal driving, overspeed driving, high-speed driving, low-speed driving, temporary stopping and traffic light stopping;
Under the condition that the application program running in the foreground of the vehicle is not detected, determining the updating frequency and/or the updating time of the voice guidance instruction set according to the driving scene;
updating at least one voice guidance instruction for performing guidance operation according to the update frequency and/or update time;
the voice guidance instruction set matched with the driving scene of the vehicle and the user portrait of the user is obtained, and the voice guidance instruction set specifically comprises:
Detecting whether the user portrait matches a pre-stored user portrait;
If the user portrait is matched with the pre-stored user portrait, extracting the voice guidance instruction set matched with the user portrait and the driving scene from a pre-stored voice guidance instruction library;
And if the user portrait is not matched with the pre-stored user portrait, configuring the voice guidance instruction set according to a cold start strategy.
2. The method of claim 1, wherein prior to the obtaining a voice guidance instruction set that matches a driving scenario of a vehicle and a user representation of a user, the method comprises:
configuring a driving scene of the vehicle and/or generating a user portrait of the user.
3. The method of claim 2, wherein configuring a driving scenario of the vehicle comprises:
collecting at least one of speed information, position information, running map data and gear information of the vehicle as scene configuration information;
and configuring the driving scene according to the scene configuration information.
4. The method of claim 2, wherein generating a user representation of the user comprises:
acquiring working condition information and historical voice records of the vehicle and the machine;
generating the user portrait according to the working condition information and the historical voice record,
The working condition information comprises touch information of the vehicle and/or an application program operated by the vehicle.
5. The method of claim 1, wherein prior to the obtaining a voice guidance instruction set that matches the driving scene and the user representation, the method comprises:
Collecting guide information, wherein the guide information comprises at least one of function instruction information, voice self-learning instruction information, popularization instruction information and user instruction information;
deleting the information which is not matched with the guiding rule in the guiding information so as to reserve guiding information to be processed;
dividing the information to be processed into a training set and a testing set;
Executing machine learning according to the training set, the test set, the pre-stored driving scene and the pre-stored user portrait to generate a plurality of groups of pre-stored voice instruction sets and index information corresponding to each group of pre-stored voice instruction sets;
and generating the voice guidance instruction library according to a plurality of groups of pre-stored voice instruction sets and the index information.
6. The method of claim 1, wherein said configuring said voice guidance instruction set according to a cold start strategy comprises:
and configuring at least one group of pre-stored voice instruction sets with highest use frequency in the voice guidance instruction library as the voice guidance instruction set.
7. The method of claim 5, wherein the extracting the voice guidance instruction set matching the user representation and the driving scene from a pre-stored voice guidance instruction library comprises:
and determining the index information with the highest matching degree with the user portrait and the driving scene, and determining the pre-stored voice instruction set corresponding to the index information as the voice guiding instruction set.
8. The method according to any one of claims 1 to 7, further comprising:
And stopping executing the guiding operation under the condition that the vehicle machine is detected to have the application program running in the foreground.
9. A voice input guiding device for a vehicle, comprising:
The acquisition unit is used for acquiring a voice guidance instruction set matched with the driving scene of the vehicle and the user portrait of the user; the driving scene comprises one of normal driving, overspeed driving, high-speed driving, low-speed driving, temporary stopping and traffic light stopping;
The execution unit is used for determining the updating frequency and/or the updating time of the voice guidance instruction set according to the driving scene under the condition that the application program operated by the vehicle with a foreground is not detected; updating at least one voice guidance instruction for performing guidance operation according to the update frequency and/or update time;
The acquisition unit includes:
a detection subunit, configured to detect whether the user portrait matches a pre-stored user portrait;
An extraction subunit, configured to extract, from a pre-stored voice guidance instruction library, a voice guidance instruction set that matches the user representation and the driving scene if the user representation matches the pre-stored user representation;
And the configuration subunit is used for configuring the voice guidance instruction set according to the cold start strategy if the user portrait is not matched with the pre-stored user portrait.
10. A vehicle machine, comprising:
The voice input guiding device of claim 9.
CN202010519922.9A 2020-06-09 2020-06-09 Voice input guiding method, device and car machine Active CN113779300B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010519922.9A CN113779300B (en) 2020-06-09 2020-06-09 Voice input guiding method, device and car machine

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010519922.9A CN113779300B (en) 2020-06-09 2020-06-09 Voice input guiding method, device and car machine

Publications (2)

Publication Number Publication Date
CN113779300A CN113779300A (en) 2021-12-10
CN113779300B true CN113779300B (en) 2024-05-07

Family

ID=78834526

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010519922.9A Active CN113779300B (en) 2020-06-09 2020-06-09 Voice input guiding method, device and car machine

Country Status (1)

Country Link
CN (1) CN113779300B (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116016578B (en) * 2022-11-22 2024-04-16 中国第一汽车股份有限公司 Intelligent voice guiding method based on equipment state and user behavior

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104335152A (en) * 2012-06-05 2015-02-04 苹果公司 Providing navigation instructions while device is in locked mode
WO2015079331A1 (en) * 2013-11-28 2015-06-04 Sony Corporation Application activation method and apparatus and electronic equipment
CN108766423A (en) * 2018-05-25 2018-11-06 三星电子(中国)研发中心 A kind of active awakening method and device based on scene
CN109377115A (en) * 2018-12-19 2019-02-22 Oppo广东移动通信有限公司 Vehicular applications recommended method, device, terminal device and storage medium
CN110096249A (en) * 2018-01-31 2019-08-06 阿里巴巴集团控股有限公司 Methods, devices and systems for prompting fast to wake up word
CN110275692A (en) * 2019-05-20 2019-09-24 北京百度网讯科技有限公司 A kind of recommended method of phonetic order, device, equipment and computer storage medium
CN110472095A (en) * 2019-08-16 2019-11-19 百度在线网络技术(北京)有限公司 Voice guide method, apparatus, equipment and medium
CN110784833A (en) * 2019-08-29 2020-02-11 腾讯科技(深圳)有限公司 Message reminding method and device, vehicle-mounted equipment and storage medium

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9997069B2 (en) * 2012-06-05 2018-06-12 Apple Inc. Context-aware voice guidance

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104335152A (en) * 2012-06-05 2015-02-04 苹果公司 Providing navigation instructions while device is in locked mode
WO2015079331A1 (en) * 2013-11-28 2015-06-04 Sony Corporation Application activation method and apparatus and electronic equipment
CN110096249A (en) * 2018-01-31 2019-08-06 阿里巴巴集团控股有限公司 Methods, devices and systems for prompting fast to wake up word
CN108766423A (en) * 2018-05-25 2018-11-06 三星电子(中国)研发中心 A kind of active awakening method and device based on scene
CN109377115A (en) * 2018-12-19 2019-02-22 Oppo广东移动通信有限公司 Vehicular applications recommended method, device, terminal device and storage medium
CN110275692A (en) * 2019-05-20 2019-09-24 北京百度网讯科技有限公司 A kind of recommended method of phonetic order, device, equipment and computer storage medium
CN110472095A (en) * 2019-08-16 2019-11-19 百度在线网络技术(北京)有限公司 Voice guide method, apparatus, equipment and medium
CN110784833A (en) * 2019-08-29 2020-02-11 腾讯科技(深圳)有限公司 Message reminding method and device, vehicle-mounted equipment and storage medium

Also Published As

Publication number Publication date
CN113779300A (en) 2021-12-10

Similar Documents

Publication Publication Date Title
US9875741B2 (en) Selective speech recognition for chat and digital personal assistant systems
CN107305769B (en) Voice interaction processing method, device, equipment and operating system
US20190304466A1 (en) Voice control method, voice control device and computer readable storage medium
EP3166023A1 (en) In-vehicle interactive system and in-vehicle information appliance
CN105989841B (en) Vehicle-mounted voice control method and device
CN111767021A (en) Voice interaction method, vehicle, server, system and storage medium
CN110286745B (en) Dialogue processing system, vehicle with dialogue processing system, and dialogue processing method
CN104239442B (en) Search result shows method and apparatus
CN102202082A (en) Vehicle-mounted communication system and method
CN113590769B (en) State tracking method and device in task-driven multi-round dialogue system
CN105335414A (en) Music recommendation method, device and terminal
CN105719648A (en) Individual automated vehicle interaction method and automated vehicle
CN110767219B (en) Semantic updating method, device, server and storage medium
CN111414512A (en) Resource recommendation method and device based on voice search and electronic equipment
US12038299B2 (en) Content-aware navigation instructions
CN105161112B (en) Audio recognition method and device
CN113779300B (en) Voice input guiding method, device and car machine
US20200286479A1 (en) Agent device, method for controlling agent device, and storage medium
CN112242143A (en) Voice interaction method and device, terminal equipment and storage medium
CN107247518A (en) Method, device, equipment and the readable storage medium storing program for executing of polynary association's input
KR20160051983A (en) Music recommendation system for vehicle and method thereof
CN112825256B (en) Guiding method, device, equipment and computer storage medium for recording voice packet function
CN112562658A (en) Groove filling method and device
KR102109186B1 (en) Method ana apparatus for recommendation contents based on conversation of vehicle's occupant
US11620994B2 (en) Method for operating and/or controlling a dialog system

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant