EP3944232A1 - Voice control for autonomous vehicles - Google Patents
Voice control for autonomous vehicles Download PDFInfo
- Publication number
- EP3944232A1 EP3944232A1 EP20205245.2A EP20205245A EP3944232A1 EP 3944232 A1 EP3944232 A1 EP 3944232A1 EP 20205245 A EP20205245 A EP 20205245A EP 3944232 A1 EP3944232 A1 EP 3944232A1
- Authority
- EP
- European Patent Office
- Prior art keywords
- vehicle
- command
- navigation command
- control
- navigation
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Withdrawn
Links
Images
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/22—Procedures used during a speech recognition process, e.g. man-machine dialogue
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W60/00—Drive control systems specially adapted for autonomous road vehicles
- B60W60/001—Planning or execution of driving tasks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/16—Sound input; Sound output
- G06F3/167—Audio in a user interface, e.g. using voice commands for navigating, audio feedback
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L17/00—Speaker identification or verification
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W2420/00—Indexing codes relating to the type of sensors based on the principle of their operation
- B60W2420/54—Audio sensitive means, e.g. ultrasound
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W2540/00—Input parameters relating to occupants
- B60W2540/043—Identity of occupants
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/22—Procedures used during a speech recognition process, e.g. man-machine dialogue
- G10L2015/223—Execution procedure of a spoken command
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/22—Procedures used during a speech recognition process, e.g. man-machine dialogue
- G10L2015/225—Feedback of the input speech
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/22—Procedures used during a speech recognition process, e.g. man-machine dialogue
- G10L2015/226—Procedures used during a speech recognition process, e.g. man-machine dialogue using non-speech characteristics
Definitions
- This disclosure relates to a method and apparatus for in-vehicle voice control for autonomous vehicles.
- Autonomous vehicles or self-driving vehicles may allow a vehicle to carry passenger to a desired destination without the need for a driver.
- the Society of Automotive Engineers publication "Taxonomy and Definitions for Terms Related to Driving Automation Systems for On-Road Motor Vehicles J3016TM", issued 16th January 2014 and revised 15th June 2018 defines six levels of automation.
- a vehicle with level 4 automation may operate safely in predefined zones without any human intervention, and a vehicle with level 5 automation may operate safely anywhere.
- Such a vehicle may be operated for example as a robot taxi.
- a method for controlling an autonomous vehicle comprising: detecting an audio signal; processing the audio signal to determine whether the speech comprises a vehicle-navigation command; determining whether the audio signal comprises speech from an authorised occupant of the vehicle; in response to the speech comprising an authorised vehicle-navigation command, determining whether execution of the vehicle-navigation command is feasible and executing the vehicle-navigation command if feasible to control at least one vehicle actuator to control the direction and speed of the autonomous vehicle.
- the method may comprise providing feedback to the authorized occupant if execution of the vehicle navigation command is not feasible.
- determining whether the audio signal comprises speech of an authorised occupant of the vehicle further comprises determining the seat location of the occupant.
- the method may comprise comparing the voice signature of the received speech with a predetermined speech signature database and determining if an occupant is an authorised occupant from the comparison.
- determining whether execution of the vehicle-navigation command is feasible may comprise: receiving navigation data via at least one of an in-vehicle navigation system and vehicle-to-x (V2X) network; and obtaining environmental context attributes from the received navigation data.
- V2X vehicle-to-x
- the method may comprise receiving an abort instruction for a previously issued command; determining whether the previously issued command is in an execution phase; cancelling the previously issued command if the previously issued command is not in an execution phase.
- an audio device comprising: a microphone; a transceiver; a processor coupled to the transceiver and the microphone and configured to: process the audio signal to determine whether the speech comprises a vehicle-navigation command; transmit a control instruction comprising the vehicle-navigation command to a vehicle control system coupled to the audio device, the vehicle control system being configured to determine whether the vehicle-navigation command is authorised; and in response to determining that the vehicle-navigation command is authorised, control at least one actuator in the vehicle dependent on the authorized vehicle-navigation command.
- the audio device may be further configured to: receive a location identifier via the transceiver from a location identification transmitter located in a vehicle; and wherein the control instruction further comprises the received location identifier and wherein the vehicle control system is further configured to determine whether the audio signal comprises speech from an authorised occupant of the vehicle from the location of the occupant provided by the location identification transmitter.
- the audio device may be configured as a mobile hearable device and comprise a wireless transceiver configured as one of a near field magnetic induction, NFMI, transceiver, a near field electromagnetic induction, NFEMI, transceiver and an Ultra-Wide-Band, UWB, transceiver.
- a wireless transceiver configured as one of a near field magnetic induction, NFMI, transceiver, a near field electromagnetic induction, NFEMI, transceiver and an Ultra-Wide-Band, UWB, transceiver.
- a vehicle control apparatus for an autonomous vehicle comprising: a vehicle function controller configured to be coupled to one or more vehicle actuators and comprising a vehicle control transceiver configured to be coupled to an audio device; wherein the vehicle control transceiver is configured to receive a vehicle-navigation command from the audio device and is further configured to determine whether the vehicle-navigation command is authorized and to control the one or more vehicle actuators to control the direction and speed of the autonomous vehicle dependent on the authorized vehicle-navigation command.
- the vehicle control apparatus may further comprise at least one location identification transmitter located in the autonomous vehicle wherein each of the at least one location identification transmitters are configured to transmit a respective location identifier for authorizing a vehicle-navigation command to the audio device.
- the at least one location identification transmitter may comprise a wireless transceiver configured as one of a near field magnetic induction, NFMI, transceiver, a near field electromagnetic induction, NFEMI, transceiver and an Ultra-Wide-Band, UWB, transceiver.
- NFMI near field magnetic induction
- NFEMI near field electromagnetic induction
- UWB Ultra-Wide-Band
- Embodiments of the audio input device and vehicle control apparatus may be included in a system for controlling an autonomous vehicle.
- a non-transitory computer readable media comprising a computer program comprising computer executable instructions which, when executed by a processing unit, causes the processing unit to control an autonomous vehicle by performing the steps of :detecting an audio signal; processing the audio signal to determine whether the speech comprises a vehicle-navigation command; determining whether the audio signal comprises speech from an authorised occupant of a vehicle; in response to the speech comprising an authorised vehicle-navigation command, determining whether execution of the vehicle-navigation command is feasible and executing the vehicle-navigation command if feasible to control at least one vehicle actuator to control the direction and speed of the autonomous vehicle.
- the non-transitory computer readable media may further comprise instructions to perform the steps of providing feedback to the authorized occupant if execution of the vehicle navigation command is not feasible.
- the non-transitory computer readable media may further comprise instructions to determine whether the audio signal comprises speech of an authorised occupant of the vehicle be determining the seat location of the occupant.
- the non-transitory computer readable media may further comprise instructions to provide an indication to the user that the action is prohibited in response to determining that the execution of the vehicle is not feasible.
- the non-transitory computer readable media may further comprise instructions to compare the voice signature of the received speech with a predetermined speech signature database and determine if an occupant is an authorised occupant from the comparison.
- the non-transitory computer readable media may further comprise instructions to receive an abort instruction for a previously issued command; determine whether the previously issued command is in an execution phase; and cancel the previously issued command if the previously issued command is not in an execution phase.
- FIG. 1 illustrates a system 100 for in-vehicle voice control of an autonomous vehicle according to an example embodiment.
- the system 100 includes an audio input device 110 coupled to a microphone 102 that captures a voice command which may be a vehicle-navigation command and provides it to the vehicle electronics implementing a vehicle function controller 120.
- the audio input device 110 and microphone 102 may be referred to herein as a microphone unit.
- the audio input device 110 may include a voice command detector 104, a voice command processor and interpreter 106 and an instruction broadcast module 108.
- the vehicle function controller 120 may be connected to one or more actuators 140 such as a steering actuator 142 braking actuator 144 and speed actuator 146 to control the navigation of the autonomous vehicle.
- the vehicle function controller 120 may include an instruction processing and interpretation module 122, an authorization module 124 and a feasibility check and execution module 126.
- the vehicle function controller 120 may also include a vehicle control transceiver (not shown) which may be a wired or wireless transceiver for connecting the audio input device 110.
- the vehicle function controller 120 may also be connected to a navigation module 150.
- the navigation module 150 may provide Global Positioning System (GPS) location information and other environment context information such as parking locations, roadworks, and other information used for the navigation of the autonomous vehicle.
- GPS Global Positioning System
- the vehicle function controller 120 may connect via a wired or wireless vehicle control transceiver (not shown) to an audio output device 130 including an audio processor 132.
- the audio output device 130 may be connected to a speaker 134 to provide acoustic feedback to the passengers of the autonomous vehicle.
- the audio output device 130 and the speaker 134 may be referred to herein as a speaker unit.
- the speaker unit may be incorporated in the vehicle or be part of a mobile hearable device.
- the vehicle function controller 120 may be connected to a screen for visual feedback instead of or as well as an audio output device 130.
- the microphone unit and the speaker unit may be part of a mobile hearable device such as a hearing aid or other hearing device which may be wirelessly coupled to the vehicle function controller 120 for example by transmitter which may be a near field magnetic induction (NFMI), a near field electromagnetic induction (NFEMI), Bluetooth, or Ultra-Wide-Band (UWB) transmitter.
- NFMI refers to any communication means that makes use of transmissions within a localized magnetic field.
- NFEMI is an extension of NFMI which also uses an electric antenna for transmissions as explained for example in US 9197986B1 and European Patent application 15164621.3 .
- the audio input device 110 may receive the audio signal from the microphone 102.
- the audio input device 110 may perform voice command detection with voice command detector 104 followed by voice command processing and interpretation by the voice command processing and interpreter 106 .
- the results of the voice command processing and interpretation may be broadcast by the broadcast instruction module 108 to the vehicle function controller 120.
- the assembled instruction may be output from the audio input device 110 and broadcast via a wired or wireless connection to the vehicle function controller 120.
- the vehicle function controller 120 may receive and process the instruction with instruction processing and interpretation module 122. Following instruction processing and interpretation a check is made by the authorization module 124 to determine to identify whether voice commands are provided by an authorized occupant of the vehicle. In some examples, identification of an authorized person may be based on either seat location, biometric data, from voice recognition, visual information.
- the feasibility check and execution module 126 may output an audio signal to the audio output device 130 if the command is not authorised or it is not feasible to execute the command for some other reason. If the command is authorised and feasible, the command may be executed to control the actuators 140 in order to navigate the autonomous vehicle based on information from the navigation module 150 and the authorized command. Audio input device 110, vehicle function controller 120 and audio output device 130 may be implemented in hardware or a combination of hardware and software.
- Figure 2 describes a method 200 of operating the audio input device 110.
- the autonomous vehicle may be switched on.
- an optional pairing action may occur between the audio input device 110 and the vehicle function controller 120. This may be done for examples of the audio input device which are wirelessly coupled to the vehicle function controller 120 using for example a Bluetooth, UWB, NFMI or NFEMI wireless connection. In examples when the audio input device 110 is fixed in the vehicle interior or connected by a wired connection, the pairing may not be required.
- the audio input device 110 may be in a listening mode.
- the audio input device may determine whether a voice has been detected. If a voice has not been detected, the method may return to step 206.
- step 210 the method may proceed to step 210 and extract keywords from the speech. Keyword extraction may be implemented for example using temporal convolution networks as described in " Temporal Convolutional Networks for Action Segmentation and Detection", Colin Lea et al, 16th November 2016, arXiv:1611.05267 . An extracted keyword may be compared against known vehicle-navigation commands in step 212. If a known command has not been detected, then the method returns to step 206. If a known command has been detected then the method proceeds to step 214 in which a command instruction may be assembled including the command keyword and additional authentication information such as for example the seat location of the occupant. In step 216 the assembled command may be transmitted to the vehicle function controller. Following step 216, the method may return to listening mode step 206.
- FIG. 3 Shows an example implementation of a microphone unit 300 which may implement the method 200.
- Microphone unit 300 includes first and second microphone 302, 302' connected to a respective microphone interface 304. 304'.
- the microphone interface outputs may be connected to a digital signal processor 306.
- the output of the digital signal processor 306 may be connected to a wired or wireless transceiver 308 for communication with for example vehicle function controller 120.
- the microphone unit may have two microphones which for example may allow beamforming which may for example be implemented by software executed on the digital signal processor 306. Using beamforming may improve the discrimination between the intended user and background noise or speech from other occupants of the vehicle.
- Figure 4 shows an example method of operation 400 which may be implemented by the vehicle function controller 120.
- an instruction from the microphone unit may be received by the vehicle function controller.
- the instruction may be processed to check whether it includes a valid command
- the extracted keyword may be compared with the database keyword instruction set to make a comparison if the keyword is valid.
- a list of valid commands may be provide in database 408.
- Step 404 may be implemented for example by the instruction processing and interpretation module 122.
- an authorization check is made to determine whether a command is provided by an authorized person.
- An authorized person may be a person that wears a paired hearable/wearable or a person that is located on a given seat.
- Authorised person and location information may be stored in database 408.
- authorization is done by pairing so that only voice commands that are provided from a paired device are authorized for execution.
- authorization may be based on Seating location.
- navigation may be authorized from a given seat within the vehicle.
- navigation may only be permitted from a particular seat location. Alternatively, navigation could be handed-over to be done from any other seat within the vehicle.
- the detection of the seat location may be implemented for example as described in European patent application 19187261.3 , where a seat ID transceiver is part of the seat.
- a seat ID transceiver is part of the seat.
- the voice signature of the authorized person may be stored as part of the authenticated persons/seats database.
- a signature may be stored for all of the keywords.
- the authentication may include face recognition. Once a voice command has been provided together with a seat location, a check is done via a camera system whether the occupant of the vehicle that provided the voice command is authorized to navigate the vehicle. In other examples a centralized microphone system may be used that includes a direction of where the voice command came from.
- a microphone may be paired with more than one microphone. For example, there could be a microphone on each side (left/right) of a centre microphone. From the relative signal strengths and time delay of the audio signal detected by each microphone, the direction of the audio source may be determined.
- step 410 checks if the command is valid.
- the voice command is classified as invalid when it has been provided by a non-authorized person, or when it is not a valid navigation instruction for the vehicle. If the command is invalid, the method proceeds to step 412 and an audio signal is generated for example an alert sound or a generated speech message to be provided to the vehicle occupants. This audio signal is then output on a speaker in step 414 which may be implemented using the audio output device 130 and speaker 134.
- the navigation instruction may be captured in step 416 and stored for execution in step 418.
- Authorization and validity checks carried out in steps 406 and 410 may be implemented for example by authorization module 124.
- Figure 5 shows an example method of operation 500 which may be implemented by the vehicle function controller 120.
- Method 500 indicates the operation for a navigation action process which may be implemented for example by feasibility check and execution module 126.
- the commands are Turn Left, Turn Right, Stop, and Park but it will be appreciated that other commands may also be similarly processed.
- the navigation instruction for execution which may be determined by method 400 is captured in step 514 and processed in step 516. If the command is to " turn left " the method proceeds to step 518 and in step 526 a determination is made whether a route exists that allows the vehicle to comply with the instruction. This determination step 526 may require environmental information to provide the information about the context in which the vehicle is operating.
- This environmental information may be provided via a V2X connection 502, V2X transceiver 504 and V2X processor 506.
- the environmental information may be provided by navigation maps 512.
- the information from V2X processor 506 is provided to the decision step 526 schematically by the connection 508.
- Map information for route details and path planning may be provided by connection 510. If the left route exists the method proceeds to an execution step 534. For the left turn instruction in step 536 the actuators of the autonomous vehicle may be controlled to re-route. If the left route does not exist, for example due to a closed street, safety related aspects such as no turn or stop permitted in the current location, the method proceeds to step 544 and an indication is provided to the used that the command cannot be serviced. This indication may be a computer generated speech or other audio indication.
- step 520 a determination is made whether a route exists that allows the vehicle to comply with the instruction. If the route exists the method proceeds to an execution step 534. For the right turn instruction in step 538 the actuators of the autonomous vehicle may be controlled to re-route. If the left route does not exist, the method proceeds to step 544 as previously described.
- step 522 the method proceeds to step 522 and in step 530 a determination is made whether a stopping zone exists that allows the vehicle to comply with the instruction. If the stopping zone does not exist the method proceeds to an execution step 534. In step 540 actuators in the autonomous vehicle may be controlled to execute the stop instruction. If it is not possible to execute the stop instruction, the method proceeds to step 544 as previously described.
- step 524 a determination is made whether a stopping zone exists that allows the vehicle to comply with the instruction. If the stopping zone does not exist the method proceeds to an execution step 534 In step 54 actuators in the autonomous vehicle may be controlled to execute the park instruction. If it is not possible to execute the stop instruction, the method proceeds to step 544 as previously described.
- Figures 6A and 6B illustrate an example operation of an " abort " command.
- Figure 6A shows the possible phases of operation 600.
- a first phase or T1 phase 602 the voice command capturing and processing is in progress.
- a second phase 604 or T2 phase the execution is in progress i.e. the action is being performed.
- FIG. 6B shows an example abort process 610.
- step 612 the captured navigation abort command is received
- step 614 a check is made to determine whether the received abort command is in the T1 phase 602. If the abort command is in the T1 phase 602 then in step 616, the previous instruction command is cancelled and in step 618 no control instructions are issued by the vehicle function controller. In step 620 an indication is generated that the abort operation was successful.
- the actions of the current instruction continue to be executed in step 622 and in step 624 the user receives feedback that the abort operation was not done and the previous instruction has been executed.
- Figures 7A, 7B and 7C illustrate a further example detail of the execution steps 534 shown in figure 5 .
- FIG. 7A shows a method 700 of executing a stop instruction which may be used to implement execution step 540 in method 500.
- the method may control the actuators to reduce speed and initiate a lane change.
- a check is made to determine whether the vehicle is on a high speed lane. If the vehicle is not on a high speed lane then the method proceeds to step 708 and stop light or hazard indicators may be activated.
- the method proceeds to step 706 and the vehicle actuators are controlled to move the autonomous vehicle to a low speed lane and then to step 708.
- the actuators may be controlled to move the vehicle to a stop zone in step 710 and then in step 712 the brakes are actuated to halt the vehicle.
- the completion of the command may be indicated by generating an audio output.
- Figure 7B shows a method of executing a park instruction 720 which may be used to implement execution step 542 in method 500.
- the method may control the actuators to reduce speed and initiate a lane change.
- a check is made to determine whether the vehicle is on a high speed lane. If the vehicle is not on a high speed lane then the method proceeds to step 728 and stop light or hazard indicators may be activated.
- the method proceeds to step 726 and the vehicle actuators are controlled to move the autonomous vehicle to a low speed lane and then to step 728.
- the actuators may be controlled to move the vehicle to a stop zone in step 730 and then in step 732 the brakes are actuated to halt the vehicle.
- the completion of the command may be indicated by generating an audio output.
- Figure 7C shows a method of executing a re-route instruction 740 which may be used to implement steps 536 and 538.
- navigation controls may be adapted depending on the command.
- step 744 the successful service of the command may be indicated.
- FIG. 8 illustrates a system 800 for in-vehicle voice control of an autonomous vehicle according to an example embodiment.
- the system 800 may implement one or more of the methods 200, 400, 500, 510, 610, 700, 720.
- the system 800 includes an audio input device 810 coupled to a microphone 802 that captures a voice command which may be a vehicle-navigation command and provides it to the vehicle electronics implementing a vehicle function controller 820.
- the audio input device 810 and microphone 802 may be referred to herein as a microphone unit.
- the audio input device 810 may include a voice command detector 804, a voice command processor and interpreter 806 an instruction assembler 812 and an instruction broadcast module 808.
- the vehicle function controller 820 may be connected to one or more actuators 840 such as a steering actuator 842 braking actuator 844 and speed actuator 846 to control the navigation of the autonomous vehicle.
- the vehicle function controller 820 may include an instruction processing and interpretation module 822, an authorization module 824 and a feasibility check and execution module 826.
- the vehicle function controller 820 may also include a vehicle control transceiver (not shown) which may be a wired or wireless transceiver for connecting the audio input device 810.
- the vehicle function controller 820 may also be connected to a navigation module 850.
- the navigation module 850 may provide GPS location information and other environment context information such as parking locations, roadworks, and other information used for the navigation of the autonomous vehicle.
- the vehicle function controller 820 may have an output connected by a wired or wireless connection to an audio output device 830 including an audio processor 832.
- the audio output device 830 may be connected to a speaker 834 to provide acoustic feedback to the passengers of the autonomous vehicle.
- the audio output device 830 and the speaker 934 may be referred to herein as a speaker unit.
- the speaker unit may be incorporated in the vehicle or be part of a mobile hearable device.
- the microphone unit and the speaker unit may be part of a mobile hearable device such as a hearing aid or other hearing device which may be wirelessly coupled to the vehicle function controller 820 for example by transmitter which may be a near field magnetic induction (NFMI) transmitter or a near field electromagnetic induction transmitter (NFEMI), Bluetooth and Ultra-Wide-Band (UWB) communication.
- NFMI near field magnetic induction
- NFEMI near field electromagnetic induction transmitter
- UWB Ultra-Wide-Band
- the audio input device 810 may receive the audio signal from the microphone 802.
- the audio input device 810 may perform voice command detection with voice command detector followed by voice command processing and interpretation.
- the instruction assemble module 812 may receive a seat identifier for example from a seat identification transceiver which may be wirelessly coupled to the audio input device 810.
- the instruction assemble module 812 may receive biomarker data from the biomarker module 852 .
- the biomarker module 852 may be included in the vehicle.
- the biomarker module 852 may be in the audio input device 810.
- the results of the voice command processing and interpretation instruction may be broadcast by the broadcast instruction module 808 to the vehicle function controller 820.
- the assembled instruction may be output from the audio input device 810 and broadcast via a wired or wireless connection to the vehicle function controller 820.
- the vehicle function controller 820 may receive and process the instruction with instruction processing and interpretation module 822. Following instruction processing and interpretation a check is made by the authorization module 824 to determine to identify whether voice commands are provided by an authorized occupant of the vehicle. In some examples, identification of an authorized person may be based on either seat location, biometric data, from voice recognition, visual information.
- the feasibility check and execution module 826 may output an audio signal to the audio output device 830 if the command is not authorised or it is not feasible to execute the command for some other reason such. If the command is authorised and feasible, the command may be executed to control the actuators 840 in order to navigate the autonomous vehicle based on information from the navigation module 850 and the authorized command. Audio input device 810, vehicle function controller 820 and audio output device 830 may be implemented in hardware or a combination of hardware and software.
- Example embodiments described provide an intelligent voice-based human-machine interface for vehicles that enable control the navigation functionality of the vehicle from personalized zones of passengers inside vehicles. Any occupant of the vehicle is enabled to interact with the vehicle by means of voice commands, for navigating the vehicle to destination from their personal seating zone.
- the control actuators include, but are not limited to, Speed Control Unit, Brake Control Unit, Steering Control Unit, Start / Stop.
- Embodiments described herein may allow the direct connection between an occupant of the vehicle providing the voice command comprising at least a navigation command to a Navigation control unit of Vehicle electronics, the authorization of an occupant of the vehicle to allow execution of the given navigation instruction.
- a voice command comprising at least a navigation command may be validated by Vehicle electronics for the allowable Navigation control commands.
- Navigation commands are checked for feasibility of execution in conjunction with at least predefined navigation maps and V2X information. Adapting the navigation controls like steering, speed etc. in conjunction with at least V2X and navigation maps for path planning and rerouting to the new destination in Realtime. The command may be aborted before the navigation controls execute the instruction and provide the feedback to the occupants of the vehicle.
- One or more embodiments may enable lane change assist, indicators etc., for path rerouting, stop or park command.
- One or more embodiments may enable navigating to stop zone, parking zone based on V2X information and initiating the stop or parking sequence.
- the audio control system as described herein may be used with autonomous vehicles configured for example as a robot taxi.
- a method and apparatus for controlling an autonomous vehicle is described.
- An audio signal is detected and processed the to determine whether the speech comprises a vehicle-navigation command.
- the audio signal is analyzed to determine whether the signal includes speech from an authorised occupant of the vehicle. If the speech includes an authorised vehicle-navigation command which is feasible to execute given the current position of the vehicle, the command is executed to control the direction and speed of the autonomous vehicle.
- the set of instructions/method steps described above are implemented as functional and software instructions embodied as a set of executable instructions which are effected on a computer or machine which is programmed with and controlled by said executable instructions. Such instructions are loaded for execution on a processor (such as one or more CPUs).
- processor includes microprocessors, microcontrollers, processor modules or subsystems (including one or more microprocessors or microcontrollers), or other control or computing devices.
- a processor can refer to a single component or to plural components.
- the set of instructions/methods illustrated herein and data and instructions associated therewith are stored in respective storage devices, which are implemented as one or more non-transient machine or computer-readable or computer-usable storage media or mediums.
- Such computer-readable or computer usable storage medium or media is (are) considered to be part of an article (or article of manufacture).
- An article or article of manufacture can refer to any manufactured single component or multiple components.
- the non-transient machine or computer usable media or mediums as defined herein excludes signals, but such media or mediums may be capable of receiving and processing information from signals and/or other transient mediums.
- Example embodiments of the material discussed in this specification can be implemented in whole or in part through network, computer, or data based devices and/or services. These may include cloud, internet, intranet, mobile, desktop, processor, look-up table, microcontroller, consumer equipment, infrastructure, or other enabling devices and services. As may be used herein and in the claims, the following non-exclusive definitions are provided.
- one or more instructions or steps discussed herein are automated.
- the terms automated or automatically mean controlled operation of an apparatus, system, and/or process using computers and/or mechanical/electrical devices without the necessity of human intervention, observation, effort and/or decision.
Abstract
A method and apparatus for controlling an autonomous vehicle is described. An audio signal is detected and processed the to determine whether the speech comprises a vehicle-navigation command. The audio signal is analyzed to determine whether the signal includes speech from an authorised occupant of the vehicle. If the speech includes an authorised vehicle-navigation command which is feasible to execute given the current position of the vehicle, the command is executed to control the direction and speed of the autonomous vehicle.
Description
- This disclosure relates to a method and apparatus for in-vehicle voice control for autonomous vehicles.
- Autonomous vehicles or self-driving vehicles may allow a vehicle to carry passenger to a desired destination without the need for a driver. The Society of Automotive Engineers publication "Taxonomy and Definitions for Terms Related to Driving Automation Systems for On-Road Motor Vehicles J3016™", issued 16th January 2014 and revised 15th June 2018 defines six levels of automation. For example, a vehicle with level 4 automation may operate safely in predefined zones without any human intervention, and a vehicle with level 5 automation may operate safely anywhere. Such a vehicle may be operated for example as a robot taxi.
- Various aspects of the disclosure are defined in the accompanying claims.
- In a first aspect there is provided a method for controlling an autonomous vehicle, the method comprising: detecting an audio signal; processing the audio signal to determine whether the speech comprises a vehicle-navigation command; determining whether the audio signal comprises speech from an authorised occupant of the vehicle; in response to the speech comprising an authorised vehicle-navigation command, determining whether execution of the vehicle-navigation command is feasible and executing the vehicle-navigation command if feasible to control at least one vehicle actuator to control the direction and speed of the autonomous vehicle.
- In one or more embodiments the method may comprise providing feedback to the authorized occupant if execution of the vehicle navigation command is not feasible.
- In one or more embodiments determining whether the audio signal comprises speech of an authorised occupant of the vehicle further comprises determining the seat location of the occupant.
- In one or more embodiments in response to determining that the execution of the vehicle is not feasible, providing an indication to the user that the action is prohibited.
- In one or more embodiments the method may comprise comparing the voice signature of the received speech with a predetermined speech signature database and determining if an occupant is an authorised occupant from the comparison.
- In one or more embodiments determining whether execution of the vehicle-navigation command is feasible may comprise: receiving navigation data via at least one of an in-vehicle navigation system and vehicle-to-x (V2X) network; and obtaining environmental context attributes from the received navigation data.
- In one or more embodiments the method may comprise receiving an abort instruction for a previously issued command; determining whether the previously issued command is in an execution phase; cancelling the previously issued command if the previously issued command is not in an execution phase.
- In a second aspect there is provided an audio device comprising: a microphone; a transceiver; a processor coupled to the transceiver and the microphone and configured to: process the audio signal to determine whether the speech comprises a vehicle-navigation command; transmit a control instruction comprising the vehicle-navigation command to a vehicle control system coupled to the audio device, the vehicle control system being configured to determine whether the vehicle-navigation command is authorised; and in response to determining that the vehicle-navigation command is authorised, control at least one actuator in the vehicle dependent on the authorized vehicle-navigation command.
- In one or more embodiments, the audio device may be further configured to: receive a location identifier via the transceiver from a location identification transmitter located in a vehicle; and wherein the control instruction further comprises the received location identifier and wherein the vehicle control system is further configured to determine whether the audio signal comprises speech from an authorised occupant of the vehicle from the location of the occupant provided by the location identification transmitter.
- In one or more embodiments, the audio device may be configured as a mobile hearable device and comprise a wireless transceiver configured as one of a near field magnetic induction, NFMI, transceiver, a near field electromagnetic induction, NFEMI, transceiver and an Ultra-Wide-Band, UWB, transceiver.
- In a third aspect there is provided a vehicle control apparatus for an autonomous vehicle comprising: a vehicle function controller configured to be coupled to one or more vehicle actuators and comprising a vehicle control transceiver configured to be coupled to an audio device; wherein the vehicle control transceiver is configured to receive a vehicle-navigation command from the audio device and is further configured to determine whether the vehicle-navigation command is authorized and to control the one or more vehicle actuators to control the direction and speed of the autonomous vehicle dependent on the authorized vehicle-navigation command.
- In one or more embodiments, the vehicle control apparatus may further comprise at least one location identification transmitter located in the autonomous vehicle wherein each of the at least one location identification transmitters are configured to transmit a respective location identifier for authorizing a vehicle-navigation command to the audio device.
- In one or more embodiments, the at least one location identification transmitter may comprise a wireless transceiver configured as one of a near field magnetic induction, NFMI, transceiver, a near field electromagnetic induction, NFEMI, transceiver and an Ultra-Wide-Band, UWB, transceiver.
- Embodiments of the audio input device and vehicle control apparatus may be included in a system for controlling an autonomous vehicle.
- In a fourth aspect there is provided a non-transitory computer readable media comprising a computer program comprising computer executable instructions which, when executed by a processing unit, causes the processing unit to control an autonomous vehicle by performing the steps of :detecting an audio signal; processing the audio signal to determine whether the speech comprises a vehicle-navigation command; determining whether the audio signal comprises speech from an authorised occupant of a vehicle; in response to the speech comprising an authorised vehicle-navigation command, determining whether execution of the vehicle-navigation command is feasible and executing the vehicle-navigation command if feasible to control at least one vehicle actuator to control the direction and speed of the autonomous vehicle.
- In one or more embodiments, the non-transitory computer readable media may further comprise instructions to perform the steps of providing feedback to the authorized occupant if execution of the vehicle navigation command is not feasible.
- In one or more embodiments, the non-transitory computer readable media may further comprise instructions to determine whether the audio signal comprises speech of an authorised occupant of the vehicle be determining the seat location of the occupant.
- In one or more embodiments, the non-transitory computer readable media may further comprise instructions to provide an indication to the user that the action is prohibited in response to determining that the execution of the vehicle is not feasible.
- In one or more embodiments, the non-transitory computer readable media may further comprise instructions to compare the voice signature of the received speech with a predetermined speech signature database and determine if an occupant is an authorised occupant from the comparison.
- In one or more embodiments, the non-transitory computer readable media may further comprise instructions to receive an abort instruction for a previously issued command; determine whether the previously issued command is in an execution phase; and cancel the previously issued command if the previously issued command is not in an execution phase.
- In the figures and description like reference numerals refer to like features. Embodiments are now described in detail, by way of example only, illustrated by the accompanying drawings in which:
-
Figure 1 Shows a system for in-vehicle voice control of an autonomous vehicle according to an embodiment. -
Figure 2 Illustrates an example operation of the microphone unit in the system offigure 1 . -
Figure 3 Shows an example implementation of an audio device to implement a microphone unit in the system offigure 1 . -
Figure 4 illustrates an example method of capturing the navigation instruction implemented by the vehicle function controller of the system offigure 1 . -
Figure 5 illustrates an example method of navigation action instruction implemented by the vehicle function controller of the system offigure 1 . -
Figures 6A illustrates different phases of operation of one or more embodiments andfigure 6B illustrate an example method of aborting an instruction implemented by the vehicle function controller of the system offigure 1 dependent on the phases of operation shown infigure 6A . -
Figure 7A illustrates an example method of a stop instruction implemented by the vehicle function controller of the system offigure 1 . -
Figure 7B illustrates an example method of a parking instruction implemented by the vehicle function controller of the system offigure 1 . -
Figure 7C illustrates an example method of rerouting implemented by the vehicle function controller of the system offigure 1 . -
Figure 8 illustrates an a system for in-vehicle voice control of an autonomous vehicle according to an embodiment. -
Figure 1 illustrates asystem 100 for in-vehicle voice control of an autonomous vehicle according to an example embodiment. Thesystem 100 includes anaudio input device 110 coupled to amicrophone 102 that captures a voice command which may be a vehicle-navigation command and provides it to the vehicle electronics implementing avehicle function controller 120. Theaudio input device 110 and microphone 102 may be referred to herein as a microphone unit. Theaudio input device 110 may include avoice command detector 104, a voice command processor andinterpreter 106 and aninstruction broadcast module 108. - The
vehicle function controller 120 may be connected to one ormore actuators 140 such as asteering actuator 142braking actuator 144 andspeed actuator 146 to control the navigation of the autonomous vehicle. Thevehicle function controller 120 may include an instruction processing andinterpretation module 122, anauthorization module 124 and a feasibility check andexecution module 126. Thevehicle function controller 120 may also include a vehicle control transceiver (not shown) which may be a wired or wireless transceiver for connecting theaudio input device 110. - The
vehicle function controller 120 may also be connected to anavigation module 150. Thenavigation module 150 may provide Global Positioning System (GPS) location information and other environment context information such as parking locations, roadworks, and other information used for the navigation of the autonomous vehicle. - The
vehicle function controller 120 may connect via a wired or wireless vehicle control transceiver (not shown) to anaudio output device 130 including anaudio processor 132. Theaudio output device 130 may be connected to aspeaker 134 to provide acoustic feedback to the passengers of the autonomous vehicle. Theaudio output device 130 and thespeaker 134 may be referred to herein as a speaker unit. The speaker unit may be incorporated in the vehicle or be part of a mobile hearable device. In other examples thevehicle function controller 120 may be connected to a screen for visual feedback instead of or as well as anaudio output device 130. - In some examples the microphone unit and the speaker unit may be part of a mobile hearable device such as a hearing aid or other hearing device which may be wirelessly coupled to the
vehicle function controller 120 for example by transmitter which may be a near field magnetic induction (NFMI), a near field electromagnetic induction (NFEMI), Bluetooth, or Ultra-Wide-Band (UWB) transmitter. NFMI refers to any communication means that makes use of transmissions within a localized magnetic field. NFEMI is an extension of NFMI which also uses an electric antenna for transmissions as explained for example inUS 9197986B1 15164621.3 - In operation, when a vehicle occupant speaks the
audio input device 110 may receive the audio signal from themicrophone 102. Theaudio input device 110 may perform voice command detection withvoice command detector 104 followed by voice command processing and interpretation by the voice command processing andinterpreter 106 . The results of the voice command processing and interpretation may be broadcast by thebroadcast instruction module 108 to thevehicle function controller 120. In some examples, the assembled instruction may be output from theaudio input device 110 and broadcast via a wired or wireless connection to thevehicle function controller 120. - The
vehicle function controller 120 may receive and process the instruction with instruction processing andinterpretation module 122. Following instruction processing and interpretation a check is made by theauthorization module 124 to determine to identify whether voice commands are provided by an authorized occupant of the vehicle. In some examples, identification of an authorized person may be based on either seat location, biometric data, from voice recognition, visual information. The feasibility check andexecution module 126 may output an audio signal to theaudio output device 130 if the command is not authorised or it is not feasible to execute the command for some other reason. If the command is authorised and feasible, the command may be executed to control theactuators 140 in order to navigate the autonomous vehicle based on information from thenavigation module 150 and the authorized command.Audio input device 110,vehicle function controller 120 andaudio output device 130 may be implemented in hardware or a combination of hardware and software. -
Figure 2 describes amethod 200 of operating theaudio input device 110. Instep 202 the autonomous vehicle may be switched on. Instep 204 an optional pairing action may occur between theaudio input device 110 and thevehicle function controller 120. This may be done for examples of the audio input device which are wirelessly coupled to thevehicle function controller 120 using for example a Bluetooth, UWB, NFMI or NFEMI wireless connection. In examples when theaudio input device 110 is fixed in the vehicle interior or connected by a wired connection, the pairing may not be required. Instep 206 theaudio input device 110 may be in a listening mode. Instep 208, the audio input device may determine whether a voice has been detected. If a voice has not been detected, the method may return to step 206. If a voice has been detected, the method may proceed to step 210 and extract keywords from the speech. Keyword extraction may be implemented for example using temporal convolution networks as described in "Temporal Convolutional Networks for Action Segmentation and Detection", Colin Lea et al, 16th November 2016, arXiv:1611.05267. An extracted keyword may be compared against known vehicle-navigation commands instep 212. If a known command has not been detected, then the method returns to step 206. If a known command has been detected then the method proceeds to step 214 in which a command instruction may be assembled including the command keyword and additional authentication information such as for example the seat location of the occupant. Instep 216 the assembled command may be transmitted to the vehicle function controller. Followingstep 216, the method may return to listeningmode step 206. -
Figure 3 Shows an example implementation of amicrophone unit 300 which may implement themethod 200.Microphone unit 300 includes first andsecond microphone 302, 302' connected to arespective microphone interface 304. 304'. The microphone interface outputs may be connected to adigital signal processor 306. The output of thedigital signal processor 306 may be connected to a wired orwireless transceiver 308 for communication with for examplevehicle function controller 120. The microphone unit may have two microphones which for example may allow beamforming which may for example be implemented by software executed on thedigital signal processor 306. Using beamforming may improve the discrimination between the intended user and background noise or speech from other occupants of the vehicle. -
Figure 4 shows an example method ofoperation 400 which may be implemented by thevehicle function controller 120. Instep 402 an instruction from the microphone unit may be received by the vehicle function controller. Instep 404 the instruction may be processed to check whether it includes a valid command The extracted keyword may be compared with the database keyword instruction set to make a comparison if the keyword is valid. A list of valid commands may be provide indatabase 408. Step 404 may be implemented for example by the instruction processing andinterpretation module 122. Instep 406 an authorization check is made to determine whether a command is provided by an authorized person. - An authorized person may be a person that wears a paired hearable/wearable or a person that is located on a given seat. Authorised person and location information may be stored in
database 408. In some examples, authorization is done by pairing so that only voice commands that are provided from a paired device are authorized for execution. In other examples authorization may be based on Seating location. In some examples, navigation may be authorized from a given seat within the vehicle. In other examples, navigation may only be permitted from a particular seat location. Alternatively, navigation could be handed-over to be done from any other seat within the vehicle. - In a navigation system using hearable devices, the detection of the seat location may be implemented for example as described in European patent application
19187261.3 - In some examples, in addition to audio detection, the authentication may include face recognition. Once a voice command has been provided together with a seat location, a check is done via a camera system whether the occupant of the vehicle that provided the voice command is authorized to navigate the vehicle. In other examples a centralized microphone system may be used that includes a direction of where the voice command came from.
- In a centralized microphone system, a microphone may be paired with more than one microphone. For example, there could be a microphone on each side (left/right) of a centre microphone. From the relative signal strengths and time delay of the audio signal detected by each microphone, the direction of the audio source may be determined.
- Following the authorization check in
step 406,the method proceeds to step 410 which checks if the command is valid. The voice command is classified as invalid when it has been provided by a non-authorized person, or when it is not a valid navigation instruction for the vehicle. If the command is invalid, the method proceeds to step 412 and an audio signal is generated for example an alert sound or a generated speech message to be provided to the vehicle occupants. This audio signal is then output on a speaker instep 414 which may be implemented using theaudio output device 130 andspeaker 134. Returning to step 410 if the command is valid and provided by an authorized person, the navigation instruction may be captured instep 416 and stored for execution instep 418. Authorization and validity checks carried out insteps authorization module 124. -
Figure 5 shows an example method ofoperation 500 which may be implemented by thevehicle function controller 120.Method 500 indicates the operation for a navigation action process which may be implemented for example by feasibility check andexecution module 126. In the example shown, the commands are Turn Left, Turn Right, Stop, and Park but it will be appreciated that other commands may also be similarly processed. The navigation instruction for execution which may be determined bymethod 400 is captured instep 514 and processed instep 516. If the command is to "turn left" the method proceeds to step 518 and in step 526 a determination is made whether a route exists that allows the vehicle to comply with the instruction. Thisdetermination step 526 may require environmental information to provide the information about the context in which the vehicle is operating. This can be traffic light related information, or safety related aspects like the presence of other vehicles, pedestrians, cyclists etc This environmental information may be provided via aV2X connection 502,V2X transceiver 504 andV2X processor 506. Alternatively or in addition the environmental information may be provided bynavigation maps 512. - As illustrated, the information from
V2X processor 506 is provided to thedecision step 526 schematically by theconnection 508. Map information for route details and path planning may be provided byconnection 510. If the left route exists the method proceeds to anexecution step 534. For the left turn instruction instep 536 the actuators of the autonomous vehicle may be controlled to re-route. If the left route does not exist, for example due to a closed street, safety related aspects such as no turn or stop permitted in the current location, the method proceeds to step 544 and an indication is provided to the used that the command cannot be serviced. This indication may be a computer generated speech or other audio indication. - Similarly if the command is to "turn right" the method proceeds to step 520 and in step 528 a determination is made whether a route exists that allows the vehicle to comply with the instruction. If the route exists the method proceeds to an
execution step 534. For the right turn instruction instep 538 the actuators of the autonomous vehicle may be controlled to re-route. If the left route does not exist, the method proceeds to step 544 as previously described. - If the command is to "stop" the method proceeds to step 522 and in step 530 a determination is made whether a stopping zone exists that allows the vehicle to comply with the instruction. If the stopping zone does not exist the method proceeds to an
execution step 534. Instep 540 actuators in the autonomous vehicle may be controlled to execute the stop instruction. If it is not possible to execute the stop instruction, the method proceeds to step 544 as previously described. - If the command is to "park" the method proceeds to step 524 and in step 532 a determination is made whether a stopping zone exists that allows the vehicle to comply with the instruction. If the stopping zone does not exist the method proceeds to an
execution step 534 In step 54 actuators in the autonomous vehicle may be controlled to execute the park instruction. If it is not possible to execute the stop instruction, the method proceeds to step 544 as previously described. -
Figures 6A and 6B illustrate an example operation of an "abort" command.Figure 6A shows the possible phases ofoperation 600. In a first phase orT1 phase 602, the voice command capturing and processing is in progress. In asecond phase 604 or T2 phase, the execution is in progress i.e. the action is being performed. -
Figure 6B shows anexample abort process 610. Instep 612 the captured navigation abort command is received, In step 614 a check is made to determine whether the received abort command is in theT1 phase 602. If the abort command is in theT1 phase 602 then instep 616, the previous instruction command is cancelled and instep 618 no control instructions are issued by the vehicle function controller. Instep 620 an indication is generated that the abort operation was successful. Returning to step 614 if the abort command is in theT2 phase 604, then the actions of the current instruction continue to be executed instep 622 and instep 624 the user receives feedback that the abort operation was not done and the previous instruction has been executed. -
Figures 7A, 7B and7C illustrate a further example detail of the execution steps 534 shown infigure 5 . -
Figure 7A shows amethod 700 of executing a stop instruction which may be used to implementexecution step 540 inmethod 500. Instep 702 the method may control the actuators to reduce speed and initiate a lane change. In step 704 a check is made to determine whether the vehicle is on a high speed lane. If the vehicle is not on a high speed lane then the method proceeds to step 708 and stop light or hazard indicators may be activated. Returning to step 704, if the vehicle is on a high speed lane then the method proceeds to step 706 and the vehicle actuators are controlled to move the autonomous vehicle to a low speed lane and then to step 708. Followingstep 708, the actuators may be controlled to move the vehicle to a stop zone instep 710 and then instep 712 the brakes are actuated to halt the vehicle. Instep 714 the completion of the command may be indicated by generating an audio output. -
Figure 7B shows a method of executing apark instruction 720 which may be used to implementexecution step 542 inmethod 500. Instep 722 the method may control the actuators to reduce speed and initiate a lane change. In step 724 a check is made to determine whether the vehicle is on a high speed lane. If the vehicle is not on a high speed lane then the method proceeds to step 728 and stop light or hazard indicators may be activated. Returning to step 724, if the vehicle is on a high speed lane then the method proceeds to step 726 and the vehicle actuators are controlled to move the autonomous vehicle to a low speed lane and then to step 728. Followingstep 728, the actuators may be controlled to move the vehicle to a stop zone instep 730 and then instep 732 the brakes are actuated to halt the vehicle. Instep 734 the completion of the command may be indicated by generating an audio output. -
Figure 7C shows a method of executing are-route instruction 740 which may be used to implementsteps step 742 navigation controls may be adapted depending on the command. Instep 744 the successful service of the command may be indicated. -
Figure 8 illustrates asystem 800 for in-vehicle voice control of an autonomous vehicle according to an example embodiment. Thesystem 800 may implement one or more of themethods system 800 includes anaudio input device 810 coupled to amicrophone 802 that captures a voice command which may be a vehicle-navigation command and provides it to the vehicle electronics implementing avehicle function controller 820. Theaudio input device 810 andmicrophone 802 may be referred to herein as a microphone unit. Theaudio input device 810 may include avoice command detector 804, a voice command processor andinterpreter 806 aninstruction assembler 812 and aninstruction broadcast module 808. - The
vehicle function controller 820 may be connected to one ormore actuators 840 such as asteering actuator 842braking actuator 844 andspeed actuator 846 to control the navigation of the autonomous vehicle. Thevehicle function controller 820 may include an instruction processing andinterpretation module 822, anauthorization module 824 and a feasibility check andexecution module 826. Thevehicle function controller 820 may also include a vehicle control transceiver (not shown) which may be a wired or wireless transceiver for connecting theaudio input device 810. - The
vehicle function controller 820 may also be connected to anavigation module 850. Thenavigation module 850 may provide GPS location information and other environment context information such as parking locations, roadworks, and other information used for the navigation of the autonomous vehicle. - The
vehicle function controller 820 may have an output connected by a wired or wireless connection to anaudio output device 830 including anaudio processor 832. Theaudio output device 830 may be connected to aspeaker 834 to provide acoustic feedback to the passengers of the autonomous vehicle. Theaudio output device 830 and the speaker 934 may be referred to herein as a speaker unit. The speaker unit may be incorporated in the vehicle or be part of a mobile hearable device. - In some examples the microphone unit and the speaker unit may be part of a mobile hearable device such as a hearing aid or other hearing device which may be wirelessly coupled to the
vehicle function controller 820 for example by transmitter which may be a near field magnetic induction (NFMI) transmitter or a near field electromagnetic induction transmitter (NFEMI), Bluetooth and Ultra-Wide-Band (UWB) communication. - In operation, when a vehicle occupant speaks the
audio input device 810 may receive the audio signal from themicrophone 802. Theaudio input device 810 may perform voice command detection with voice command detector followed by voice command processing and interpretation. The instruction assemblemodule 812 may receive a seat identifier for example from a seat identification transceiver which may be wirelessly coupled to theaudio input device 810. The instruction assemblemodule 812 may receive biomarker data from thebiomarker module 852 . In some examples thebiomarker module 852 may be included in the vehicle. Alternatively or in addition, in other examples thebiomarker module 852 may be in theaudio input device 810. The results of the voice command processing and interpretation instruction may be broadcast by thebroadcast instruction module 808 to thevehicle function controller 820. In some examples, the assembled instruction may be output from theaudio input device 810 and broadcast via a wired or wireless connection to thevehicle function controller 820. - The
vehicle function controller 820 may receive and process the instruction with instruction processing andinterpretation module 822. Following instruction processing and interpretation a check is made by theauthorization module 824 to determine to identify whether voice commands are provided by an authorized occupant of the vehicle. In some examples, identification of an authorized person may be based on either seat location, biometric data, from voice recognition, visual information. The feasibility check andexecution module 826 may output an audio signal to theaudio output device 830 if the command is not authorised or it is not feasible to execute the command for some other reason such. If the command is authorised and feasible, the command may be executed to control theactuators 840 in order to navigate the autonomous vehicle based on information from thenavigation module 850 and the authorized command.Audio input device 810,vehicle function controller 820 andaudio output device 830 may be implemented in hardware or a combination of hardware and software. - Example embodiments described provide an intelligent voice-based human-machine interface for vehicles that enable control the navigation functionality of the vehicle from personalized zones of passengers inside vehicles. Any occupant of the vehicle is enabled to interact with the vehicle by means of voice commands, for navigating the vehicle to destination from their personal seating zone. The control actuators include, but are not limited to, Speed Control Unit, Brake Control Unit, Steering Control Unit, Start / Stop. Embodiments described herein may allow the direct connection between an occupant of the vehicle providing the voice command comprising at least a navigation command to a Navigation control unit of Vehicle electronics, the authorization of an occupant of the vehicle to allow execution of the given navigation instruction. A voice command comprising at least a navigation command may be validated by Vehicle electronics for the allowable Navigation control commands. Navigation commands are checked for feasibility of execution in conjunction with at least predefined navigation maps and V2X information. Adapting the navigation controls like steering, speed etc. in conjunction with at least V2X and navigation maps for path planning and rerouting to the new destination in Realtime. The command may be aborted before the navigation controls execute the instruction and provide the feedback to the occupants of the vehicle. One or more embodiments may enable lane change assist, indicators etc., for path rerouting, stop or park command. One or more embodiments may enable navigating to stop zone, parking zone based on V2X information and initiating the stop or parking sequence. The audio control system as described herein may be used with autonomous vehicles configured for example as a robot taxi.
- A method and apparatus for controlling an autonomous vehicle is described. An audio signal is detected and processed the to determine whether the speech comprises a vehicle-navigation command. The audio signal is analyzed to determine whether the signal includes speech from an authorised occupant of the vehicle. If the speech includes an authorised vehicle-navigation command which is feasible to execute given the current position of the vehicle, the command is executed to control the direction and speed of the autonomous vehicle.
- In some example embodiments the set of instructions/method steps described above are implemented as functional and software instructions embodied as a set of executable instructions which are effected on a computer or machine which is programmed with and controlled by said executable instructions. Such instructions are loaded for execution on a processor (such as one or more CPUs). The term processor includes microprocessors, microcontrollers, processor modules or subsystems (including one or more microprocessors or microcontrollers), or other control or computing devices. A processor can refer to a single component or to plural components.
- In other examples, the set of instructions/methods illustrated herein and data and instructions associated therewith are stored in respective storage devices, which are implemented as one or more non-transient machine or computer-readable or computer-usable storage media or mediums. Such computer-readable or computer usable storage medium or media is (are) considered to be part of an article (or article of manufacture). An article or article of manufacture can refer to any manufactured single component or multiple components. The non-transient machine or computer usable media or mediums as defined herein excludes signals, but such media or mediums may be capable of receiving and processing information from signals and/or other transient mediums.
- Example embodiments of the material discussed in this specification can be implemented in whole or in part through network, computer, or data based devices and/or services. These may include cloud, internet, intranet, mobile, desktop, processor, look-up table, microcontroller, consumer equipment, infrastructure, or other enabling devices and services. As may be used herein and in the claims, the following non-exclusive definitions are provided.
- In one example, one or more instructions or steps discussed herein are automated. The terms automated or automatically (and like variations thereof) mean controlled operation of an apparatus, system, and/or process using computers and/or mechanical/electrical devices without the necessity of human intervention, observation, effort and/or decision.
- Although the appended claims are directed to particular combinations of features, it should be understood that the scope of the disclosure of the present invention also includes any novel feature or any novel combination of features disclosed herein either explicitly or implicitly or any generalisation thereof, whether or not it relates to the same invention as presently claimed in any claim and whether or not it mitigates any or all of the same technical problems as does the present invention.
- Features which are described in the context of separate embodiments may also be provided in combination in a single embodiment. Conversely, various features which are, for brevity, described in the context of a single embodiment, may also be provided separately or in any suitable sub combination.
- The applicant hereby gives notice that new claims may be formulated to such features and/or combinations of such features during the prosecution of the present application or of any further application derived therefrom.
- For the sake of completeness it is also stated that the term "comprising" does not exclude other elements or steps, the term "a" or "an" does not exclude a plurality, a single processor or other unit may fulfil the functions of several means recited in the claims and reference signs in the claims shall not be construed as limiting the scope of the claims.
Claims (15)
- A method for controlling an autonomous vehicle, the method comprising:detecting an audio signal;processing the audio signal to determine whether the speech comprises a vehicle-navigation command;determining whether the audio signal comprises speech from an authorised occupant of the vehicle;in response to the speech comprising an authorised vehicle-navigation command, determining whether execution of the vehicle-navigation command is feasible andexecuting the vehicle-navigation command if feasible to control at least one vehicle actuator to control the direction and speed of the autonomous vehicle.
- The method of claim 1 further comprising providing feedback to the authorized occupant if execution of the vehicle navigation command is not feasible.
- The method of any preceding claim wherein determining whether the audio signal comprises speech of an authorised occupant of the vehicle further comprises determining the seat location of the occupant.
- The method of any preceding claim further comprising in response to determining that the execution of the vehicle is not feasible, providing an indication to the user that the action is prohibited.
- The method of any preceding claim further comprising comparing the voice signature of the received speech with a predetermined speech signature database and determining if an occupant is an authorised occupant from the comparison.
- The method of any preceding claim wherein determining whether execution of the vehicle-navigation command is feasible comprises:receiving navigation data via at least one of an in-vehicle navigation system and vehicle-to-x (V2X) network; andobtaining environmental context attributes from the received navigation data.
- The method of any preceding claim further comprising:receiving an abort instruction for a previously issued command;determining whether the previously issued command is in an execution phase;cancelling the previously issued command if the previously issued command is not in an execution phase.
- An audio device comprising: a microphone; a transceiver; a processor coupled to the transceiver and the microphone and configured to:process the audio signal to determine whether the speech comprises a vehicle-navigation command;transmit a control instruction comprising the vehicle-navigation command to a vehicle control system coupled to the audio device, the vehicle control system being configured to determine whether the vehicle-navigation command is authorised; andin response to determining that the vehicle-navigation command is authorized , control at least one actuator in the vehicle dependent on the authorized vehicle-navigation command.
- The audio device of claim 8 further configured to: receive a location identifier via the transceiver from a location identification transmitter located in a vehicle; and wherein the control instruction further comprises the received location identifier and wherein the vehicle control system is further configured to determine whether the audio signal comprises speech from an authorised occupant of the vehicle from the location of the occupant provided by the location identification transmitter.
- The audio device of claim 8 or 9 configured as a mobile hearable device and comprising a wireless transceiver configured as one of a near field magnetic induction, NFMI, transceiver, a near field electromagnetic induction, NFEMI, transceiver and an Ultra-Wide-Band, UWB, transceiver.
- A vehicle control apparatus for an autonomous vehicle comprising:a vehicle function controller configured to be coupled to one or more vehicle actuators and comprising a vehicle control transceiver configured to be coupled to an audio device;wherein the vehicle control transceiver is configured to receive a vehicle-navigation command from the audio device and is further configured to determine whether the vehicle-navigation command is authorized and to control the one or more vehicle actuators to control the direction and speed of the autonomous vehicle dependent on the authorized vehicle-navigation command.
- The vehicle control apparatus of claim 11 further comprising at least one location identification transmitter located in the autonomous vehicle wherein each of the at least one location identification transmitters are configured to transmit a respective location identifier for authorizing a vehicle-navigation command to the audio device.
- The vehicle control apparatus of claim 12 wherein the at least one location identification transmitter comprises a wireless transceiver configured as one of a near field magnetic induction, NFMI, transceiver, a near field electromagnetic induction, NFEMI, transceiver and an Ultra-Wide-Band, UWB, transceiver.
- A system for controlling an autonomous vehicle comprising the audio input device of any of claims 8 to 10 and the vehicle control apparatus of claims 11 to 13.
- A non-transitory computer readable media comprising a computer program comprising computer executable instructions which, when executed by a processing unit, causes the processing unit to control an autonomous vehicle by performing the steps of:detecting an audio signal;processing the audio signal to determine whether the speech comprises a vehicle-navigation command;determining whether the audio signal comprises speech from an authorised occupant of a vehicle;in response to the speech comprising an authorised vehicle-navigation command, determining whether execution of the vehicle-navigation command is feasible andexecuting the vehicle-navigation command if feasible to control at least one vehicle actuator to control the direction and speed of the autonomous vehicle.
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
IN202011031938 | 2020-07-25 |
Publications (1)
Publication Number | Publication Date |
---|---|
EP3944232A1 true EP3944232A1 (en) | 2022-01-26 |
Family
ID=73059440
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
EP20205245.2A Withdrawn EP3944232A1 (en) | 2020-07-25 | 2020-11-02 | Voice control for autonomous vehicles |
Country Status (1)
Country | Link |
---|---|
EP (1) | EP3944232A1 (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN115547313A (en) * | 2022-09-20 | 2022-12-30 | 海南大学 | Method for controlling sudden stop of running vehicle based on voice of driver |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20140365228A1 (en) * | 2013-03-15 | 2014-12-11 | Honda Motor Co., Ltd. | Interpretation of ambiguous vehicle instructions |
US9197986B1 (en) | 2014-06-12 | 2015-11-24 | Nxp, B.V. | Electromagnetic induction radio |
US20170349184A1 (en) * | 2016-06-06 | 2017-12-07 | GM Global Technology Operations LLC | Speech-based group interactions in autonomous vehicles |
US20180194366A1 (en) * | 2017-01-10 | 2018-07-12 | Ford Global Technologies, Llc | Autonomous-Vehicle-Control System And Method Incorporating Occupant Preferences |
US20180201226A1 (en) * | 2017-01-17 | 2018-07-19 | NextEv USA, Inc. | Voice Biometric Pre-Purchase Enrollment for Autonomous Vehicles |
-
2020
- 2020-11-02 EP EP20205245.2A patent/EP3944232A1/en not_active Withdrawn
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20140365228A1 (en) * | 2013-03-15 | 2014-12-11 | Honda Motor Co., Ltd. | Interpretation of ambiguous vehicle instructions |
US9197986B1 (en) | 2014-06-12 | 2015-11-24 | Nxp, B.V. | Electromagnetic induction radio |
US20170349184A1 (en) * | 2016-06-06 | 2017-12-07 | GM Global Technology Operations LLC | Speech-based group interactions in autonomous vehicles |
US20180194366A1 (en) * | 2017-01-10 | 2018-07-12 | Ford Global Technologies, Llc | Autonomous-Vehicle-Control System And Method Incorporating Occupant Preferences |
US20180201226A1 (en) * | 2017-01-17 | 2018-07-19 | NextEv USA, Inc. | Voice Biometric Pre-Purchase Enrollment for Autonomous Vehicles |
Non-Patent Citations (1)
Title |
---|
COLIN LEA ET AL., TEMPORAL CONVOLUTIONAL NETWORKS FOR ACTION SEGMENTATION AND DETECTION, 16 November 2016 (2016-11-16) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN115547313A (en) * | 2022-09-20 | 2022-12-30 | 海南大学 | Method for controlling sudden stop of running vehicle based on voice of driver |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US11756429B2 (en) | Transport vehicle configuration for impaired riders | |
US11180116B2 (en) | Vehicle dispatch management system and vehicle dispatch management server | |
CN109572702B (en) | Vehicle control device and vehicle including the same | |
US10266182B2 (en) | Autonomous-vehicle-control system and method incorporating occupant preferences | |
US11577742B2 (en) | Methods and systems for increasing autonomous vehicle safety and flexibility using voice interaction | |
CN112180921B (en) | Automatic driving algorithm training system and method | |
CN109774729A (en) | Realize the quick stopping method and apparatus of automatic driving vehicle | |
US11054818B2 (en) | Vehicle control arbitration | |
CN104737226A (en) | Speech recognition in a motor vehicle | |
JP6011584B2 (en) | Speech recognition apparatus and speech recognition system | |
CN108016383B (en) | Autonomous vehicle ingress and egress system and method | |
CN105957523A (en) | Vehicular system control method and device | |
KR20190006741A (en) | Connectivity Integration Management Method and Connected Car thereof | |
WO2014181434A1 (en) | Driving assist system | |
WO2019131116A1 (en) | Information processing device, moving device and method, and program | |
EP3944232A1 (en) | Voice control for autonomous vehicles | |
CN110366852B (en) | Information processing apparatus, information processing method, and recording medium | |
CN113557174A (en) | Information processing apparatus, information processing method, movement control apparatus, and movement control method | |
JP3979080B2 (en) | Voice recognition device | |
WO2022176038A1 (en) | Voice recognition device and voice recognition method | |
CN112238827A (en) | Mobile audible device | |
KR20190074344A (en) | Dialogue processing apparatus and dialogue processing method | |
JP2020059401A (en) | Vehicle control device, vehicle control method and program | |
CN109345855A (en) | A kind of user management method | |
WO2023090057A1 (en) | Information processing device, information processing method, and information processing program |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PUAI | Public reference made under article 153(3) epc to a published international application that has entered the european phase |
Free format text: ORIGINAL CODE: 0009012 |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: THE APPLICATION HAS BEEN PUBLISHED |
|
AK | Designated contracting states |
Kind code of ref document: A1 Designated state(s): AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: THE APPLICATION IS DEEMED TO BE WITHDRAWN |
|
18D | Application deemed to be withdrawn |
Effective date: 20220727 |