WO2016032990A1 - Integrated wearable article for interactive vehicle control system - Google Patents
Integrated wearable article for interactive vehicle control system Download PDFInfo
- Publication number
- WO2016032990A1 WO2016032990A1 PCT/US2015/046626 US2015046626W WO2016032990A1 WO 2016032990 A1 WO2016032990 A1 WO 2016032990A1 US 2015046626 W US2015046626 W US 2015046626W WO 2016032990 A1 WO2016032990 A1 WO 2016032990A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- vehicle
- user
- function
- wearable article
- input
- Prior art date
Links
- 230000002452 interceptive effect Effects 0.000 title description 43
- 238000000034 method Methods 0.000 claims abstract description 17
- 230000006870 function Effects 0.000 claims description 111
- 230000009471 action Effects 0.000 claims description 14
- 238000004891 communication Methods 0.000 claims description 5
- 239000004984 smart glass Substances 0.000 claims description 3
- 210000003128 head Anatomy 0.000 description 13
- 238000010586 diagram Methods 0.000 description 8
- 230000033001 locomotion Effects 0.000 description 8
- 230000008569 process Effects 0.000 description 8
- 238000013459 approach Methods 0.000 description 7
- 239000000446 fuel Substances 0.000 description 7
- 238000012545 processing Methods 0.000 description 7
- 230000003993 interaction Effects 0.000 description 6
- 238000003860 storage Methods 0.000 description 6
- 239000011248 coating agent Substances 0.000 description 5
- 238000000576 coating method Methods 0.000 description 5
- 230000008901 benefit Effects 0.000 description 4
- 230000001815 facial effect Effects 0.000 description 4
- 238000005516 engineering process Methods 0.000 description 3
- 241001669679 Eleotris Species 0.000 description 2
- 239000004983 Polymer Dispersed Liquid Crystal Substances 0.000 description 2
- 238000007405 data analysis Methods 0.000 description 2
- 230000002542 deteriorative effect Effects 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 238000012552 review Methods 0.000 description 2
- 238000002604 ultrasonography Methods 0.000 description 2
- 230000003044 adaptive effect Effects 0.000 description 1
- 238000004458 analytical method Methods 0.000 description 1
- 238000013475 authorization Methods 0.000 description 1
- 230000015572 biosynthetic process Effects 0.000 description 1
- 230000001149 cognitive effect Effects 0.000 description 1
- 230000001419 dependent effect Effects 0.000 description 1
- 238000001514 detection method Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 238000005108 dry cleaning Methods 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 230000036541 health Effects 0.000 description 1
- 230000010354 integration Effects 0.000 description 1
- 238000012423 maintenance Methods 0.000 description 1
- 238000013507 mapping Methods 0.000 description 1
- 238000005259 measurement Methods 0.000 description 1
- 239000003607 modifier Substances 0.000 description 1
- 238000003825 pressing Methods 0.000 description 1
- 230000009467 reduction Effects 0.000 description 1
- 230000004044 response Effects 0.000 description 1
- 238000005096 rolling process Methods 0.000 description 1
- 230000029305 taxis Effects 0.000 description 1
- 230000001960 triggered effect Effects 0.000 description 1
- 238000012559 user support system Methods 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F21/00—Security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
- G06F21/30—Authentication, i.e. establishing the identity or authorisation of security principals
- G06F21/31—User authentication
- G06F21/32—User authentication using biometric data, e.g. fingerprints, iris scans or voiceprints
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04W—WIRELESS COMMUNICATION NETWORKS
- H04W4/00—Services specially adapted for wireless communication networks; Facilities therefor
- H04W4/30—Services specially adapted for particular environments, situations or purposes
- H04W4/40—Services specially adapted for particular environments, situations or purposes for vehicles, e.g. vehicle-to-pedestrians [V2P]
- H04W4/48—Services specially adapted for particular environments, situations or purposes for vehicles, e.g. vehicle-to-pedestrians [V2P] for in-vehicle communication
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60K—ARRANGEMENT OR MOUNTING OF PROPULSION UNITS OR OF TRANSMISSIONS IN VEHICLES; ARRANGEMENT OR MOUNTING OF PLURAL DIVERSE PRIME-MOVERS IN VEHICLES; AUXILIARY DRIVES FOR VEHICLES; INSTRUMENTATION OR DASHBOARDS FOR VEHICLES; ARRANGEMENTS IN CONNECTION WITH COOLING, AIR INTAKE, GAS EXHAUST OR FUEL SUPPLY OF PROPULSION UNITS IN VEHICLES
- B60K35/00—Instruments specially adapted for vehicles; Arrangement of instruments in or on vehicles
- B60K35/10—Input arrangements, i.e. from user to vehicle, associated with vehicle functions or specially adapted therefor
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F1/00—Details not covered by groups G06F3/00 - G06F13/00 and G06F21/00
- G06F1/16—Constructional details or arrangements
- G06F1/1613—Constructional details or arrangements for portable computers
- G06F1/163—Wearable computers, e.g. on a belt
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F21/00—Security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
- G06F21/30—Authentication, i.e. establishing the identity or authorisation of security principals
- G06F21/31—User authentication
- G06F21/34—User authentication involving the use of external additional devices, e.g. dongles or smart cards
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F21/00—Security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
- G06F21/30—Authentication, i.e. establishing the identity or authorisation of security principals
- G06F21/31—User authentication
- G06F21/36—User authentication by graphic or iconic representation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/017—Gesture based interaction, e.g. based on a set of recognized hand gestures
Definitions
- the present disclosure relates to a vehicle and more particularly to systems and methods therefor.
- a key fob allows a driver to perform remote control functions such as locking or starting a vehicle.
- the driver needs to carry the key fob around, such as in a key chain, a pocket, or a purse.
- wearable technology has made it possible for people to interact with the vehicle through a wearable article, such as a smart watch or wristband.
- a system can include a user input subsystem and a user recognition and authentication subsystem in communication with the user input subsystem.
- the user input subsystem includes a wearable article and is configured to receive inputs from a user.
- the user recognition and authentication subsystem is configured to detect and authenticate the user based on the inputs received from the wearable article, or the vehicle, or both.
- the inputs received from the wearable article can include, for example, a user input indicating a vehicle function to be performed, a gesture input by the user for authentication, or both.
- the inputs received from the vehicle can include, for example, a gesture input by the user for authentication.
- the wearable article can include a wearable computing device configured to perform at least one vehicle function on a vehicle.
- the wearable article can be, for example, a smart watch, a smart clothing article, a transdermal chip or a wearable sensor.
- a driver score can be generated for driver actions associated with vehicle functions performed by the user on the vehicle. The driver score can be transmitted to the wearable article, and can be uploaded, for example, to a home computer or an external database via cloud.
- a method for operating a vehicle can include receiving, at a wearable article, a first input from a user indicating a vehicle function to be performed on a vehicle, receiving a second input indicating a gesture by the user for authentication, and generating, at the wearable article, a control signal for performing the vehicle function on the vehicle based on a successful authentication of the user.
- the wearable article can be, for example, a smart watch, a smart clothing article, a transdermal chip or a wearable sensor.
- a driver score can be generated for driver actions associated with vehicle functions performed by the user on the vehicle. The driver score can be transmitted to the wearable article, and can be uploaded, for example, to a home computer or an external database via cloud.
- a wearable article can include one or more processors and a memory.
- the memory stores data and program instructions that can be executed by the processors.
- the wearable article can be a wearable computing device, such as a smart watch or a wearable sensor.
- the processors are configured to execute instructions stored in the memory.
- the instructions include receiving a first input from a user indicating a vehicle function to be performed on a vehicle, receiving a second input indicating a gesture by the user for authentication, and generating a control signal to perform the vehicle function based on a successful authentication of the user.
- the wearable article can be, for example, a smart watch, a smart clothing article, a transdermal chip or a wearable sensor.
- a driver score can be generated for driver actions associated with vehicle functions performed by the user on the vehicle. The driver score can be transmitted to the wearable article, and can be uploaded, for example, to a home computer or an external database via cloud.
- Figure 1 is a pictorial representation of an example vehicle for use with an interactive vehicle window display system
- Figure 2 is a schematic block diagram of the interactive vehicle window display system according to one non-limiting embodiment
- Figure 3 is a partial interior view of the vehicle with the interactive vehicle window display system
- Figure 4 is a top view of the vehicle illustrating an exterior user identification subsystem of the interactive vehicle window display system
- Figure 5 is a pictorial representation of the vehicle illustrating user identification via a wearable article, skeletal joint relationship, key fob and/or user gesture;
- Figure 6 is a schematic block diagram of an algorithm for operation of the system according to one non-limiting embodiment
- Figure 7 is a pictorial representation of an example skeletal joint relationship recognizable by the system
- Figure 8 is an illustration of an example wearable article and an example user gesture recognizable by the system according to one non-limiting embodiment
- Figure 9 is an example landing page displayed by the interactive vehicle window display system
- Figure 10 is an example route page displayed by the interactive vehicle window display system
- Figure 11 is an example calendar page displayed by the interactive vehicle window display system
- Figure 12 is an example weather page displayed by the interactive vehicle window display system
- Figure 13 is an example vehicle status page displayed by the interactive vehicle window display system
- Figure 14 is an example to-do page displayed by the interactive vehicle window display system
- Figure 15 is a partial interior view of a vehicle cabin illustrating an interactive environment for the driver and/or passengers to utilize functionalities of a vehicle head unit;
- Figure 16 is a partial interior view of the vehicle cabin illustrating discrimination of a driver and/or passenger to selectively permit utilization of functionalities of a vehicle head unit during vehicle operation;
- Figure 17 is a pictorial representation of a vehicle passenger facial map for use with the system to track occupant location;
- Figure 18 is an overhead interior view of the vehicle illustrating a sensor arrangement to track occupant location within the vehicle cabin;
- Figure 19 is a schematic block diagram of a wearable article according to one non-limiting embodiment;
- Figure 20A is an example wearable article
- Figure 20B is another example wearable article
- Figure 20C is another example wearable article
- Figure 21 is a flow diagram of process performed by a wearable article according to one non-limiting embodiment
- Figures 22A-C are example screen displays of an example wearable article according to one implementation
- Figures 23A-F are example screens of the remote control mode of the example wearable article in Figures 22A-C;
- Figure 24A is an example screen that displays a notification of a vehicle function performed on a vehicle
- Figure 24B is an example driver score alert screen for the example wearable article in Figures 22A-22C.
- Figures 25A-25C are examples screens of the panic mode for the example wearable article in Figures 22A-22C.
- a wearable article can be integrated with a key fob for operating an automobile so that the user will no longer need to carry the key fob separately.
- a user can put on the smart watch every morning without struggling to find the key. The user will no longer need to carry the key fob separately.
- the smart watch can be worn by the user to perform vehicle functions on the vehicle.
- the wearable article can be implemented to receive user indications of the vehicle functions to be performed on the vehicle and gesture inputs for user authentication in order for such vehicle functions to be performed.
- a control signal to perform the vehicle functions can be generated based on a successful authentication.
- the gesture input can be used in the multi-factor authentication in conjunction with the user input received from the wearable article to authenticate the user.
- the second gesture input can be detected by a sensor, which can be coupled to the vehicle or the wearable article. The sensor can be used to detect whether the user is wearing the wearable article. When the user is not wearing the wearable article, access to some or all vehicle functions can be denied.
- the wearable article can be used to sense information from a user, such as biometrics and driving information.
- a driver score can be generated for driver actions performed on the vehicle.
- the driver score can be used to help the driver improving driving skills and alert the driver when deteriorating performance occurs.
- the driver score can be an option selectable from the menu screen of the wearable article and calculated based on sensed information of driver actions.
- the driver score can be sent to the wearable article for display, storage or further processing.
- the driver score received at the wearable article can be uploaded to, for example, a personal computer, a smartphone, or an external database via the cloud.
- Data analysis can be used to improve safe driving skills or for other application- specific purposes.
- the driver scores can be placed into a social ranking system and the user can review and evaluate his driving performances relative to others in the social ranking system.
- FIG. 1 schematically illustrates a vehicle 20 with a window 22 and an interactive vehicle window display system 30.
- the window 22 is here shown as a driver's side passenger window of a minivan type vehicle in the disclosed, non-limiting embodiment, it should be appreciated that various vehicle types and windows will also benefit herefrom.
- the system 30 generally includes an interactive display subsystem 32, a control subsystem 34, a user input subsystem 36, a user identification subsystem 38, and a user location subsystem 39.
- the user identification subsystem 38 can be implemented as a user recognition and authentication subsystem. It should be appreciated that although particular subsystems are separately defined, each or any of the subsystems may be combined or segregated via hardware and/or software of the system 30. Additionally, each or any of the subsystems can be implemented using one or more computing devices including conventional central processing units or other devices capable of manipulating or processing information.
- the interactive display subsystem 32 can include any device or devices capable of displaying images on a vehicle window 22 under the control of system 30, and can be adapted for viewing from outside the vehicle, inside the vehicle, or both.
- the interactive display subsystem 32 can include a display device integral to the window 22, such as an LCD.
- Such a display can be illuminated by ambient light or by one or more light sources under the control of system 30.
- Such light sources can be mounted at any operable locations enabling light emission onto a window from inside or outside the vehicle, depending on whether the display is to be viewed by a user located outside or inside the vehicle. Examples of such mounting locations can include in the floor, in the vehicle headliner, within the vehicle door structure, or in the exterior door panel.
- the interactive display subsystem 32 can include a coating 40 and a projector 42.
- the coating 40 may be a polymer dispersed liquid crystal (PDLC) film, applied to the window 22 to provide both transparency when inactive and partial or complete opacity when active.
- PDLC polymer dispersed liquid crystal
- the window 22 treated with the coating 40 is thereby operable to display content as a projection page visible from outside and/or inside the vehicle 20 ( Figure 1).
- the projector 42 can be mounted in the floor ( Figure 3) or other locations within the vehicle 20, such as the vehicle headliner or within the vehicle door structure as well as in locations on the vehicle exterior such as in an exterior door panel.
- the illustrated shaded area extending from the projector 42 toward the window 22 schematically represents the projection of output in the form of content pages provided by the projector 42.
- the coating 40 changes from transparent to opaque so that the projector 42 may project the output onto the window 22.
- the displayed content can include personalized information or entertainment content such as videos, games, maps, navigation, vehicle diagnostics, calendar information, weather information, vehicle climate controls, vehicle entertainment controls, email, internet browsing, or any other interactive applications associated with the recognized user, whether the information originates onboard and/or off board the vehicle 20.
- personalized information or entertainment content such as videos, games, maps, navigation, vehicle diagnostics, calendar information, weather information, vehicle climate controls, vehicle entertainment controls, email, internet browsing, or any other interactive applications associated with the recognized user, whether the information originates onboard and/or off board the vehicle 20.
- the control subsystem 34 generally includes a control module 50 with a processor 52, a memory 54, and an interface 56.
- the processor 52 may be any type of microprocessor having desired performance characteristics.
- the memory 54 may include any type of computer readable medium which stores the data and control algorithms described herein such as a user support system algorithm 58.
- the functions of the algorithm 58 are disclosed in terms of functional block diagrams ( Figure 6) and representative pages ( Figures 9-14), and it should be understood by those skilled in the art with the benefit of this disclosure that these functions may be enacted in either dedicated hardware circuitry or programmed software routines capable of execution in a microprocessor based electronics control embodiment.
- control module 50 may be a portion of a central vehicle control, a stand-alone unit, or other system such as a cloud-based system.
- Other operational software for the processor 52 may also be stored in the memory 54.
- the interface 56 facilitates communication with other subsystems such as the interactive display subsystem 32, the user input subsystem 36, the user identification subsystem 38, and the user location subsystem 39. It should be understood that the interface 56 may also communicate with other onboard vehicle systems and offboard vehicle systems.
- Onboard systems include but are not limited to, a vehicle head unit 300 which communicates with vehicle sensors that provide, for example, vehicle tire pressure, fuel level and other vehicle diagnostic information.
- Offboard vehicle systems can provide information which includes but is not limited to, weather reports, traffic, and other information which may be provided via cloud 70.
- the user input subsystem 36 can include one or more input sensors including onboard input sensors 60, offboard input devices, or both.
- Onboard input sensors 60 can include one or more motion cameras or other light sensors configured to detect gesture commands, one or more touch sensors configured to detect touch commands, one or more microphones configured to detect voice commands, or other onboard devices configured to detect user input.
- the user input subsystem can also include offboard input devices such as a wearable article 61, a key fob 62 and/or a personal electronic device 63 of the user, e.g. a tablet, smart phone, or other mobile device.
- the wearable article 61 can be a wearable computing device, such as a smart watch or a wearable sensor.
- the wearable article 61 can be integrated with the key fob 62 so that the user will no longer need to carry the key fob 62 separately.
- the wearable article 61 can be configured to receive user inputs to indicate the vehicle functions to be performed on the vehicle 20.
- the wearable article 61 can also be configured to receive gesture inputs from the user for authentication before such vehicle functions can be performed.
- the system 30 utilizes a multi-factor authentication for security and authorization.
- the authentication can be implemented, for example, in the user identification subsystem 38.
- Example multi-factor authentication may include receiving inputs from the wearable article 61, the key fob 62, skeleton joint relationship recognition (Figure 5), and/or a gesture password ( Figure 8).
- the user may be provisionally identified with one of these factors, but may require a total of at least two factors to authenticate the user prior to display of certain content. That is, the user will not be granted access to all the features in user mode 104 until a multi-factor authentication is passed and the user is within a predetermine range of the vehicle 20. This authentication process ensures the security of the vehicle and the personal information embedded in the system 30.
- the first authentication factor can be the wearable article 61, which integrates the functionalities of a digital key fob
- the second can be the skeleton joint relationship ( Figure 7) of the user. If the user does not have their wearable article 61 or the key fob 62, the skeleton joint relationship may become the first authentication factor and a gesture password such as a wave or particular arm movement ( Figure 8) becomes the second.
- the first authentication factor can be the wearable article with the integrated key fob functionalities
- the second can be a gesture input from the user, such as the gesture password or the skeleton joint relationship.
- Other combinations of authentication factors are also possible and the second factor can be optional.
- the second factor can be required when there is a heightened need for security, e.g., when the vehicle is parked in a public location or in a high crime area.
- the user can be authenticated only when it is detected that he is wearing the wearable article 61 or carrying the key fob 62.
- the wearable article 61 in one disclosed non-limiting embodiment may be encrypted to uniquely identify each user to the system 30. Additional security protocols such as a rolling time key to ensure that even the encrypted key cannot be intercepted and re -used by unauthorized devices may additionally be utilized.
- the wearable article 61 Once the wearable article 61 is recognized, the user will be welcomed and pre-authenticated to allow limited access to selected content in the user mode 104. This will provide the user with enough time to cycle through multiple content features during the walkup experience, yet maintain security with respect to other content features e.g., a destination. Once the user has been fully authenticated, all content features, e.g. destination made during the pre-authenticated state, are validated for display. If the authentication fails, the user will not be granted access to the vehicle 20 or any sensitive information.
- the system 30 in this disclosed non-limiting embodiment allows pre-authenticated access at about 30-40 feet and full access at about 15- 25 feet from the vehicle.
- Skeleton joint relationships in this disclosed non-limiting embodiment facilitate pre-authentication but not full authentication that grants full access to the vehicle 20. However, if the user has been pre- authenticated via the wearable article 61 or the key fob 62, a matching skeleton joint relationship will fully authenticate the user. That is, the user identification subsystem 38 may utilize skeleton joint relationships as the second point of identification.
- FIG 19 is a block diagram of a computing device 1000 for implementing a wearable article 61.
- the wearable article 61 can include some or all functionalities of a digital key fob such as key fob 62.
- a digital key fob such as key fob 62.
- the computing device 1000 can be any type of wearable, handheld, or other form of single computing device, or can include multiple computing devices.
- the computing device 1000 can be a smart watch 2002 ( Figure 20A), a personal mobile device, a smart clothing article 2004 ( Figure 20B), a transdermal chip (not shown), a wearable sensor (not shown), or a smartglasses article 2006 ( Figure 20C).
- the processing unit in the computing device 1000 can be a conventional central processing unit (CPU) 1102 or any other type of device, or multiple devices, capable of manipulating or processing information.
- the memory 1104 in the computing device 1000 can be a random access memory device (RAM) or any other suitable type of storage device.
- the memory 1104 can include data 1106 that is accessed by the CPU using a bus 1108.
- the memory 1104 can also include an operating system 1110 and installed applications 1112, the installed applications 1112 including programs that permit the CPU 1102 to implement the instructions to generate control signals for performing vehicle functions on a vehicle as described.
- the instructions can also include performing functions that are not related to a vehicle, such as to track a user's biometrics or to display time.
- the computing device 1100 can also include secondary, additional or external storage 1114, for example, a memory card, flash drive, or any other form of computer readable medium.
- the installed applications 1112 can be stored in a whole or in part in the external storage 1114 and loaded into the memory 1104 as needed for processing.
- the computing device 1000 can include one or more output devices, such as a display 1116 and one or more input 1118 devices, such as a keypad, a touch sensitive device, a sensor, or a gesture sensitive input device that can receive user inputs.
- the computing device 1000 can be in communication with one or more of the subsystems via a communication device (not shown) such as a transponder/transceiver device or a Wi-Fi, infrared, or Bluetooth device.
- a communication device such as a transponder/transceiver device or a Wi-Fi, infrared, or Bluetooth device.
- the computing device 1000 can communicate with the control subsystem 34 through interface 56.
- the computing device 1000 can be coupled to one or more vehicle devices configured to receive inputs from the user and provide feedback to the driver of the vehicle 20.
- the computing device 1000 can also include a sensor (not shown) to take sensed information from the user such as voice commands, ultrasound, gesture or other inputs.
- the computing device 1000 can be a wearable computing device configured to perform vehicle functions on the vehicle 20.
- vehicle functions can be implemented in the installed applications described above.
- the vehicle functions can include, among others, various remote control functions (Figures 23A-F), driver score function (Figure 24B), panic mode ( Figures 25A-C), a navigation function, an audio/video function, a climate control function, an internet access function, and a remote control function for controlling the vehicle.
- the remote control functions can include, for example, unlocking, locking (2308 in Figure 23A), switching on flashlight (2310 in Figure 23A), switching off flashlight, horn honk (2312 in Figure 23A), starting (2302 in Figure 23A), stopping (2306 in Figure 23C), powering on, or powering off of the vehicle.
- the 1112 stored in the memory 1104 can include vehicle applications such as application 2204 shown in Figure 22B.
- the applications 1112 can also include autonomous driving applications such as a data analyzer, a path planner, a target value generator, an error detector, an adaptive module, or any other application configured to implement the autonomous driving system by performing such actions as identifying the driver, planning a route for autonomous operation of the vehicle, and improving positional accuracy of the vehicle.
- FIGs 20A-C show several non-limiting examples of the wearable article 61.
- the wearable article 61 can be implemented as a smart watch 2002.
- the smart watch 2002 can be worn by the user to perform vehicle functions on the vehicle 20.
- the integrated wearable article 61 can be implemented as a smart clothing article 2004. The user can operate the smart clothing article 2004 by controlling certain vehicle functions incorporated into the smart clothing article 2004.
- the integrated wearable article 61 can also be implemented as smartglasses 2006. Other implementations of the wearable article 61 are also possible.
- the wearable article 61 can be implemented as a smart phone device, a transdermal chip, a wearable sensor, or a remote access key fob.
- Figure 21 is a flow diagram of an example process 2100 performed by the wearable computing device 1000 of Figure 19, which will be described in details further below.
- Figures 22A through 25C are example screen displays 2202 of a wearable article 2200 according to one implementation.
- Figure 22A illustrates a screen that displays date and time.
- Figure 22B illustrates a main menu from which a vehicle application icon 2204 can be selected.
- Figure 22C illustrates a menu screen with three icons: remote control 2206, driver score 2208, and panic mode 2210.
- Figures 23A-F are example screen displays when remote control 2206 is selected.
- Figure 23A illustrates a remote control screen showing a list of remote control functions to be selected by the user. The user can navigate the list up and down by scrolling or using up and down buttons. For example, when the user clicks on "Vehicle START" icon 2302, a control signal for starting the vehicle is sent to vehicle. A status message 2304 can be displayed on the control screen, as illustrated in Figure 23B. Once the start function is performed, the icon can be switched or toggled to show "VEHICLE STOP" icon 2306 as illustrated in Figure 23C.
- "Horn Honk” icon 2312 can be selected and a status message 2314 can be generated for display on the screen. Once the horn honk is performed, the user can click on "Horn Honk” icon 2316 again to repeat the action.
- Figure 24A illustrates a screen that displays an example notification
- Figure 24B illustrates a driver score alert screen 2404, which will be described in more details below.
- Figures 25A-C are example screen displays for the panic mode of the vehicle application 2204 on the wearable article 2200.
- Figure 25 A illustrates the main menu where the panic mode icon 2210 can be selected.
- Figure 25B illustrates a screen 2502 where the user is indicated to long press the select button to activate the panic mode.
- Figure 25C shows a message 2504 indicating the panic mode has been activated.
- vehicle functions to be performed on the vehicle 20 can include various remote control functions (Figures 23A-F), driver score function (Figure 24B), and panic mode ( Figures 25A-C).
- the remote control functions can include, for example, unlocking, locking (2308 in Figure 23A), switching on flashlight (2310 in Figure 23A), switching off flashlight, horn honk (2312 in Figure 23A), starting (2302 in Figure 23A), stopping (2306 in Figure 23C), powering on, or powering off of the vehicle.
- Other vehicle functions can include, for example, a navigation function, an audio/video function, a climate control function, or an internet access function.
- the wearable article can be used to sense information from a user, such as biometrics and driving information.
- the wearable article 61 can be used to store and forward a driver score.
- the driver score can be sent from the vehicle 20 to the wearable article 61.
- the wearable article 61 can upload the driver score to a remote server or cloud 70, which can perform further analysis on the driver score to help the driver improving driving skills and become a safer driver.
- data analysis based on the driver scores can be used to improve safe driving skills or for other purposes. For example, the user can review and evaluate his driving performances relative to others in a social ranking system based on the driver scores.
- a driver score can be generated for driver actions performed on the vehicle 20.
- the driver score can be used to help the driver improving driving skills and alert the driver when deteriorating performance occurs.
- the driver actions can be associated with or triggered by the vehicle functions performed on the vehicle 20.
- the driver score can be calculated based on, for example, information from a sensor such as a motion camera or light sensor that detect gesture commands, an onboard device, and/or the wearable article 61. For example, once the vehicle is started, information about driver actions can be gathered for calculating the driver score. Once the vehicle is stopped and locked, the driver action information gathered during this trip can be used to calculate the driver score. In case the driver score is calculated by a device other than the wearable article 61, it can be transmitted to the wearable article 61 for display and/or storage.
- the driver score can be an option selectable from the menu screen of the wearable article. As shown in Figure 22C, driver score option 2208 can be selected on the menu screen 2202 of the wearable article 2200. Referring to Figure 24B, an example driver score of "87" is generated and displayed on a driver score alert screen 2404 when the driver score option 2208 is selected. As previously discussed, the driver score can be uploaded and further processed for various applications-specific purposes such as to improve driving skills.
- the wearable article 61 can also be used to control multiple vehicles, or to allow multiple users to share control of a vehicle.
- encryption techniques may be implemented on the wearable articles and/or some of the vehicle subsystems for security purposes.
- At least one onboard input sensor 60 or offboard input device can be integrated into, or operate in conjunction with, the interactive display subsystem 32.
- the interactive display subsystem 32 includes an LCD display integrated into a window 22 and can operate in conjunction with one or more touch sensors integrated into the window 22, causing the window to function as a
- the interactive display subsystem 32 includes a projector 42 and coating 40 on the window 22 and can operate in conjunction with one or more motion detectors configured to detect user gesture commands, causing the window to operate as a gesture-based interactive display.
- Subsystem combinations involving the interactive display subsystem 32 and the user input subsystem and enabling user interaction with a display on a vehicle window 22 will be referred to herein as an interactive window display.
- the user identification subsystem 38 also referred to herein as the user recognition and authentication subsystem, includes one or more identification sensors 64 such as a closed-circuit television (CCTV) camera, infrared, thermal or other sensor mounted to the vehicle 20 to provide a desired field of view external to the vehicle 20 as shown in Figure 4, internal to the vehicle, or both.
- identification sensors 64 such as a closed-circuit television (CCTV) camera, infrared, thermal or other sensor mounted to the vehicle 20 to provide a desired field of view external to the vehicle 20 as shown in Figure 4, internal to the vehicle, or both.
- One example user identification subsystem 38 can recognize the driver and/or passenger based on image data captured by identification sensors 64, e.g. a skeletal joint relationship 66 and/or other user form data (Figure 5), separate from, or along with, wireless devices such as the wearable article 61 associated with that particular driver and/or passenger.
- the wearable article 61 can also include a sensor (not shown) to take sensed information from the user, such as pulse or heart rate.
- the sensor on the wearable article 61 can be, for example, a wrist- mounted sensor that recognizes the user based on voice commands, ultrasound, gesture or other inputs. Based at least in part on this identification, the system 30 provides access to interactive interfaces on the interactive display subsystem 32 associated with the particular driver and/or passenger.
- FIG 21 is a flow diagram of an example process 2100 performed by the wearable computing device 1000 of Figure 19.
- Process 2100 can be implemented as a software program executed by computing device 1000.
- the software program can include machine-readable instructions that may be stored in a memory such as memory 1104, and that, when executed by a processor, such as CPU 1102, may cause the wearable computing device 1000 to perform process 2100.
- Process 2100 can also be implemented using specialized hardware or firmware.
- a user input indicative of a vehicle function to be performed on the vehicle can be received on the wearable article, such as the wearable computing device 1000.
- the user can press on a touch button on a display of the smart watch 2002 to activate the vehicle function to unlock the front door of the vehicle 20.
- the user can select a vehicle function by pressing an icon on the smart clothing article 2004 worn by the user.
- the user input can also include, for example, an indication to activate the window display in the interactive display subsystem. Other types of inputs are also possible.
- the user can use voice commands to activate the vehicle functions.
- a gesture input by the user can be received at the wearable article such as the wearable computing device 1000.
- the gesture input can be used for authentication of the user.
- the gesture input can be used in the multi-factor authentication in conjunction with the user input received from the wearable article at step 2102 to authenticate the user.
- the user can be authenticated based on a first input received from the wearable article indicating a vehicle function to be performed and a second gesture input detected by a sensor.
- the sensor can be, for example, coupled to the vehicle 20 or to the wearable article 61.
- the sensor can also be integrated with the vehicle 20 or the wearable article 61.
- the senor can be an onboard input sensor, such as a camera or light sensor configured to detect gesture commands, or a microphone configured to detect voice commands.
- the sensor can also be an offboard input device coupled with the wearable article 61, or another device, such as key fob 62 or personal electronic device 63.
- the second input can include a gesture input by the user detected by the sensor when the user is within a predetermined range of the vehicle 20.
- the sensor can be used to detect whether the user is wearing the wearable article. When the user is not wearing the wearable article, access to some or all vehicle functions can be denied.
- a vehicle function can be performed on the vehicle 20 based on a successful authentication of the user.
- the vehicle function can be one indicated at step 2102, such as, for example, an input indication to unlock or switch on the vehicle 20.
- the vehicle functions can include various remote control functions, a driver score function, and a panic mode.
- the remote control functions can include, for example unlocking, locking, switching on flashlight, switching off flashlight, horn honk, starting, stopping, powering on, or powering off of the vehicle.
- the vehicle functions can also include a navigation function, an audio/video function, a climate control function, or an internet access function.
- Information used for user authentication can include user input received at step 2102, the gesture input received at step 2104, or any of the multi-factors described above.
- Example multi-factor authentication may include receiving inputs from the wearable article 61, the key fob 62, skeleton joint relationship recognition (Figure 5), and/or a gesture password ( Figure 8). The user may be provisionally identified with one of these factors, but may require a total of at least two factors to perform some or all of the vehicle functions.
- control signal for performing the vehicle function indicated at step 2102 can be generated and sent to the vehicle 20.
- the user can interact with the display subsystem through the integrated wearable article 61 and output can be generated for display on the vehicle window.
- the system 30 can store user profiles of known users, the user profiles including identification information relevant to individual users.
- a user profile can contain skeleton joint relationship data or facial recognition data useable by the user identification subsystem 38 to identify or authenticate a user.
- a user profile can additionally contain personal interest information, such as personal calendar and event information, driving/destination history, web browsing history, entertainment preferences, climate preferences, etc.
- any or all information contained in a user profile can be stored on or shared with the wearable article 61, a personal electronic device 63, remote server, or other cloud 70 based system.
- Such offboard storage or sharing of user profile data can facilitate utilization of user profile data in other vehicles such as any additional vehicles owned by the user, rental vehicles, etc.
- Such user profile data can be secured by being accessible through a password protected application running on the cloud 70 based system, by biometric authentication, or by other effective means.
- a user profile can additionally contain user access information; data pertaining to whether the user is allowed to control a given vehicle function.
- the user profile associated with a user can indicate full user access, or function control rights for that user. This can be analogous to the control rights of the administrator of a personal computer.
- a user profile can alternatively indicate restricted user access.
- the user profile associated with a child can be set to block the user from accessing certain audio or video controls, the navigation system, altering user profiles, or the like.
- Registration of various user profiles with the system 30 can be completed in any manner, for example, over the internet or with a direct vehicle interface.
- User profiles can be based on the identities of individual users known to or registered with the system, or to user categories, such as "unknown user", or “valet”.
- a default user category such as "unknown user” or “valet” can be associated with limited, default access, or can be associated with no access, i.e. complete prohibition of access to the system 30.
- the user location subsystem 39 operable to determine the location of one or more users inside or outside the vehicle, includes one or more location sensors 66 such as a pressure sensor, temperature sensor, or camera deployed inside or outside the vehicle.
- a device can serve as both an identification sensor 64 and a location sensor 66.
- a camera mounted within the vehicle can provide information on a user's specific identity, by means described above, and on the user' s location within the vehicle, such as the driver's seat or the front-row passenger's seat.
- elements of the interactive display subsystem 32 can also operate as location sensors 66 within the user location subsystem 39.
- pressure sensors within a smartscreen or motion detectors operating as part of an interactive display can be used to obtain user location information.
- user access can be based on user location as determined by the user location subsystem 39. For example, second or third row passengers can be allowed or disallowed access to various vehicle functions such as the navigation system.
- a user with a user profile that is associated with unlimited access per the access information associated with the user profile can specify such settings.
- user access can be based on a combination of the user profile as applied by the user identification subsystem 38, and the user location as detected by the user location subsystem 39. For example, a user with unlimited access as specified by the applied user profile can nonetheless be blocked from accessing certain vehicle functions when occupying the driver's seat of a moving vehicle.
- operation of the system 30 according to one disclosed non-limiting embodiment generally includes a sleeper mode 100, a watcher mode 102 and a user mode 104. It should be appreciated that other modes may additionally or alternatively be provided.
- the system 30 If the system 30 is active but has yet to detect a user, the system 30 will be in sleeper mode 100 until awakened by the user identification subsystem 38. After detection but prior to identification by the system 30, the watcher mode 102 may be utilized to interact with authenticated as well as un-authenticated persons. For example, when a person approaches the vehicle 20, the system 30 recognizes the direction from which the person has approached then activates the interactive display subsystem 32 to display an avatar, eyes or other graphic. The graphic may be directed specifically toward the direction from which the person approaches, e.g., the graphical eyes "look" toward their approach.
- an audio capability allows the system 30 to respond to commands and initiate interaction from a blind side of the vehicle 20, i.e., a side without the interactive display subsystem 32.
- the watcher mode 102 utilizes the user identification subsystem 38 to discriminate between authenticated and un-authenticated persons.
- the user mode 104 allows a user with a known operator and/or passenger user profile in the system 30 to make decisions on approach to the vehicle 20 so that so that certain vehicle interactions need not await entry into the vehicle 20.
- the user mode 104 reduces distractions through the reduction of travel-associated decisions from the driver's cognitive, visual and manual workload streams once within the vehicle 20.
- the user is presented with an overview of information to include, for example, weather, traffic, calendar events and vehicle health.
- predictive functions of the system 30 identify likely actions, and offer optimal paths to completion, such as planning an efficient route.
- a maximum range of content provision by the interactive display subsystem 32 may be associated with a maximum distance at which that content can be effectively interacted with by the user.
- the maximum range of each content feature is prioritized with respect to legibility range of content displayed by the interactive display subsystem 32. This range metric facilitates the determination of the order in which content appears in the walkup experience. Access to prioritized content with greater maximum range allows the walkup experience to begin further from the vehicle 20 to provide the user with more overall time to interact with the system 30.
- the interactive display subsystem can also enable the user to interact with the display subsystem through the integrated wearable article 61 and generate output for display on the vehicle window.
- "home” page 200 provides a summary of alerts and important information to the user.
- the landing page 200 provides the user with a readily reviewable overview of the status of the vehicle and how it may affect his schedule and activities.
- the content includes time information, vehicle diagnostic information, and personal calendar information.
- a low fuel warning is provided in addition to a traffic-based route update for use by the vehicle navigation system and a calendar event reminder to "Pick up Kids in 20 minutes.”
- the system 30 will include a fuel station as a stop during route guidance if the destination is a distance greater than the available fuel range.
- preferred fuel stations or other stops may be predefined in the user profile.
- the landing page 200 further displays a plurality of icons to indicate additional content pages that can be viewed by the authorized user.
- the landing page 200 itself may be accessed on each content page as an icon such as a vehicle manufacturer mark icon on each content page.
- the landing page 200 allows the authorized user to understand what vehicle systems or personal user profile items may require further attention and provides access to additional content feature details with regard to these items in the form of navigable icons that lead to additional content pages.
- the landing page 200 can additionally or alternatively integrate an interactive display, for example, a smart page or video game. Other interactive vehicle display page configurations are also possible.
- Selection of content is accomplished with, for example, wearable article 61, the key fob 62, user gestures, voice commands, touch inputs, etc.
- the user utilizes the input of the wearable article 61 to cycle through various pages displayed by the interactive display subsystem 32.
- the wearable article 61 may include a four button directional pad and two auxiliary buttons.
- hand gestures may be used to "swipe" between pages.
- the user can utilizes the key fob 62 to cycle through the pages. It should be appreciated that although particular pages are illustrated in the disclosed non-limiting embodiment, various alternative or additional pages may be provided.
- a route page 202 defaults to the predicted best route for the user with respect to an explicit or inferred next destination. Any alternate destinations or routes that can be explicit or inferred with confidence from, for example, a user personal electronic device such as the wearable article 61, are presented to permit user selection by scrolling through the options.
- the suggested route screen is here shown accessed using the folded- map icon, however, other icons may be utilized.
- a calendar page 204 displays the user's calendar.
- the view is near-term, and shows only the next 2-3 upcoming appointments. If the event includes location information the user is also given the option to use the event for destination selection.
- the calendar page 204 provides content with respect to the next appointment highlighted for the user and provides a reminder to "Pick Up Kids.”
- the calendar screen is here shown accessed using a flip calendar icon, however, other icons may be utilized.
- a weather page 206 leverages information about the route to provide relevant weather information - this may be especially effective when the user is travelling away from home. For example, the system 30 determines whether it is more valuable to present the user with local weather information, destination weather information, or both, depending on the settings selected by the user or the type of weather information available.
- the weather forecast is chronological.
- the weather page 206 can be accessed with a sun icon, however, other icons may be utilized.
- weather conditions can be utilized to generate a reminder for display on the landing screen 200 that, for example, suggests an umbrella be placed in the vehicle if rain is forecasted.
- a vehicle status page 208 provides the user with a view of impending vehicle maintenance needs that requires attention.
- Notifications can include source details of the notification, severity, and options to resolve the potential issue. For example, given the notification of "Low Fuel,” the system 30 can suggest a route to a nearby fuel station within the range of the vehicle.
- the vehicle status page 208 is here shown accessed with a vehicle icon, however, other icons may be utilized.
- a to-do list page 210 presents the authorized user with information from any associated to-do list available on, for example, that user's personal wearable device 61.
- the recognized user is tasked to "Send Package," "Submit Taxes," and "Renew Car Registration,” among other items.
- the to-do list page 210 can alternatively be integrated into the route selection page if location information is included in a given list item in the personal electronic device to-do list.
- An example of this integration includes the provision of route details to a dry cleaner if the dry cleaning pickup is on the to-do list and the current route is proximate to the location of the dry cleaner location.
- the to-do list page is here shown accessed using a check-mark icon, however, other icons may be utilized.
- information of this nature can in some variations be stored on or shared with wearable device 61, another a personal electronic device 63, remote server, or other cloud 70 based system, facilitating utilization in more than one vehicle. Any such information can be secured by being accessible through a password protected application running on the cloud 70 based system, by biometric authentication, or by other effective means.
- a first user can be granted partial or complete access to a second user's profile by password sharing, for example.
- Such sharing of access could enable a first user to write reminders or tasks from a remote location to the user profile of a second user, such as a family member, such that the reminders or tasks written by the first user will be displayed on a window when the second user approaches or enters the vehicle, or any vehicle equipped with system 30 enabled to access the user profile of the second user.
- user access to various vehicle functions can include direct or remote access to utilize functionalities of a vehicle head unit 300.
- the vehicle passengers may, for example, interact with the system 30 and thereby the vehicle head unit 300 via an interactive window display or through the wearable article 61 which communicates therewith, through Bluetooth, RFID or other wireless technology standards to exchange data.
- the system 30 may permit the formation of personal area networks (PANs) for vehicle passengers to share information.
- PANs personal area networks
- a passenger's wearable article 61 may include a mapping app operable to communicate with the vehicle navigation system on the vehicle head unit 300 with no features locked out such that the passenger can search destinations and selectively send to the vehicle navigation system via the vehicle head unit 300.
- Interaction of the system 30 with the vehicle head unit 300 also allows the driver and/or passengers to select content for other vehicle passengers and/or the driver.
- one of the passengers can select a destination to display on the navigation system for the driver while the vehicle is in motion.
- the driver can select entertainment content for display to child passengers.
- the passenger can control infotainment or climate control features controlled by the vehicle head unit 300.
- the system 30, by utilizing user location subsystem 39, is operable to track the location or position of the vehicle occupants within the vehicle cabin 400 (Figure 18) through skeletal position (Figure 16), facial map data (Figure 17), pressure sensors, interactive window display input sensors, or others.
- Figure 18 For a three row vehicle, for example, three distinct areas are tracked - front row, middle row and rear row.
- at least two sensors 402 per row are required to track a state of each occupant within the vehicle 20. In some instances, each individual seat in the vehicle 20 can be tracked.
- the data from all sensors 402 may alternatively or additionally be combined to create one central map (2D or 3D) for use by the system 30. It should be appreciated that the sensors 402 may communicate with, or be a portion of, the user identification subsystem 38, the user location subsystem 39, or both.
- the multi-point skeletal joint relationship and facial recognition map data provides a relatively accurate position of each occupant captured on an XYZ axis map that can track, to a desired level of precision, the state of each occupant at a specific snapshot in time.
- the state of each occupant facilitates further tailored operations for various vehicle functions.
- the user location subsystem 39 detects and discriminates between a driver's hand from that of a vehicle front row passenger hand to selectively unlock various head unit functionality such as navigation route selection ( Figure 16).
- Head unit functionality such as navigation route selection ( Figure 16).
- Dependent for example, on which user (driver or passenger) is attempting to access the system 30 and whether the vehicle is in motion, content menu items of the vehicle head unit 300 are selectively displayed.
- the system 30 beneficially recognizes a user with a first and second point of identification to display information for that particular, authorized user. This authentication process ensures the security of the vehicle and the personal information embedded in the system 30 yet permits vehicle interaction prior to user entry into the vehicle cabin. The system 30 also beneficially discriminates passengers from the driver to selectively permit access to personalized content or specific vehicle system interfaces.
- the vehicle 20 is not limited to an automobile, as the integrated wearable article can also be implemented with other vehicles generally controlled by a driver, or operator, such as airplanes, boats etc.
- the vehicle 20 need not be limited to one controlled by a driver or operator, and could be one or more robots or robotic tools performing operations under the control of an equivalent application to a path planner application. It is therefore to be appreciated that within the scope of the appended claims, the disclosure may be practiced other than as specifically described. For that reason the appended claims should be studied to determine true scope and content. The scope of the claims is thus to be accorded the broadest interpretation so as to encompass all such modifications and equivalent structures as is permitted under the law.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Computer Security & Cryptography (AREA)
- General Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Computer Hardware Design (AREA)
- Software Systems (AREA)
- Human Computer Interaction (AREA)
- Signal Processing (AREA)
- Computer Networks & Wireless Communication (AREA)
- Chemical & Material Sciences (AREA)
- Combustion & Propulsion (AREA)
- Transportation (AREA)
- Mechanical Engineering (AREA)
- User Interface Of Digital Computer (AREA)
- Lock And Its Accessories (AREA)
- Telephone Function (AREA)
- Selective Calling Equipment (AREA)
Priority Applications (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR1020177008192A KR101854633B1 (ko) | 2014-08-26 | 2015-08-25 | 인터랙티브 차량 제어 시스템을 위한 통합된 웨어러블 용품 |
DE112015003882.5T DE112015003882B4 (de) | 2014-08-26 | 2015-08-25 | Integrierter tragbarer Artikel für interaktives Fahrzeugsteuerungssystem |
JP2017511229A JP6337199B2 (ja) | 2014-08-26 | 2015-08-25 | 対話式移動体制御システムのための一体化ウェアラブル用品 |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US14/469,041 | 2014-08-26 | ||
US14/469,041 US9760698B2 (en) | 2013-09-17 | 2014-08-26 | Integrated wearable article for interactive vehicle control system |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2016032990A1 true WO2016032990A1 (en) | 2016-03-03 |
Family
ID=54066206
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/US2015/046626 WO2016032990A1 (en) | 2014-08-26 | 2015-08-25 | Integrated wearable article for interactive vehicle control system |
Country Status (4)
Country | Link |
---|---|
JP (1) | JP6337199B2 (ja) |
KR (1) | KR101854633B1 (ja) |
DE (1) | DE112015003882B4 (ja) |
WO (1) | WO2016032990A1 (ja) |
Cited By (93)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20170151956A1 (en) * | 2015-11-27 | 2017-06-01 | Bragi GmbH | Vehicle with wearable for identifying role of one or more users and adjustment of user settings |
WO2018019889A1 (en) * | 2016-07-26 | 2018-02-01 | Bragi GmbH | Earpiece with vehicle forced settings |
DE102016215434A1 (de) * | 2016-08-18 | 2018-02-22 | Continental Automotive Gmbh | Anzeigeanordnung für ein Fahrzeug sowie Fahrzeug mit einer derartigen Anzeigeanordnung |
US9978278B2 (en) | 2015-11-27 | 2018-05-22 | Bragi GmbH | Vehicle to vehicle communications using ear pieces |
US10013542B2 (en) | 2016-04-28 | 2018-07-03 | Bragi GmbH | Biometric interface system and method |
US10015579B2 (en) | 2016-04-08 | 2018-07-03 | Bragi GmbH | Audio accelerometric feedback through bilateral ear worn device system and method |
CN108327722A (zh) * | 2017-01-20 | 2018-07-27 | 本田技研工业株式会社 | 用于通过移动模式来识别车辆驾驶员的系统和方法 |
US10045117B2 (en) | 2016-11-04 | 2018-08-07 | Bragi GmbH | Earpiece with modified ambient environment over-ride function |
US10040423B2 (en) | 2015-11-27 | 2018-08-07 | Bragi GmbH | Vehicle with wearable for identifying one or more vehicle occupants |
US10045112B2 (en) | 2016-11-04 | 2018-08-07 | Bragi GmbH | Earpiece with added ambient environment |
US10045110B2 (en) | 2016-07-06 | 2018-08-07 | Bragi GmbH | Selective sound field environment processing system and method |
US10045116B2 (en) | 2016-03-14 | 2018-08-07 | Bragi GmbH | Explosive sound pressure level active noise cancellation utilizing completely wireless earpieces system and method |
US10049184B2 (en) | 2016-10-07 | 2018-08-14 | Bragi GmbH | Software application transmission via body interface using a wearable device in conjunction with removable body sensor arrays system and method |
US10045736B2 (en) | 2016-07-06 | 2018-08-14 | Bragi GmbH | Detection of metabolic disorders using wireless earpieces |
US10062373B2 (en) | 2016-11-03 | 2018-08-28 | Bragi GmbH | Selective audio isolation from body generated sound system and method |
US10063957B2 (en) | 2016-11-04 | 2018-08-28 | Bragi GmbH | Earpiece with source selection within ambient environment |
US10058282B2 (en) | 2016-11-04 | 2018-08-28 | Bragi GmbH | Manual operation assistance with earpiece with 3D sound cues |
WO2018166906A1 (de) * | 2017-03-13 | 2018-09-20 | HELLA GmbH & Co. KGaA | System für ein kraftfahrzeug, fernbedienung, verfahren zur identifikation eines benutzers einer fernbedienung, computerprogrammprodukt und computerlesbares medium |
US10085082B2 (en) | 2016-03-11 | 2018-09-25 | Bragi GmbH | Earpiece with GPS receiver |
US10085091B2 (en) | 2016-02-09 | 2018-09-25 | Bragi GmbH | Ambient volume modification through environmental microphone feedback loop system and method |
US10104464B2 (en) | 2016-08-25 | 2018-10-16 | Bragi GmbH | Wireless earpiece and smart glasses system and method |
US10104460B2 (en) | 2015-11-27 | 2018-10-16 | Bragi GmbH | Vehicle with interaction between entertainment systems and wearable devices |
US10099636B2 (en) | 2015-11-27 | 2018-10-16 | Bragi GmbH | System and method for determining a user role and user settings associated with a vehicle |
US10104487B2 (en) | 2015-08-29 | 2018-10-16 | Bragi GmbH | Production line PCB serial programming and testing method and system |
US10117604B2 (en) | 2016-11-02 | 2018-11-06 | Bragi GmbH | 3D sound positioning with distributed sensors |
US10122421B2 (en) | 2015-08-29 | 2018-11-06 | Bragi GmbH | Multimodal communication system using induction and radio and method |
US10158934B2 (en) | 2016-07-07 | 2018-12-18 | Bragi GmbH | Case for multiple earpiece pairs |
US10165350B2 (en) | 2016-07-07 | 2018-12-25 | Bragi GmbH | Earpiece with app environment |
CN109249940A (zh) * | 2017-07-04 | 2019-01-22 | 现代自动车株式会社 | 车辆系统及其控制方法 |
US10200780B2 (en) | 2016-08-29 | 2019-02-05 | Bragi GmbH | Method and apparatus for conveying battery life of wireless earpiece |
US10205814B2 (en) | 2016-11-03 | 2019-02-12 | Bragi GmbH | Wireless earpiece with walkie-talkie functionality |
US10212505B2 (en) | 2015-10-20 | 2019-02-19 | Bragi GmbH | Multi-point multiple sensor array for data sensing and processing system and method |
US10216474B2 (en) | 2016-07-06 | 2019-02-26 | Bragi GmbH | Variable computing engine for interactive media based upon user biometrics |
US10225638B2 (en) | 2016-11-03 | 2019-03-05 | Bragi GmbH | Ear piece with pseudolite connectivity |
US10297911B2 (en) | 2015-08-29 | 2019-05-21 | Bragi GmbH | Antenna for use in a wearable device |
US10313779B2 (en) | 2016-08-26 | 2019-06-04 | Bragi GmbH | Voice assistant system for wireless earpieces |
US10327082B2 (en) | 2016-03-02 | 2019-06-18 | Bragi GmbH | Location based tracking using a wireless earpiece device, system, and method |
US10334346B2 (en) | 2016-03-24 | 2019-06-25 | Bragi GmbH | Real-time multivariable biometric analysis and display system and method |
US10344960B2 (en) | 2017-09-19 | 2019-07-09 | Bragi GmbH | Wireless earpiece controlled medical headlight |
US10382854B2 (en) | 2015-08-29 | 2019-08-13 | Bragi GmbH | Near field gesture control system and method |
US10397686B2 (en) | 2016-08-15 | 2019-08-27 | Bragi GmbH | Detection of movement adjacent an earpiece device |
US10397688B2 (en) | 2015-08-29 | 2019-08-27 | Bragi GmbH | Power control for battery powered personal area network device system and method |
US10405081B2 (en) | 2017-02-08 | 2019-09-03 | Bragi GmbH | Intelligent wireless headset system |
US10412478B2 (en) | 2015-08-29 | 2019-09-10 | Bragi GmbH | Reproduction of ambient environmental sound for acoustic transparency of ear canal device system and method |
US10409091B2 (en) | 2016-08-25 | 2019-09-10 | Bragi GmbH | Wearable with lenses |
US10433788B2 (en) | 2016-03-23 | 2019-10-08 | Bragi GmbH | Earpiece life monitor with capability of automatic notification system and method |
US10455313B2 (en) | 2016-10-31 | 2019-10-22 | Bragi GmbH | Wireless earpiece with force feedback |
US10460095B2 (en) | 2016-09-30 | 2019-10-29 | Bragi GmbH | Earpiece with biometric identifiers |
US10469931B2 (en) | 2016-07-07 | 2019-11-05 | Bragi GmbH | Comparative analysis of sensors to control power status for wireless earpieces |
US10506322B2 (en) | 2015-10-20 | 2019-12-10 | Bragi GmbH | Wearable device onboard applications system and method |
US10506327B2 (en) | 2016-12-27 | 2019-12-10 | Bragi GmbH | Ambient environmental sound field manipulation based on user defined voice and audio recognition pattern analysis system and method |
US10555700B2 (en) | 2016-07-06 | 2020-02-11 | Bragi GmbH | Combined optical sensor for audio and pulse oximetry system and method |
US10575086B2 (en) | 2017-03-22 | 2020-02-25 | Bragi GmbH | System and method for sharing wireless earpieces |
US10580282B2 (en) | 2016-09-12 | 2020-03-03 | Bragi GmbH | Ear based contextual environment and biometric pattern recognition system and method |
US10582289B2 (en) | 2015-10-20 | 2020-03-03 | Bragi GmbH | Enhanced biometric control systems for detection of emergency events system and method |
US10582328B2 (en) | 2016-07-06 | 2020-03-03 | Bragi GmbH | Audio response based on user worn microphones to direct or adapt program responses system and method |
US10582290B2 (en) | 2017-02-21 | 2020-03-03 | Bragi GmbH | Earpiece with tap functionality |
US10587943B2 (en) | 2016-07-09 | 2020-03-10 | Bragi GmbH | Earpiece with wirelessly recharging battery |
US10598506B2 (en) | 2016-09-12 | 2020-03-24 | Bragi GmbH | Audio navigation using short range bilateral earpieces |
US10621583B2 (en) | 2016-07-07 | 2020-04-14 | Bragi GmbH | Wearable earpiece multifactorial biometric analysis system and method |
US10617297B2 (en) | 2016-11-02 | 2020-04-14 | Bragi GmbH | Earpiece with in-ear electrodes |
US10620698B2 (en) | 2015-12-21 | 2020-04-14 | Bragi GmbH | Voice dictation systems using earpiece microphone system and method |
US10667033B2 (en) | 2016-03-02 | 2020-05-26 | Bragi GmbH | Multifactorial unlocking function for smart wearable device and method |
US10672239B2 (en) | 2015-08-29 | 2020-06-02 | Bragi GmbH | Responsive visual communication system and method |
US10698983B2 (en) | 2016-10-31 | 2020-06-30 | Bragi GmbH | Wireless earpiece with a medical engine |
US10708699B2 (en) | 2017-05-03 | 2020-07-07 | Bragi GmbH | Hearing aid with added functionality |
US10747337B2 (en) | 2016-04-26 | 2020-08-18 | Bragi GmbH | Mechanical detection of a touch movement using a sensor and a special surface pattern system and method |
US10771877B2 (en) | 2016-10-31 | 2020-09-08 | Bragi GmbH | Dual earpieces for same ear |
US10771881B2 (en) | 2017-02-27 | 2020-09-08 | Bragi GmbH | Earpiece with audio 3D menu |
US10821361B2 (en) | 2016-11-03 | 2020-11-03 | Bragi GmbH | Gaming with earpiece 3D audio |
CN111918807A (zh) * | 2018-03-28 | 2020-11-10 | 日立汽车系统株式会社 | 车辆控制装置以及车辆控制系统 |
US10852829B2 (en) | 2016-09-13 | 2020-12-01 | Bragi GmbH | Measurement of facial muscle EMG potentials for predictive analysis using a smart wearable system and method |
US10856809B2 (en) | 2016-03-24 | 2020-12-08 | Bragi GmbH | Earpiece with glucose sensor and system |
US10887679B2 (en) | 2016-08-26 | 2021-01-05 | Bragi GmbH | Earpiece for audiograms |
US10888039B2 (en) | 2016-07-06 | 2021-01-05 | Bragi GmbH | Shielded case for wireless earpieces |
EP3767920A1 (en) * | 2017-10-03 | 2021-01-20 | Google LLC | Multi-factor authentication and access control in a vehicular environment |
US10904653B2 (en) | 2015-12-21 | 2021-01-26 | Bragi GmbH | Microphone natural speech capture voice dictation system and method |
US10942701B2 (en) | 2016-10-31 | 2021-03-09 | Bragi GmbH | Input and edit functions utilizing accelerometer based earpiece movement system and method |
US10977348B2 (en) | 2016-08-24 | 2021-04-13 | Bragi GmbH | Digital signature using phonometry and compiled biometric data system and method |
US11013445B2 (en) | 2017-06-08 | 2021-05-25 | Bragi GmbH | Wireless earpiece with transcranial stimulation |
US11064408B2 (en) | 2015-10-20 | 2021-07-13 | Bragi GmbH | Diversity bluetooth system and method |
US11086593B2 (en) | 2016-08-26 | 2021-08-10 | Bragi GmbH | Voice assistant for wireless earpieces |
US11085871B2 (en) | 2016-07-06 | 2021-08-10 | Bragi GmbH | Optical vibration detection system and method |
US11116415B2 (en) | 2017-06-07 | 2021-09-14 | Bragi GmbH | Use of body-worn radar for biometric measurements, contextual awareness and identification |
US11200026B2 (en) | 2016-08-26 | 2021-12-14 | Bragi GmbH | Wireless earpiece with a passive virtual assistant |
US11272367B2 (en) | 2017-09-20 | 2022-03-08 | Bragi GmbH | Wireless earpieces for hub communications |
US11283742B2 (en) | 2016-09-27 | 2022-03-22 | Bragi GmbH | Audio-based social media platform |
US11380430B2 (en) | 2017-03-22 | 2022-07-05 | Bragi GmbH | System and method for populating electronic medical records with wireless earpieces |
US11490858B2 (en) | 2016-08-31 | 2022-11-08 | Bragi GmbH | Disposable sensor array wearable device sleeve system and method |
US11544104B2 (en) | 2017-03-22 | 2023-01-03 | Bragi GmbH | Load sharing between wireless earpieces |
US11694771B2 (en) | 2017-03-22 | 2023-07-04 | Bragi GmbH | System and method for populating electronic health records with wireless earpieces |
US11772603B2 (en) | 2021-05-18 | 2023-10-03 | Motional Ad Llc | Passenger authentication and entry for autonomous vehicles |
US11799852B2 (en) | 2016-03-29 | 2023-10-24 | Bragi GmbH | Wireless dongle for communications with wireless earpieces |
Families Citing this family (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11009963B2 (en) | 2016-05-20 | 2021-05-18 | Ford Global Technologies, Llc | Sign language inputs to a vehicle user interface |
DE102017007275A1 (de) | 2017-08-01 | 2018-04-19 | Daimler Ag | Verfahren zur Ausgabe einer fahrzeugrelevanten Information |
KR102417524B1 (ko) * | 2017-10-13 | 2022-07-07 | 현대자동차주식회사 | 음성 인식 기반의 자동차 제어 방법 |
WO2019181143A1 (ja) * | 2018-03-22 | 2019-09-26 | 三菱自動車工業株式会社 | 車両制御システム |
DE102018220429A1 (de) * | 2018-11-28 | 2020-05-28 | Volkswagen Aktiengesellschaft | Verfahren zum Anzeigen von Ladeinformationen eines Fahrzeugs |
DE102019103867A1 (de) * | 2019-02-15 | 2020-08-20 | Bayerische Motoren Werke Aktiengesellschaft | Anzeigeeinrichtung für eine Außenverkleidung eines Kraftfahrzeugs und Kraftfahrzeug mit einer solchen Anzeigeeinrichtung |
JP2020203546A (ja) * | 2019-06-14 | 2020-12-24 | 株式会社シマノ | 検出装置、検出方法、生成方法、コンピュータプログラム、および記憶媒体 |
DE102019118183A1 (de) * | 2019-07-05 | 2021-01-07 | Bayerische Motoren Werke Aktiengesellschaft | System und Verfahren zur Information |
DE102019119534A1 (de) * | 2019-07-18 | 2021-01-21 | Bayerische Motoren Werke Aktiengesellschaft | Vorrichtung, system und verfahren zum bedienen eines fahrzeugs sowie fahrzeug, insbesondere kraftfahrzeug |
DE102019133663A1 (de) * | 2019-12-10 | 2021-06-10 | Bayerische Motoren Werke Aktiengesellschaft | Vorrichtung zum Anzeigen und Steuern von Funktionen eines Fahrzeugs |
DE102019219272A1 (de) * | 2019-12-10 | 2021-06-10 | Continental Automotive Gmbh | Fahrzeug mit einem Innenraum und Verfahren zur Steuerung einer Anzeigeeinheit in einem Fahrzeug |
US20230036233A1 (en) * | 2020-01-29 | 2023-02-02 | Nec Corporation | Information processing device, information processing method, and storage medium |
DE102020107066A1 (de) | 2020-03-14 | 2021-09-16 | Audi Aktiengesellschaft | Verfahren zum Betreiben einer Ausgabevorrichtung eines Kraftfahrzeugs, Steuereinrichtung, und Kraftfahrzeug |
DE102021124586A1 (de) | 2021-09-22 | 2023-03-23 | Bayerische Motoren Werke Aktiengesellschaft | System und Verfahren zur agilen, intuitiven Steuerung von Fahrzeugfunktionen |
DE102022206977A1 (de) | 2022-07-08 | 2024-01-11 | Zf Friedrichshafen Ag | Computer-implementiertes Verfahren und Vorrichtung zum Steuern eines zumindest teilweise autonom, insbesondere vollständig autonom fahrenden Fahrzeugs |
Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP1970265A2 (en) * | 2007-03-15 | 2008-09-17 | Jaguar Cars Limited | Security system for a motor vehicle |
US20090146947A1 (en) * | 2007-12-07 | 2009-06-11 | James Ng | Universal wearable input and authentication device |
US20100075656A1 (en) * | 2008-09-24 | 2010-03-25 | Embarq Holdings Company, Llc | System and method for key free access to a vehicle |
US20100083373A1 (en) * | 2008-09-29 | 2010-04-01 | Scott White | Methods and apparatus for determining user authorization from motion of a gesture-based control unit |
US20120296559A1 (en) * | 2011-05-18 | 2012-11-22 | Gueziec Andre | System for providing traffic data and driving efficiency data |
US8463488B1 (en) * | 2010-06-24 | 2013-06-11 | Paul Hart | Vehicle profile control and monitoring |
US20130231800A1 (en) * | 2011-11-16 | 2013-09-05 | Flextronics Ap, Llc | Vehicle occupant health data gathering and monitoring |
DE102012203535A1 (de) * | 2012-03-06 | 2013-09-12 | Bayerische Motoren Werke Aktiengesellschaft | Tastenloser Fahrzeugschlüssel mit Gestikerkennung |
Family Cites Families (20)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6227862B1 (en) | 1999-02-12 | 2001-05-08 | Advanced Drivers Education Products And Training, Inc. | Driver training system |
JP2001304896A (ja) * | 2000-04-25 | 2001-10-31 | Mitsubishi Motors Corp | 車両用ナビゲーション装置 |
US7474592B2 (en) | 2001-12-05 | 2009-01-06 | Hewlett-Packard Development Company, L.P. | Secure operation of a versatile device based on whether an authenticated user continues to wear the versatile device after initiating its use |
US7205884B2 (en) | 2002-12-19 | 2007-04-17 | Denso Corporation | Vehicle electronic key system |
DE10361115A1 (de) | 2003-12-22 | 2005-07-21 | Daimlerchrysler Ag | Verfahren zur Fernbedienung von Türen und/oder Klappen für Fahrzeuge und zugehöriges Fernbediensystem |
US20060012679A1 (en) | 2004-07-14 | 2006-01-19 | Ressler Galen E | Multifunction vehicle interior imaging system |
US7248151B2 (en) | 2005-01-05 | 2007-07-24 | General Motors Corporation | Virtual keypad for vehicle entry control |
JP2007210457A (ja) * | 2006-02-09 | 2007-08-23 | Fujitsu Ten Ltd | 車両自動設定装置及び設定方法 |
US8427276B2 (en) | 2006-02-24 | 2013-04-23 | Denso International America, Inc. | Apparatus for automatically initiating sequence of vehicle functions |
JP4441887B2 (ja) | 2006-03-31 | 2010-03-31 | 株式会社デンソー | 自動車用ユーザーもてなしシステム |
JP2008143220A (ja) * | 2006-12-06 | 2008-06-26 | Tokai Rika Co Ltd | 個人認証システム |
DE102007002257A1 (de) | 2007-01-16 | 2008-07-17 | GM Global Technology Operations, Inc., Detroit | Fensterscheibenanordnung, Verfahren zum Erfassen von mechanischen Einflüssen auf eine Fensterscheibe und Kraftfahrzeug |
KR101331827B1 (ko) * | 2007-01-31 | 2013-11-22 | 최윤정 | 차량용 디스플레이 장치 및 이를 이용한 디스플레이 방법 |
JP2008225889A (ja) * | 2007-03-13 | 2008-09-25 | Pioneer Electronic Corp | 情報提供装置及び情報提供方法等 |
US20080238667A1 (en) | 2007-03-30 | 2008-10-02 | Proxwear, Llc | Clothing and Accessories that Operate Radio Frequency Identification Enabled Security Devices |
US8606430B2 (en) * | 2010-10-08 | 2013-12-10 | GM Global Technology Operations LLC | External presentation of information on full glass display |
US20120249291A1 (en) * | 2011-03-29 | 2012-10-04 | Denso Corporation | Systems and methods for vehicle passive entry |
US9032493B2 (en) | 2011-03-31 | 2015-05-12 | Intel Corporation | Connecting mobile devices, internet-connected vehicles, and cloud services |
EP2798626A4 (en) | 2011-12-29 | 2016-01-06 | Intel Corp | PERSONALIZED RECONFIGURABLE VEHICLE INDICATORS |
JP2014088730A (ja) * | 2012-10-31 | 2014-05-15 | Mitsubishi Electric Corp | 携帯通信機およびドア制御装置 |
-
2015
- 2015-08-25 KR KR1020177008192A patent/KR101854633B1/ko active IP Right Grant
- 2015-08-25 JP JP2017511229A patent/JP6337199B2/ja active Active
- 2015-08-25 WO PCT/US2015/046626 patent/WO2016032990A1/en active Application Filing
- 2015-08-25 DE DE112015003882.5T patent/DE112015003882B4/de active Active
Patent Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP1970265A2 (en) * | 2007-03-15 | 2008-09-17 | Jaguar Cars Limited | Security system for a motor vehicle |
US20090146947A1 (en) * | 2007-12-07 | 2009-06-11 | James Ng | Universal wearable input and authentication device |
US20100075656A1 (en) * | 2008-09-24 | 2010-03-25 | Embarq Holdings Company, Llc | System and method for key free access to a vehicle |
US20100083373A1 (en) * | 2008-09-29 | 2010-04-01 | Scott White | Methods and apparatus for determining user authorization from motion of a gesture-based control unit |
US8463488B1 (en) * | 2010-06-24 | 2013-06-11 | Paul Hart | Vehicle profile control and monitoring |
US20120296559A1 (en) * | 2011-05-18 | 2012-11-22 | Gueziec Andre | System for providing traffic data and driving efficiency data |
US20130231800A1 (en) * | 2011-11-16 | 2013-09-05 | Flextronics Ap, Llc | Vehicle occupant health data gathering and monitoring |
DE102012203535A1 (de) * | 2012-03-06 | 2013-09-12 | Bayerische Motoren Werke Aktiengesellschaft | Tastenloser Fahrzeugschlüssel mit Gestikerkennung |
US20140379175A1 (en) * | 2012-03-06 | 2014-12-25 | Bayerische Motoren Werke Aktiengesellschaft | Buttonless Vehicle Key Having Gesture Recognition |
Cited By (139)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10122421B2 (en) | 2015-08-29 | 2018-11-06 | Bragi GmbH | Multimodal communication system using induction and radio and method |
US10104487B2 (en) | 2015-08-29 | 2018-10-16 | Bragi GmbH | Production line PCB serial programming and testing method and system |
US10297911B2 (en) | 2015-08-29 | 2019-05-21 | Bragi GmbH | Antenna for use in a wearable device |
US10412478B2 (en) | 2015-08-29 | 2019-09-10 | Bragi GmbH | Reproduction of ambient environmental sound for acoustic transparency of ear canal device system and method |
US10439679B2 (en) | 2015-08-29 | 2019-10-08 | Bragi GmbH | Multimodal communication system using induction and radio and method |
US10672239B2 (en) | 2015-08-29 | 2020-06-02 | Bragi GmbH | Responsive visual communication system and method |
US10397688B2 (en) | 2015-08-29 | 2019-08-27 | Bragi GmbH | Power control for battery powered personal area network device system and method |
US10382854B2 (en) | 2015-08-29 | 2019-08-13 | Bragi GmbH | Near field gesture control system and method |
US10506322B2 (en) | 2015-10-20 | 2019-12-10 | Bragi GmbH | Wearable device onboard applications system and method |
US11064408B2 (en) | 2015-10-20 | 2021-07-13 | Bragi GmbH | Diversity bluetooth system and method |
US11683735B2 (en) | 2015-10-20 | 2023-06-20 | Bragi GmbH | Diversity bluetooth system and method |
US11419026B2 (en) | 2015-10-20 | 2022-08-16 | Bragi GmbH | Diversity Bluetooth system and method |
US10582289B2 (en) | 2015-10-20 | 2020-03-03 | Bragi GmbH | Enhanced biometric control systems for detection of emergency events system and method |
US10212505B2 (en) | 2015-10-20 | 2019-02-19 | Bragi GmbH | Multi-point multiple sensor array for data sensing and processing system and method |
US10104460B2 (en) | 2015-11-27 | 2018-10-16 | Bragi GmbH | Vehicle with interaction between entertainment systems and wearable devices |
US9978278B2 (en) | 2015-11-27 | 2018-05-22 | Bragi GmbH | Vehicle to vehicle communications using ear pieces |
US9944295B2 (en) | 2015-11-27 | 2018-04-17 | Bragi GmbH | Vehicle with wearable for identifying role of one or more users and adjustment of user settings |
US20170151956A1 (en) * | 2015-11-27 | 2017-06-01 | Bragi GmbH | Vehicle with wearable for identifying role of one or more users and adjustment of user settings |
US10155524B2 (en) | 2015-11-27 | 2018-12-18 | Bragi GmbH | Vehicle with wearable for identifying role of one or more users and adjustment of user settings |
US10040423B2 (en) | 2015-11-27 | 2018-08-07 | Bragi GmbH | Vehicle with wearable for identifying one or more vehicle occupants |
US10099636B2 (en) | 2015-11-27 | 2018-10-16 | Bragi GmbH | System and method for determining a user role and user settings associated with a vehicle |
US10904653B2 (en) | 2015-12-21 | 2021-01-26 | Bragi GmbH | Microphone natural speech capture voice dictation system and method |
US11496827B2 (en) | 2015-12-21 | 2022-11-08 | Bragi GmbH | Microphone natural speech capture voice dictation system and method |
US10620698B2 (en) | 2015-12-21 | 2020-04-14 | Bragi GmbH | Voice dictation systems using earpiece microphone system and method |
US10085091B2 (en) | 2016-02-09 | 2018-09-25 | Bragi GmbH | Ambient volume modification through environmental microphone feedback loop system and method |
US10412493B2 (en) | 2016-02-09 | 2019-09-10 | Bragi GmbH | Ambient volume modification through environmental microphone feedback loop system and method |
US10327082B2 (en) | 2016-03-02 | 2019-06-18 | Bragi GmbH | Location based tracking using a wireless earpiece device, system, and method |
US10667033B2 (en) | 2016-03-02 | 2020-05-26 | Bragi GmbH | Multifactorial unlocking function for smart wearable device and method |
US10085082B2 (en) | 2016-03-11 | 2018-09-25 | Bragi GmbH | Earpiece with GPS receiver |
US11336989B2 (en) | 2016-03-11 | 2022-05-17 | Bragi GmbH | Earpiece with GPS receiver |
US11968491B2 (en) | 2016-03-11 | 2024-04-23 | Bragi GmbH | Earpiece with GPS receiver |
US11700475B2 (en) | 2016-03-11 | 2023-07-11 | Bragi GmbH | Earpiece with GPS receiver |
US10893353B2 (en) | 2016-03-11 | 2021-01-12 | Bragi GmbH | Earpiece with GPS receiver |
US10506328B2 (en) | 2016-03-14 | 2019-12-10 | Bragi GmbH | Explosive sound pressure level active noise cancellation |
US10045116B2 (en) | 2016-03-14 | 2018-08-07 | Bragi GmbH | Explosive sound pressure level active noise cancellation utilizing completely wireless earpieces system and method |
US10433788B2 (en) | 2016-03-23 | 2019-10-08 | Bragi GmbH | Earpiece life monitor with capability of automatic notification system and method |
US10856809B2 (en) | 2016-03-24 | 2020-12-08 | Bragi GmbH | Earpiece with glucose sensor and system |
US10334346B2 (en) | 2016-03-24 | 2019-06-25 | Bragi GmbH | Real-time multivariable biometric analysis and display system and method |
US11799852B2 (en) | 2016-03-29 | 2023-10-24 | Bragi GmbH | Wireless dongle for communications with wireless earpieces |
US10015579B2 (en) | 2016-04-08 | 2018-07-03 | Bragi GmbH | Audio accelerometric feedback through bilateral ear worn device system and method |
US10313781B2 (en) | 2016-04-08 | 2019-06-04 | Bragi GmbH | Audio accelerometric feedback through bilateral ear worn device system and method |
US10747337B2 (en) | 2016-04-26 | 2020-08-18 | Bragi GmbH | Mechanical detection of a touch movement using a sensor and a special surface pattern system and method |
US10013542B2 (en) | 2016-04-28 | 2018-07-03 | Bragi GmbH | Biometric interface system and method |
US10169561B2 (en) | 2016-04-28 | 2019-01-01 | Bragi GmbH | Biometric interface system and method |
US10582328B2 (en) | 2016-07-06 | 2020-03-03 | Bragi GmbH | Audio response based on user worn microphones to direct or adapt program responses system and method |
US10470709B2 (en) | 2016-07-06 | 2019-11-12 | Bragi GmbH | Detection of metabolic disorders using wireless earpieces |
US11497150B2 (en) | 2016-07-06 | 2022-11-08 | Bragi GmbH | Shielded case for wireless earpieces |
US10045110B2 (en) | 2016-07-06 | 2018-08-07 | Bragi GmbH | Selective sound field environment processing system and method |
US11770918B2 (en) | 2016-07-06 | 2023-09-26 | Bragi GmbH | Shielded case for wireless earpieces |
US11085871B2 (en) | 2016-07-06 | 2021-08-10 | Bragi GmbH | Optical vibration detection system and method |
US10045736B2 (en) | 2016-07-06 | 2018-08-14 | Bragi GmbH | Detection of metabolic disorders using wireless earpieces |
US11781971B2 (en) | 2016-07-06 | 2023-10-10 | Bragi GmbH | Optical vibration detection system and method |
US10888039B2 (en) | 2016-07-06 | 2021-01-05 | Bragi GmbH | Shielded case for wireless earpieces |
US10216474B2 (en) | 2016-07-06 | 2019-02-26 | Bragi GmbH | Variable computing engine for interactive media based upon user biometrics |
US10201309B2 (en) | 2016-07-06 | 2019-02-12 | Bragi GmbH | Detection of physiological data using radar/lidar of wireless earpieces |
US10448139B2 (en) | 2016-07-06 | 2019-10-15 | Bragi GmbH | Selective sound field environment processing system and method |
US10555700B2 (en) | 2016-07-06 | 2020-02-11 | Bragi GmbH | Combined optical sensor for audio and pulse oximetry system and method |
US10516930B2 (en) | 2016-07-07 | 2019-12-24 | Bragi GmbH | Comparative analysis of sensors to control power status for wireless earpieces |
US10469931B2 (en) | 2016-07-07 | 2019-11-05 | Bragi GmbH | Comparative analysis of sensors to control power status for wireless earpieces |
US10158934B2 (en) | 2016-07-07 | 2018-12-18 | Bragi GmbH | Case for multiple earpiece pairs |
US10621583B2 (en) | 2016-07-07 | 2020-04-14 | Bragi GmbH | Wearable earpiece multifactorial biometric analysis system and method |
US10165350B2 (en) | 2016-07-07 | 2018-12-25 | Bragi GmbH | Earpiece with app environment |
US10587943B2 (en) | 2016-07-09 | 2020-03-10 | Bragi GmbH | Earpiece with wirelessly recharging battery |
WO2018019889A1 (en) * | 2016-07-26 | 2018-02-01 | Bragi GmbH | Earpiece with vehicle forced settings |
US10397686B2 (en) | 2016-08-15 | 2019-08-27 | Bragi GmbH | Detection of movement adjacent an earpiece device |
DE102016215434A1 (de) * | 2016-08-18 | 2018-02-22 | Continental Automotive Gmbh | Anzeigeanordnung für ein Fahrzeug sowie Fahrzeug mit einer derartigen Anzeigeanordnung |
US10977348B2 (en) | 2016-08-24 | 2021-04-13 | Bragi GmbH | Digital signature using phonometry and compiled biometric data system and method |
US11620368B2 (en) | 2016-08-24 | 2023-04-04 | Bragi GmbH | Digital signature using phonometry and compiled biometric data system and method |
US12001537B2 (en) | 2016-08-24 | 2024-06-04 | Bragi GmbH | Digital signature using phonometry and compiled biometric data system and method |
US10104464B2 (en) | 2016-08-25 | 2018-10-16 | Bragi GmbH | Wireless earpiece and smart glasses system and method |
US10409091B2 (en) | 2016-08-25 | 2019-09-10 | Bragi GmbH | Wearable with lenses |
US10887679B2 (en) | 2016-08-26 | 2021-01-05 | Bragi GmbH | Earpiece for audiograms |
US11200026B2 (en) | 2016-08-26 | 2021-12-14 | Bragi GmbH | Wireless earpiece with a passive virtual assistant |
US11861266B2 (en) | 2016-08-26 | 2024-01-02 | Bragi GmbH | Voice assistant for wireless earpieces |
US11573763B2 (en) | 2016-08-26 | 2023-02-07 | Bragi GmbH | Voice assistant for wireless earpieces |
US11086593B2 (en) | 2016-08-26 | 2021-08-10 | Bragi GmbH | Voice assistant for wireless earpieces |
US10313779B2 (en) | 2016-08-26 | 2019-06-04 | Bragi GmbH | Voice assistant system for wireless earpieces |
US10200780B2 (en) | 2016-08-29 | 2019-02-05 | Bragi GmbH | Method and apparatus for conveying battery life of wireless earpiece |
US11490858B2 (en) | 2016-08-31 | 2022-11-08 | Bragi GmbH | Disposable sensor array wearable device sleeve system and method |
US10598506B2 (en) | 2016-09-12 | 2020-03-24 | Bragi GmbH | Audio navigation using short range bilateral earpieces |
US10580282B2 (en) | 2016-09-12 | 2020-03-03 | Bragi GmbH | Ear based contextual environment and biometric pattern recognition system and method |
US11675437B2 (en) | 2016-09-13 | 2023-06-13 | Bragi GmbH | Measurement of facial muscle EMG potentials for predictive analysis using a smart wearable system and method |
US11294466B2 (en) | 2016-09-13 | 2022-04-05 | Bragi GmbH | Measurement of facial muscle EMG potentials for predictive analysis using a smart wearable system and method |
US10852829B2 (en) | 2016-09-13 | 2020-12-01 | Bragi GmbH | Measurement of facial muscle EMG potentials for predictive analysis using a smart wearable system and method |
US11956191B2 (en) | 2016-09-27 | 2024-04-09 | Bragi GmbH | Audio-based social media platform |
US11283742B2 (en) | 2016-09-27 | 2022-03-22 | Bragi GmbH | Audio-based social media platform |
US11627105B2 (en) | 2016-09-27 | 2023-04-11 | Bragi GmbH | Audio-based social media platform |
US10460095B2 (en) | 2016-09-30 | 2019-10-29 | Bragi GmbH | Earpiece with biometric identifiers |
US10049184B2 (en) | 2016-10-07 | 2018-08-14 | Bragi GmbH | Software application transmission via body interface using a wearable device in conjunction with removable body sensor arrays system and method |
US10455313B2 (en) | 2016-10-31 | 2019-10-22 | Bragi GmbH | Wireless earpiece with force feedback |
US10771877B2 (en) | 2016-10-31 | 2020-09-08 | Bragi GmbH | Dual earpieces for same ear |
US11947874B2 (en) | 2016-10-31 | 2024-04-02 | Bragi GmbH | Input and edit functions utilizing accelerometer based earpiece movement system and method |
US10698983B2 (en) | 2016-10-31 | 2020-06-30 | Bragi GmbH | Wireless earpiece with a medical engine |
US10942701B2 (en) | 2016-10-31 | 2021-03-09 | Bragi GmbH | Input and edit functions utilizing accelerometer based earpiece movement system and method |
US11599333B2 (en) | 2016-10-31 | 2023-03-07 | Bragi GmbH | Input and edit functions utilizing accelerometer based earpiece movement system and method |
US10617297B2 (en) | 2016-11-02 | 2020-04-14 | Bragi GmbH | Earpiece with in-ear electrodes |
US10117604B2 (en) | 2016-11-02 | 2018-11-06 | Bragi GmbH | 3D sound positioning with distributed sensors |
US11806621B2 (en) | 2016-11-03 | 2023-11-07 | Bragi GmbH | Gaming with earpiece 3D audio |
US10225638B2 (en) | 2016-11-03 | 2019-03-05 | Bragi GmbH | Ear piece with pseudolite connectivity |
US10205814B2 (en) | 2016-11-03 | 2019-02-12 | Bragi GmbH | Wireless earpiece with walkie-talkie functionality |
US11908442B2 (en) | 2016-11-03 | 2024-02-20 | Bragi GmbH | Selective audio isolation from body generated sound system and method |
US10062373B2 (en) | 2016-11-03 | 2018-08-28 | Bragi GmbH | Selective audio isolation from body generated sound system and method |
US10821361B2 (en) | 2016-11-03 | 2020-11-03 | Bragi GmbH | Gaming with earpiece 3D audio |
US10896665B2 (en) | 2016-11-03 | 2021-01-19 | Bragi GmbH | Selective audio isolation from body generated sound system and method |
US11417307B2 (en) | 2016-11-03 | 2022-08-16 | Bragi GmbH | Selective audio isolation from body generated sound system and method |
US11325039B2 (en) | 2016-11-03 | 2022-05-10 | Bragi GmbH | Gaming with earpiece 3D audio |
US10058282B2 (en) | 2016-11-04 | 2018-08-28 | Bragi GmbH | Manual operation assistance with earpiece with 3D sound cues |
US10045117B2 (en) | 2016-11-04 | 2018-08-07 | Bragi GmbH | Earpiece with modified ambient environment over-ride function |
US10063957B2 (en) | 2016-11-04 | 2018-08-28 | Bragi GmbH | Earpiece with source selection within ambient environment |
US10045112B2 (en) | 2016-11-04 | 2018-08-07 | Bragi GmbH | Earpiece with added ambient environment |
US10681449B2 (en) | 2016-11-04 | 2020-06-09 | Bragi GmbH | Earpiece with added ambient environment |
US10681450B2 (en) | 2016-11-04 | 2020-06-09 | Bragi GmbH | Earpiece with source selection within ambient environment |
US10397690B2 (en) | 2016-11-04 | 2019-08-27 | Bragi GmbH | Earpiece with modified ambient environment over-ride function |
US10398374B2 (en) | 2016-11-04 | 2019-09-03 | Bragi GmbH | Manual operation assistance with earpiece with 3D sound cues |
US10506327B2 (en) | 2016-12-27 | 2019-12-10 | Bragi GmbH | Ambient environmental sound field manipulation based on user defined voice and audio recognition pattern analysis system and method |
CN108327722A (zh) * | 2017-01-20 | 2018-07-27 | 本田技研工业株式会社 | 用于通过移动模式来识别车辆驾驶员的系统和方法 |
US10405081B2 (en) | 2017-02-08 | 2019-09-03 | Bragi GmbH | Intelligent wireless headset system |
US10582290B2 (en) | 2017-02-21 | 2020-03-03 | Bragi GmbH | Earpiece with tap functionality |
US10771881B2 (en) | 2017-02-27 | 2020-09-08 | Bragi GmbH | Earpiece with audio 3D menu |
WO2018166906A1 (de) * | 2017-03-13 | 2018-09-20 | HELLA GmbH & Co. KGaA | System für ein kraftfahrzeug, fernbedienung, verfahren zur identifikation eines benutzers einer fernbedienung, computerprogrammprodukt und computerlesbares medium |
US11544104B2 (en) | 2017-03-22 | 2023-01-03 | Bragi GmbH | Load sharing between wireless earpieces |
US10575086B2 (en) | 2017-03-22 | 2020-02-25 | Bragi GmbH | System and method for sharing wireless earpieces |
US11380430B2 (en) | 2017-03-22 | 2022-07-05 | Bragi GmbH | System and method for populating electronic medical records with wireless earpieces |
US11694771B2 (en) | 2017-03-22 | 2023-07-04 | Bragi GmbH | System and method for populating electronic health records with wireless earpieces |
US11710545B2 (en) | 2017-03-22 | 2023-07-25 | Bragi GmbH | System and method for populating electronic medical records with wireless earpieces |
US10708699B2 (en) | 2017-05-03 | 2020-07-07 | Bragi GmbH | Hearing aid with added functionality |
US11116415B2 (en) | 2017-06-07 | 2021-09-14 | Bragi GmbH | Use of body-worn radar for biometric measurements, contextual awareness and identification |
US11013445B2 (en) | 2017-06-08 | 2021-05-25 | Bragi GmbH | Wireless earpiece with transcranial stimulation |
US11911163B2 (en) | 2017-06-08 | 2024-02-27 | Bragi GmbH | Wireless earpiece with transcranial stimulation |
CN109249940A (zh) * | 2017-07-04 | 2019-01-22 | 现代自动车株式会社 | 车辆系统及其控制方法 |
US10344960B2 (en) | 2017-09-19 | 2019-07-09 | Bragi GmbH | Wireless earpiece controlled medical headlight |
US11711695B2 (en) | 2017-09-20 | 2023-07-25 | Bragi GmbH | Wireless earpieces for hub communications |
US11272367B2 (en) | 2017-09-20 | 2022-03-08 | Bragi GmbH | Wireless earpieces for hub communications |
US20230052913A1 (en) | 2017-10-03 | 2023-02-16 | Google Llc | Multi-factor authentication and access control in a vehicular environment |
EP3767920A1 (en) * | 2017-10-03 | 2021-01-20 | Google LLC | Multi-factor authentication and access control in a vehicular environment |
US11856399B2 (en) | 2017-10-03 | 2023-12-26 | Google Llc | Multi-factor authentication and access control in a vehicular environment |
US11528605B2 (en) | 2017-10-03 | 2022-12-13 | Google Llc | Multi-factor authentication and access control in a vehicular environment |
CN111918807A (zh) * | 2018-03-28 | 2020-11-10 | 日立汽车系统株式会社 | 车辆控制装置以及车辆控制系统 |
US11772603B2 (en) | 2021-05-18 | 2023-10-03 | Motional Ad Llc | Passenger authentication and entry for autonomous vehicles |
Also Published As
Publication number | Publication date |
---|---|
JP6337199B2 (ja) | 2018-06-06 |
KR20170044731A (ko) | 2017-04-25 |
DE112015003882T5 (de) | 2017-06-01 |
KR101854633B1 (ko) | 2018-05-04 |
JP2017533609A (ja) | 2017-11-09 |
DE112015003882B4 (de) | 2023-04-27 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US9760698B2 (en) | Integrated wearable article for interactive vehicle control system | |
KR101854633B1 (ko) | 인터랙티브 차량 제어 시스템을 위한 통합된 웨어러블 용품 | |
EP3047236B1 (en) | Interactive vehicle window display system with user identification and vehicle function control | |
US9902266B2 (en) | Interactive vehicle window display system with personal convenience reminders | |
EP2985571B1 (en) | Method for remote communication with and through a vehicle | |
US9400564B2 (en) | Interactive vehicle window display system with a safe driving reminder system | |
US9387824B2 (en) | Interactive vehicle window display system with user identification and image recording | |
US9807196B2 (en) | Automated social network interaction system for a vehicle | |
CN104691449B (zh) | 车辆控制设备及其方法 | |
US9977593B2 (en) | Gesture recognition for on-board display | |
KR102504746B1 (ko) | 신뢰 모바일 컴퓨팅 디바이스와 함께 차량 내 카메라를 사용한 무결절 운전자 인증 | |
KR101698102B1 (ko) | 차량 제어 장치 및 이의 제어 방법 | |
KR102220367B1 (ko) | 이동 단말기 및 그 제어 방법 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 15760560 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 2017511229 Country of ref document: JP Kind code of ref document: A |
|
WWE | Wipo information: entry into national phase |
Ref document number: 112015003882 Country of ref document: DE |
|
ENP | Entry into the national phase |
Ref document number: 20177008192 Country of ref document: KR Kind code of ref document: A |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 15760560 Country of ref document: EP Kind code of ref document: A1 |