GB2489906A - Tactile navigation unit for the visually impaired - Google Patents
Tactile navigation unit for the visually impaired Download PDFInfo
- Publication number
- GB2489906A GB2489906A GB201011874A GB201011874A GB2489906A GB 2489906 A GB2489906 A GB 2489906A GB 201011874 A GB201011874 A GB 201011874A GB 201011874 A GB201011874 A GB 201011874A GB 2489906 A GB2489906 A GB 2489906A
- Authority
- GB
- United Kingdom
- Prior art keywords
- user
- tactile
- output
- unit
- data
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Withdrawn
Links
Classifications
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01C—MEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
- G01C21/00—Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00
- G01C21/26—Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00 specially adapted for navigation in a road network
- G01C21/34—Route searching; Route guidance
- G01C21/36—Input/output arrangements for on-board computers
- G01C21/3626—Details of the output of route guidance instructions
- G01C21/3652—Guidance using non-audiovisual output, e.g. tactile, haptic or electric stimuli
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01C—MEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
- G01C21/00—Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00
- G01C21/20—Instruments for performing navigational calculations
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09B—EDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
- G09B21/00—Teaching, or communicating with, the blind, deaf or mute
- G09B21/001—Teaching or communicating with blind persons
- G09B21/003—Teaching or communicating with blind persons using tactile presentation of the information, e.g. Braille displays
Abstract
Personal guidance apparatus for directing a visually impaired or blind user from a first location to a predefined destination, the apparatus comprising an input unit operable to receive first location data and input destination data; a controller in communication with the input unit and operable to receive first location data and input destination data from the input unit and generate output direction data; and an output unit provided with an array of tactile zones 78a, 78b, 78c, the output unit operable to receive output direction data from the controller and generate a tactually-detectable output pattern at the array to communicate directional information to the user. The tactile zones can produce vibratory signals which can be pulsed. The guidance unit also allows the user to alter the output signals, such as changing the intensity of the stimulation, to suit their individual needs. The tactile elements of the device could be attached to the handle of a walking stick or the handle of a wheelchair.
Description
TACTILE FEEDBACK UNIT
çgpundtotheipention The present invention relates to a guidance device incorporating a tactile interface, and preferably such a guidance device for use by a visually impaired user.
There are several conventional guidance devices on the market intended to assist with travelling to a destination, Most of these rely on satellite navigation and are intended for use when navigating a vehicle. Although personal guidance systems exist, these are often bulky and consist of multiple units, so are inconvenient for carrying, particularly by pedestrians. Furthermore, such personal guidance systems, such as, for example, those downloadabie onto Mi mobile phones and the like, require the operation of a small user interface and visualise directional information on a small display screen. These systems are clearly unsuitable for the visually impaired.
There are a few systems avai.iabl.e for use by the visuall.y impaired to assist with guiding the user from a first location to a specified destination. However, these also have limitations. For example, the feedback systems to provide directional information iuay be too prescriptive and not suitable for all users. In addition, some systems may involve large or obtrusive units or include multiple units for location on different parts of the body with noticeable cables connecting each unit.
As visually impaired users are typically very independent, any systems that draw the attention of others to their impainnent are clearly undesirable.
There is therefore a need for a system which is unobtrusive and provides directional information in an accessible way which is suitable for use by all users.
gryfffiçJnvenition The present invention seeks to address the problems of the prior art.
A first aspect of the present invention relates to a personal guidance apparatus for directing a user from a first location to a predefincd destination, the apparatus comprising: an input unit operable to receive first location data and input destination data; a controller in communication with the input unit and operable to receive first location data and input destination data from the input unit and generate output direction data; and an output unit provided with an array of tactile zones, the output unit operable to receive output direction data from the controller and generate a tactualiy-dctectable output pattern at the array to communicate directional information to the user.
The provision of a tactually detectable output pattern to communicate directional information to a user enables the personal guidance apparatus to provide effective navigation to users including those with visual impairment.
In one embodiment, the tactile zones within the array may be operable to selectively produce a tactuallydetectable output relative to one another.
The selectively produced tactually detectable output from each tactile zone relative to the other tactile zones enables th.e directional information being provided to the user to be accurate and detailed.
In a further embodiment, the output pattern is independently programmable by the user.
The ability for the user to independently programme the output pattern enables the personal guidance device to provide an output pattern which is tailored to provide the directional information in a manner personal to the user which enables effective and accurate navigation. This personalisation of the output pattern enables the preferences of the user to determine the best and most personally suitable and/or effective communication of the directional information for that specific user.
For example, there may be many variables of the taetually-detectable output pattern that can be specified to suit the preferences of a user such as, but not limited to, one or more of the following: * the combination of specific tactile zones stimulated within the array; * the intensity of the stimulation (and thus the intensity of the tactually detectable output pattern); * the frequency of the stimulation of the tactile zones -for example, the zones may he stimulated constantly during the communication of specific directional information to a user, or may alternatively the stimulation may be pulsed at a specifically selected frequency; and * on each item of directional information communicated, the length of the tactuallydetectabie output pattern delivered may be predetermined based on the preferences of the user.
In addition, it will be appreciated that any other parameter of the tactually-detectable output pattern, suitable for modification by a user may be predetermined by the user.
Furthermore, the ability of the user to predetermine the parameters relating to the delivery of the directional information may he restricted or enhanced either manually or automatically for the purposes of safety. For example, the level of stimulation may be restricted so that a user cannot select a stimulation level that is below the conventional detection level of a user durin.g practical day to day use.
Another example of userselected preferences is the provision of reassurance' tactually-detectable output patterns to confirm to a user that they are following the correct route, particularly where there are no changes in direction to noti.fy the user of In this case, a more nervous user may wish to have a reassurance' output pattern which confirms they are following the correct route delivered over relatively short intervals, for example, every ten seconds or so when no ot.her directional information is being delivered. A less nervous or less visually impaired user may be happy to have a predetermined reassurance' output pattern communicated to them at longer intervals, for example very thirty seconds or so in the absence of the delivery of other direction information via the output unit.
The tactually detectable output pattern may correspond to predefined directional information types. For example, a specific output pattern may indicate that the user should turn right whereas another different output pattern may indicate that a user should turn left or continue straight on.
The provision of an output pattern which corresponds to predefined directional infornrntion enables provision of complex directional information to be imparted to a user in simple, recognisable manner.
In an apparatus according to the present invention, at least one of the tactile zones may be operable to produce a vibratory signal.
Alternatively, a plurality of the tactile zones may he operabl.e to produce a vibratory signal.
The provision of a tactually detectable output which is a vibratory signal enables users with visual impairment to receive easily interpreted signals which correspond to directional information.
In a further embodiment, the apparatus may further comprise an n-axis accelerometer operable to transmit directional data andlor speed of travel data to the controller, Thus, the apparatus can detect directional change and, for example, change of pace of a user during use of the apparatus. This is important as i.t allows the apparatus to detect when the user is approaching a point of directional change on the route being followed and ensure that directional information is provided in
S
good time for the user to take the necessary directional change action to successfully stay on route.
Preferably, the apparatus stores the speed of travel data over specified intervals S during a journey by the user and, over time, determines an average speed of travel of a user. This may be related to specific terrain types as defined by stored mapping data or other data accessible to the controller. This allows the apparatus to learn' a more personally appropriate delivery time of directional information to a user in advance of change of direction action being required to be taken by that user. For example, a user with a reiativel.y fast pace of travel would be provided with a tactually-detectable output pattern alerting the user to a required change of direction when they were further from the point of required directional change than another user with a slower pace of travel. Although such difference between pedestrian users may be relatively small, if the apparatus is being used, for example, on a mobility scooter or similar, this becomes quiet an important issue as the delivery of directional infonnation in good time before change of direction action is required is essential to ensure the user can successfully follow the route based on delivered instructions.
The tactile zones may produce pulsed signals. For example, the vibratory signal may be a pulsed vibratory signal.
The provision of a pulsed vibratory signal may provide users with a more easily recognised tactually detectable signal and also increase the number of output patterns available to a user.
In one embodiment, the output unit is remote from the input unit.
Such an arrangement would enable a user to carry the output unit in discreet contact \vith their body, for example in their hand, \vllilst carrying the input unit in a suitable receptacle such as, but not limited to a bag, pocket or case.
It will be understood that throughout the description the term tactual is intended to include the conventionally accepted meaning of arising from touch' and due to touch'. In addition, throughout the description, the term tactile is intended to include the conventionally accepted meaning of capable of being touched' and allows for being touched'.
jçfescrition of the Drawig An embodiment of the invention will now be described, by way of example only, and with reference to the accompanying drawings, in which: Figure 1 illustrates a first embodiment of an input unit of a personal guidance apparatus in accordance with the present invention; and Figure 2 illustrates a first embodiment of an output unit of the personal guidance apparatus of Figure 1.
Detailedçgition of the Invention Figures 1 and 2 show a first embodiment of a personal guidance apparatus 10 in accordance with the present invention. With reference to Figure 1, the personal guidance apparatus 10 comprises an i.nput unit I 8a which includes a microprocessor 34, Flash ROM 36 connected to the microprocessor 34 via memory controller 38 and RAM 40 connected to the microprocessor 34 via memory controller 42, The input unit 1 8a fUrther includes a wireless module, in this case Bluetooth module 44 which is in bidirectional communication with the microprocessor 34 and a digital compass 46, which is also connected to the microprocessor 34. The input unit 1 8a is also provided with a connection unit 48, in this case a serial bus connection, a power unit 50, an input module 52 which provides input to microprocessor 34; output module 54 which is operable to receive data output from microprocessor 34; motion detection module 56; navigation module 58, in this case a satellite navigation module and plug in module 59, each of which are operable to provide data input to the microprocessor 34, In this embodiment, the input module 52 comprises an input pad which is integrally formed within the input unit 1 8a.
With reference to Figure 2, there is shown an output unit 1 8b of personal guidance apparatus 10 which, in use, is in communication with the input unit 1 8a of Figure I. The output unit 1 8b comprises a microprocessor 64, a connection unit 68, in this case a serial bus connection; a power unit 70 and a wireless module, in this case Bluetooth module 74 which is in communication with wireless module 44 of input unit 1 8a. The microprocessor 64 is also in communication with an array of tactile elements, in this case three tactile elements 78a, b, c, via respective driver elements 76a, b and c. The tactile elements 78ac are, in this example, piezoelcctric motors however it will be clearly understood that any suitable elements which generate a tactually detectable output pattern may be used including, but not limited to pager motors, vibratable membranes, buzzers or the like. The piezoelectric motors 78ac in this example, when activated, generate a pulsed vibration which acts as a tactually detectable output.
In this embodiment, the Fl.ash ROM 36 of the input unit 18a of navigation unit 18 is provi.ded with digital mappin.g information which enables a journey start point, a journey end point and a route between the two to be selected. The selection is, in this case, carried out using input pad 52. The input unit 18a determines the route which can be used for a safe pedestrian navigation and receives data from navigation module 58 determining the current position of the input unit 1 8a and provides this to the microprocessor 34. The microprocessor 34 also determines which direction the input unit I 8a is pointing by means of the data received from the digital compass 46. The microprocessor 34 establishes the location pointed to from the calculated route, and then determines the bearing to a next calculated location and in turn determines the direction in which the user must travel. The microprocessor 34 generates an instruction signal contaiuing data representing the direction of travel information and determining which tactile element 78 a-c, in tactile unit 1 8b, will need to be activated to generate a tactually-detectable output pattern which reflects the required direction of travel. The instruction signal is coded into a suitable transmission protocol, which in this case is Bluetooth and is then provided to the wireless module 44 for transmission to tactile unit I 8b.
The tactile unit 18b receives the instruction signal via wireless module 74 and provides this to microprocessor 64 which generates an output signal corresponding to the selected tactile element to be activated appropriate signals to activate tactile elements 78a-c. The signal, in this embodiment, is provided to a suitable driver which drives the inten.ded element to produce a tactually-detectable output pattern to the user to indicate the direction in which they should move.
If, as the user starts to travel, the angle between the direction from the current point and the next point, and the direction from the current point to the expected point is greater than a certain predetermined critical value, the input unit 1 8a generates a tactually-detectable correction signal which is provided to the user by tactile unit 1 8b. If, as the user starts to travel, the angle between the direction from the current point and the next point, an.d the direction from the current poi.n.t to the expected point is within a predetermined acceptable range, the reception unit 18a generates a tactually-detectabie output pattern which represents a reassurance signal which is provided, by way of tactile unit I 8b, to the user that the correct navigation direction is being taken.
The instruction signal generated by input unit 1 8a sviii be appropriate to activate tactile elements 78a-c to generate a tactually-detectable output pattern which reflects the directional data to be imparted to the user. The tactually detectable output pattern acts to guide the user of the personal guidance apparatus 10 in the required direction as determined by the microprocessor 34. In the tactile unit I 8b of this embodiment, the requirement for the user to take a right turn is indicated to the user by the generation of a tactually detectable output in the form of a vibration of tactile element 78c in response to an instruction signal generated by the input unit 18a.
In one embodiment of tactile unit 1 8h, the microprocessor 64 can incorporate feedback circuitry (not shown) which determines whether to check that the correct tactile element 78a-c as been activated. The microprocessor uses the data from the feedback circuitry to generate a feedback signal which is provided to wireless module 74 for transmission to input unit 1 8a which can then determine if the appropriate one of tactile elements 78ac was successfully or unsuccessfully activated.
The tactually detectable output generated by tactile elements 78ac can be varied in intensity as required by the sensitivity of the user. In one embodiment of the personal guidance unit 10 the tactile elements are configurable to the preferences of the user to ensure the utility of the personal guidance unit 10 is maximised.
In a further embodiment, the output pattern provided by the tactile elements 76ac can be independently programmed by the user thus enabling the personal guidance devi.ce to provide an output pattern which is tailored to provide the directional information in a manner personal to the user which enables effective and accurate navigation. in a yet further embodiment, the tactually detectable output pattern can be activated such that a predetermined specific output corresponds to predefined directional information types. The provision of an output pattern which corresponds to predefined directional information enables provision of complex directional information to be imparted to a user in simple, recognisabie manner.
In use, the personal guidance unit 10 will, by the provision of a tactually detectable output pattern, facilitate th.e navigation of a user such as a person with visual impairment. The tactile elemen.ts 76a-c must he in contact wi.th the user which requires the tactile unit 1 8b to he carried or worn against the users skin, for example in their hand. The remainder of the personal guidance unit 10, such as the input unit 18a can be kept in a pocket of a jacket which the user is wearing, or can be carried by the user in a bag or backpack. The tactually detectable output generated by the tactile elements 76a-c provides unequivocal directionality by giving an indication of the direction of motion required for a location to be reached.
In use, the tactile elements 76a-c can be presented to the user in a variety of ways.
For example, the tactile elements 76a-c may be incorporated in a small handheld device such as a mobile phone. However, the tactile elements 76a-c can alternatively be incorporated in a ergonomic unit such as a glove, in which the finger tip pads will each be in direct contact with one of the tactile elements 76-c, alternatively, the tactile elements could be arranged in the glove to come into contact with any desired part of the hand, In a further embodiment two such glove ergonomic units may be employed within the personal guidance apparatus which allows the tactually detectable output pattern to be delivered to a preferred glove to indicate specific direction information, such as delivery of the tactually detectable pattern. For example, a specific tactually detectable pattern may be delivered to the right hand glove when indicating an immediate right turn should be taken.
The tactile elements could be provided i.n a mouldable unit which can be applied to an aid which the user carries, such as a walking stick, so that the elements will be in contact with the user's hand. For example, the mouldable element could be applied to the top of the walking stick handle where it will become shaped to correspond to the user's hand, for example by heat or pressure deformation, to provide a high level of comfort whilst also enabling the user to take advantage of the navigation aid. Similarly, the mouldable tactile element could be applied to the handle or arm of a wheelchair such that it conforms to a users hand and is comfortable whilst the tactile elements are in contact wi.th the users hand to facilitate navigation.
Although aspects of the invention have been described with reference to the embodiment shown in the accompanying drawings, it is to be understood that the invention is not limited to the precise embodiment shown and that various changes and modifications may be effected without further inventive skill and effort, for example, it will be appreciated that the present invention may also provide advantages to hearing-impaired users who cannot usefully interact with a guidance device that provides audible signals only, or to literary-disadvantaged individuals who cannot usefully interact with visual text-based input or output data requirements. In addition it will be clearly understood that whilst the microprocessor 34 of the input unit 1 8a has been described as producing a suitable instruction signal coded in to a suitable transmission protocol for wireless module 44, the instruction signal could be coded to be transmitted through for example, a wireless transceiver, such as, for example, Bluetooth or Zigbee module or any other suitable module utilising a suitable communication protocol.
The microprocessor 34 has also been detailed as determining which direction the input unit 1 8a is pointing by means of the data received from the digital compass 46. However, it will be un.derstood that the input unit 1. 8a may also include an n-axis accelerometer (not shown) and data from the accelerometer may also be used in helping to determine the di.rection in which the input unit I 8a i.s facing and determining changes in pace, resulting in changes in the output signal.
The signal from microprocessor 64 has been described as being sent to a suitable driver which drives the tactile element, However, it will be understood that the signal can, in an alternative embodiment, be sent directly to a motor or equivalent tactile element which will then generate a tactually detectable output pattern. In the embodiments detailed above the tactile elements 76a-c generate a pulsed vibration which acts as a tactually detectable output however the tactually detectable output may be vibration without pulsing or may be any other suitable physical stimulus which provides a user with a signal indicative of directional information.
The tactile unit 1 8b has been described above as being provided with an array of three tactile elements 78a, b and c however it will be clearly understood that any suitable number of tactile elements can be used to provide the tactually detectable output pattern. It will also be understood that the tactile unit may be provided as an array of tactile elements whereby the tactile element array may optionally be segregated into zones. The navigation aid can communicate with the tactile unit when it is in contact with the hand of a user and the user will be able to select, by pre-programming, the tactile elements or zones which become activated when specific directional instructions are output during navigation, It will be further understood that whilst a tactile unit has been described as being in contact with a users hand, the tactile unit could be put in contact with any chosen body part on a personal preference basis, for example, it may be in contact with a forearm or a foot.
It will be clearly understood that whilst the personal guidance apparatus 10 has been described in the above embodiment as comprising a input uni.t 1 8a and a tactile unit 1 8b which are in. communication wirelessly, the i.nput unit 1 8a and tactile unit 1 8b could he in communication. by means of a cable (not shown) which will necessitate the use of a communication protocol suitable for wired communication, such as for example, a serial link. It will be further understood that whilst the input unit I 8a and tactile unit 1 8b have been described as separate units, they could be incorporated within a single body (not shown). In addition, if the input unit and tactile unit are combined in a single body, the function of microprocessor 34 an.d microprocessor 64 can be implemented by a singl.e microprocessor (not shown) which carries out the signal reception and position establishment as well as driving the tactile elements 76a-c either directly or indirectly through the drivers 78ac.
Claims (9)
- CLAIMS1. Personal guidance apparatus for directing a user from a first location to a predefined destination, the apparatus comprising: an input unit operable to receive first location data and input destination data; a controller in communication with the input unit and operable to receive first location data and input destination data from the input unit and generate output direction data; and an output unit provided with an array of tactile zones, the output unit operable to receive output direction data from the controller and generate a tactually-detectable output pattern at the array to communicate directional information to the user.
- 2. Personal guidance apparatus as claimed in claim 1, wherein the tactile zones within the array are operable to selectively produce a tactually-detectable output relative to one another.
- 3. Personal guidance apparatus as claimed in any preceding claim, wherein the output pattern is independently programmable by the user.
- 4. Personal guidance apparatus as claimed in any preceding claim wherein the output pattern corresponds to predefined directional information types.
- 5. Personal guidance apparatus as claimed in any preceding claim, wherein at least one of the tactile zones is operable to produce a vibratory signal.
- 6. Personal guidance apparatus as claimed in claim 5, wherein all of the tactile zones are operable to produce a vibratory signal.
- 7. Personal guidance apparatus as claimed in claim 5 or claim 6, wherein the vibratory signal is a pulsed vibratory signal.
- 8. Personal guidance apparatus as claimed in any preceding claim, wherein the output unit is remote from the input unit.
- 9. Personal guidance apparatus substantially as hereinbefore described and with reference to the accompanying drawings.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
GB201011874A GB2489906A (en) | 2010-07-15 | 2010-07-15 | Tactile navigation unit for the visually impaired |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
GB201011874A GB2489906A (en) | 2010-07-15 | 2010-07-15 | Tactile navigation unit for the visually impaired |
Publications (2)
Publication Number | Publication Date |
---|---|
GB201011874D0 GB201011874D0 (en) | 2010-09-01 |
GB2489906A true GB2489906A (en) | 2012-10-17 |
Family
ID=42734957
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
GB201011874A Withdrawn GB2489906A (en) | 2010-07-15 | 2010-07-15 | Tactile navigation unit for the visually impaired |
Country Status (1)
Country | Link |
---|---|
GB (1) | GB2489906A (en) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2014170539A1 (en) * | 2013-04-17 | 2014-10-23 | Nokia Corporation | Haptic device for pedestrian navigation |
US11079249B1 (en) | 2020-12-14 | 2021-08-03 | International Business Machines Corporation | Haptic navigation device |
Families Citing this family (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110068354B (en) * | 2019-05-30 | 2023-12-29 | 漳州卫生职业学院 | Navigation reminding device and navigation method for blind person |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
GB2287535A (en) * | 1994-03-17 | 1995-09-20 | Univ Surrey | Personal navigation system |
US6502032B1 (en) * | 2001-06-25 | 2002-12-31 | The United States Of America As Represented By The Secretary Of The Air Force | GPS urban navigation system for the blind |
JP2007248478A (en) * | 2007-06-21 | 2007-09-27 | National Institute Of Advanced Industrial & Technology | Haptic type human navigation system |
US20080120029A1 (en) * | 2006-02-16 | 2008-05-22 | Zelek John S | Wearable tactile navigation system |
US7706212B1 (en) * | 2007-01-30 | 2010-04-27 | Campbell Terry L | Mobility director device and cane for the visually impaired |
-
2010
- 2010-07-15 GB GB201011874A patent/GB2489906A/en not_active Withdrawn
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
GB2287535A (en) * | 1994-03-17 | 1995-09-20 | Univ Surrey | Personal navigation system |
US6502032B1 (en) * | 2001-06-25 | 2002-12-31 | The United States Of America As Represented By The Secretary Of The Air Force | GPS urban navigation system for the blind |
US20080120029A1 (en) * | 2006-02-16 | 2008-05-22 | Zelek John S | Wearable tactile navigation system |
US7706212B1 (en) * | 2007-01-30 | 2010-04-27 | Campbell Terry L | Mobility director device and cane for the visually impaired |
JP2007248478A (en) * | 2007-06-21 | 2007-09-27 | National Institute Of Advanced Industrial & Technology | Haptic type human navigation system |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2014170539A1 (en) * | 2013-04-17 | 2014-10-23 | Nokia Corporation | Haptic device for pedestrian navigation |
US10466787B2 (en) | 2013-04-17 | 2019-11-05 | Provenance Asset Group Llc | Haptic device for pedestrian navigation |
US11079249B1 (en) | 2020-12-14 | 2021-08-03 | International Business Machines Corporation | Haptic navigation device |
Also Published As
Publication number | Publication date |
---|---|
GB201011874D0 (en) | 2010-09-01 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP6514222B2 (en) | Method and system for transmitting haptic instructions to the human body | |
US10371544B2 (en) | Vibrating haptic device for the blind | |
CN105748265B (en) | Navigation device and method | |
RU2708973C2 (en) | Navigation devices and methods of navigation | |
US8629766B2 (en) | Portable information display apparatus | |
US20080120029A1 (en) | Wearable tactile navigation system | |
US20140077945A1 (en) | Biological information notifying apparatus, biological information notifying method, and computer-readable storage medium having biological information notifying program stored thereon | |
JP4612928B2 (en) | Body motion sensing device | |
EP1625841A4 (en) | Device and method of applying skin sensory stimulation | |
KR101715472B1 (en) | Smart walking assistance device for the blind and Smart walking assistance system using the same | |
EP3197309A1 (en) | Signaling sole for shoes, shoe provided with said sole and kit comprising at least one of said sole | |
GB2489906A (en) | Tactile navigation unit for the visually impaired | |
KR20180013353A (en) | Wearable information providing device for blind | |
US11684537B2 (en) | Human-interface device and a guiding apparatus for a visually impaired user including such human-interface device | |
WO2017153708A1 (en) | Cyclist accessory system | |
EP2084494A1 (en) | Method and apparatus for communicating navigational instructions to a user | |
KR20060073111A (en) | Guide equipment for the sight disabled | |
KR20160098566A (en) | Information providing system for changing features for a body parts by using wearable device and the method thereof | |
KR102097795B1 (en) | The smart braille cane for blind person | |
JP2015219361A (en) | Glove for information transmission and information transmission system | |
KR101706770B1 (en) | Smart glove for bicycle | |
KR102238029B1 (en) | Tactile information supply devide | |
KR20180033484A (en) | Wearable information providing device for blind | |
KR101628599B1 (en) | Multi function smart fashion belt | |
KR20170035742A (en) | The Electric Walking Assistive Device |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AT | Applications terminated before publication under section 16(1) | ||
S20A | Reinstatement of application (sect. 20a/patents act 1977) |
Free format text: REQUEST FOR REINSTATEMENT FILED Effective date: 20120621 Free format text: REQUEST FOR REINSTATEMENT ALLOWED Effective date: 20120704 |
|
WAP | Application withdrawn, taken to be withdrawn or refused ** after publication under section 16(1) |