GB2489906A - Tactile navigation unit for the visually impaired - Google Patents

Tactile navigation unit for the visually impaired Download PDF

Info

Publication number
GB2489906A
GB2489906A GB201011874A GB201011874A GB2489906A GB 2489906 A GB2489906 A GB 2489906A GB 201011874 A GB201011874 A GB 201011874A GB 201011874 A GB201011874 A GB 201011874A GB 2489906 A GB2489906 A GB 2489906A
Authority
GB
United Kingdom
Prior art keywords
user
tactile
output
unit
data
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Withdrawn
Application number
GB201011874A
Other versions
GB201011874D0 (en
Inventor
Trevor Ajan Reginald
Michael Kiernan
Daniele Doneddu
Jason Perkins
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
PEEPO GPS Ltd
Original Assignee
PEEPO GPS Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by PEEPO GPS Ltd filed Critical PEEPO GPS Ltd
Priority to GB201011874A priority Critical patent/GB2489906A/en
Publication of GB201011874D0 publication Critical patent/GB201011874D0/en
Publication of GB2489906A publication Critical patent/GB2489906A/en
Withdrawn legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G01MEASURING; TESTING
    • G01CMEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
    • G01C21/00Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00
    • G01C21/26Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00 specially adapted for navigation in a road network
    • G01C21/34Route searching; Route guidance
    • G01C21/36Input/output arrangements for on-board computers
    • G01C21/3626Details of the output of route guidance instructions
    • G01C21/3652Guidance using non-audiovisual output, e.g. tactile, haptic or electric stimuli
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01CMEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
    • G01C21/00Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00
    • G01C21/20Instruments for performing navigational calculations
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09BEDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
    • G09B21/00Teaching, or communicating with, the blind, deaf or mute
    • G09B21/001Teaching or communicating with blind persons
    • G09B21/003Teaching or communicating with blind persons using tactile presentation of the information, e.g. Braille displays

Abstract

Personal guidance apparatus for directing a visually impaired or blind user from a first location to a predefined destination, the apparatus comprising an input unit operable to receive first location data and input destination data; a controller in communication with the input unit and operable to receive first location data and input destination data from the input unit and generate output direction data; and an output unit provided with an array of tactile zones 78a, 78b, 78c, the output unit operable to receive output direction data from the controller and generate a tactually-detectable output pattern at the array to communicate directional information to the user. The tactile zones can produce vibratory signals which can be pulsed. The guidance unit also allows the user to alter the output signals, such as changing the intensity of the stimulation, to suit their individual needs. The tactile elements of the device could be attached to the handle of a walking stick or the handle of a wheelchair.

Description

TACTILE FEEDBACK UNIT
çgpundtotheipention The present invention relates to a guidance device incorporating a tactile interface, and preferably such a guidance device for use by a visually impaired user.
There are several conventional guidance devices on the market intended to assist with travelling to a destination, Most of these rely on satellite navigation and are intended for use when navigating a vehicle. Although personal guidance systems exist, these are often bulky and consist of multiple units, so are inconvenient for carrying, particularly by pedestrians. Furthermore, such personal guidance systems, such as, for example, those downloadabie onto Mi mobile phones and the like, require the operation of a small user interface and visualise directional information on a small display screen. These systems are clearly unsuitable for the visually impaired.
There are a few systems avai.iabl.e for use by the visuall.y impaired to assist with guiding the user from a first location to a specified destination. However, these also have limitations. For example, the feedback systems to provide directional information iuay be too prescriptive and not suitable for all users. In addition, some systems may involve large or obtrusive units or include multiple units for location on different parts of the body with noticeable cables connecting each unit.
As visually impaired users are typically very independent, any systems that draw the attention of others to their impainnent are clearly undesirable.
There is therefore a need for a system which is unobtrusive and provides directional information in an accessible way which is suitable for use by all users.
gryfffiçJnvenition The present invention seeks to address the problems of the prior art.
A first aspect of the present invention relates to a personal guidance apparatus for directing a user from a first location to a predefincd destination, the apparatus comprising: an input unit operable to receive first location data and input destination data; a controller in communication with the input unit and operable to receive first location data and input destination data from the input unit and generate output direction data; and an output unit provided with an array of tactile zones, the output unit operable to receive output direction data from the controller and generate a tactualiy-dctectable output pattern at the array to communicate directional information to the user.
The provision of a tactually detectable output pattern to communicate directional information to a user enables the personal guidance apparatus to provide effective navigation to users including those with visual impairment.
In one embodiment, the tactile zones within the array may be operable to selectively produce a tactuallydetectable output relative to one another.
The selectively produced tactually detectable output from each tactile zone relative to the other tactile zones enables th.e directional information being provided to the user to be accurate and detailed.
In a further embodiment, the output pattern is independently programmable by the user.
The ability for the user to independently programme the output pattern enables the personal guidance device to provide an output pattern which is tailored to provide the directional information in a manner personal to the user which enables effective and accurate navigation. This personalisation of the output pattern enables the preferences of the user to determine the best and most personally suitable and/or effective communication of the directional information for that specific user.
For example, there may be many variables of the taetually-detectable output pattern that can be specified to suit the preferences of a user such as, but not limited to, one or more of the following: * the combination of specific tactile zones stimulated within the array; * the intensity of the stimulation (and thus the intensity of the tactually detectable output pattern); * the frequency of the stimulation of the tactile zones -for example, the zones may he stimulated constantly during the communication of specific directional information to a user, or may alternatively the stimulation may be pulsed at a specifically selected frequency; and * on each item of directional information communicated, the length of the tactuallydetectabie output pattern delivered may be predetermined based on the preferences of the user.
In addition, it will be appreciated that any other parameter of the tactually-detectable output pattern, suitable for modification by a user may be predetermined by the user.
Furthermore, the ability of the user to predetermine the parameters relating to the delivery of the directional information may he restricted or enhanced either manually or automatically for the purposes of safety. For example, the level of stimulation may be restricted so that a user cannot select a stimulation level that is below the conventional detection level of a user durin.g practical day to day use.
Another example of userselected preferences is the provision of reassurance' tactually-detectable output patterns to confirm to a user that they are following the correct route, particularly where there are no changes in direction to noti.fy the user of In this case, a more nervous user may wish to have a reassurance' output pattern which confirms they are following the correct route delivered over relatively short intervals, for example, every ten seconds or so when no ot.her directional information is being delivered. A less nervous or less visually impaired user may be happy to have a predetermined reassurance' output pattern communicated to them at longer intervals, for example very thirty seconds or so in the absence of the delivery of other direction information via the output unit.
The tactually detectable output pattern may correspond to predefined directional information types. For example, a specific output pattern may indicate that the user should turn right whereas another different output pattern may indicate that a user should turn left or continue straight on.
The provision of an output pattern which corresponds to predefined directional infornrntion enables provision of complex directional information to be imparted to a user in simple, recognisable manner.
In an apparatus according to the present invention, at least one of the tactile zones may be operable to produce a vibratory signal.
Alternatively, a plurality of the tactile zones may he operabl.e to produce a vibratory signal.
The provision of a tactually detectable output which is a vibratory signal enables users with visual impairment to receive easily interpreted signals which correspond to directional information.
In a further embodiment, the apparatus may further comprise an n-axis accelerometer operable to transmit directional data andlor speed of travel data to the controller, Thus, the apparatus can detect directional change and, for example, change of pace of a user during use of the apparatus. This is important as i.t allows the apparatus to detect when the user is approaching a point of directional change on the route being followed and ensure that directional information is provided in
S
good time for the user to take the necessary directional change action to successfully stay on route.
Preferably, the apparatus stores the speed of travel data over specified intervals S during a journey by the user and, over time, determines an average speed of travel of a user. This may be related to specific terrain types as defined by stored mapping data or other data accessible to the controller. This allows the apparatus to learn' a more personally appropriate delivery time of directional information to a user in advance of change of direction action being required to be taken by that user. For example, a user with a reiativel.y fast pace of travel would be provided with a tactually-detectable output pattern alerting the user to a required change of direction when they were further from the point of required directional change than another user with a slower pace of travel. Although such difference between pedestrian users may be relatively small, if the apparatus is being used, for example, on a mobility scooter or similar, this becomes quiet an important issue as the delivery of directional infonnation in good time before change of direction action is required is essential to ensure the user can successfully follow the route based on delivered instructions.
The tactile zones may produce pulsed signals. For example, the vibratory signal may be a pulsed vibratory signal.
The provision of a pulsed vibratory signal may provide users with a more easily recognised tactually detectable signal and also increase the number of output patterns available to a user.
In one embodiment, the output unit is remote from the input unit.
Such an arrangement would enable a user to carry the output unit in discreet contact \vith their body, for example in their hand, \vllilst carrying the input unit in a suitable receptacle such as, but not limited to a bag, pocket or case.
It will be understood that throughout the description the term tactual is intended to include the conventionally accepted meaning of arising from touch' and due to touch'. In addition, throughout the description, the term tactile is intended to include the conventionally accepted meaning of capable of being touched' and allows for being touched'.
jçfescrition of the Drawig An embodiment of the invention will now be described, by way of example only, and with reference to the accompanying drawings, in which: Figure 1 illustrates a first embodiment of an input unit of a personal guidance apparatus in accordance with the present invention; and Figure 2 illustrates a first embodiment of an output unit of the personal guidance apparatus of Figure 1.
Detailedçgition of the Invention Figures 1 and 2 show a first embodiment of a personal guidance apparatus 10 in accordance with the present invention. With reference to Figure 1, the personal guidance apparatus 10 comprises an i.nput unit I 8a which includes a microprocessor 34, Flash ROM 36 connected to the microprocessor 34 via memory controller 38 and RAM 40 connected to the microprocessor 34 via memory controller 42, The input unit 1 8a fUrther includes a wireless module, in this case Bluetooth module 44 which is in bidirectional communication with the microprocessor 34 and a digital compass 46, which is also connected to the microprocessor 34. The input unit 1 8a is also provided with a connection unit 48, in this case a serial bus connection, a power unit 50, an input module 52 which provides input to microprocessor 34; output module 54 which is operable to receive data output from microprocessor 34; motion detection module 56; navigation module 58, in this case a satellite navigation module and plug in module 59, each of which are operable to provide data input to the microprocessor 34, In this embodiment, the input module 52 comprises an input pad which is integrally formed within the input unit 1 8a.
With reference to Figure 2, there is shown an output unit 1 8b of personal guidance apparatus 10 which, in use, is in communication with the input unit 1 8a of Figure I. The output unit 1 8b comprises a microprocessor 64, a connection unit 68, in this case a serial bus connection; a power unit 70 and a wireless module, in this case Bluetooth module 74 which is in communication with wireless module 44 of input unit 1 8a. The microprocessor 64 is also in communication with an array of tactile elements, in this case three tactile elements 78a, b, c, via respective driver elements 76a, b and c. The tactile elements 78ac are, in this example, piezoelcctric motors however it will be clearly understood that any suitable elements which generate a tactually detectable output pattern may be used including, but not limited to pager motors, vibratable membranes, buzzers or the like. The piezoelectric motors 78ac in this example, when activated, generate a pulsed vibration which acts as a tactually detectable output.
In this embodiment, the Fl.ash ROM 36 of the input unit 18a of navigation unit 18 is provi.ded with digital mappin.g information which enables a journey start point, a journey end point and a route between the two to be selected. The selection is, in this case, carried out using input pad 52. The input unit 18a determines the route which can be used for a safe pedestrian navigation and receives data from navigation module 58 determining the current position of the input unit 1 8a and provides this to the microprocessor 34. The microprocessor 34 also determines which direction the input unit I 8a is pointing by means of the data received from the digital compass 46. The microprocessor 34 establishes the location pointed to from the calculated route, and then determines the bearing to a next calculated location and in turn determines the direction in which the user must travel. The microprocessor 34 generates an instruction signal contaiuing data representing the direction of travel information and determining which tactile element 78 a-c, in tactile unit 1 8b, will need to be activated to generate a tactually-detectable output pattern which reflects the required direction of travel. The instruction signal is coded into a suitable transmission protocol, which in this case is Bluetooth and is then provided to the wireless module 44 for transmission to tactile unit I 8b.
The tactile unit 18b receives the instruction signal via wireless module 74 and provides this to microprocessor 64 which generates an output signal corresponding to the selected tactile element to be activated appropriate signals to activate tactile elements 78a-c. The signal, in this embodiment, is provided to a suitable driver which drives the inten.ded element to produce a tactually-detectable output pattern to the user to indicate the direction in which they should move.
If, as the user starts to travel, the angle between the direction from the current point and the next point, and the direction from the current point to the expected point is greater than a certain predetermined critical value, the input unit 1 8a generates a tactually-detectable correction signal which is provided to the user by tactile unit 1 8b. If, as the user starts to travel, the angle between the direction from the current point and the next point, an.d the direction from the current poi.n.t to the expected point is within a predetermined acceptable range, the reception unit 18a generates a tactually-detectabie output pattern which represents a reassurance signal which is provided, by way of tactile unit I 8b, to the user that the correct navigation direction is being taken.
The instruction signal generated by input unit 1 8a sviii be appropriate to activate tactile elements 78a-c to generate a tactually-detectable output pattern which reflects the directional data to be imparted to the user. The tactually detectable output pattern acts to guide the user of the personal guidance apparatus 10 in the required direction as determined by the microprocessor 34. In the tactile unit I 8b of this embodiment, the requirement for the user to take a right turn is indicated to the user by the generation of a tactually detectable output in the form of a vibration of tactile element 78c in response to an instruction signal generated by the input unit 18a.
In one embodiment of tactile unit 1 8h, the microprocessor 64 can incorporate feedback circuitry (not shown) which determines whether to check that the correct tactile element 78a-c as been activated. The microprocessor uses the data from the feedback circuitry to generate a feedback signal which is provided to wireless module 74 for transmission to input unit 1 8a which can then determine if the appropriate one of tactile elements 78ac was successfully or unsuccessfully activated.
The tactually detectable output generated by tactile elements 78ac can be varied in intensity as required by the sensitivity of the user. In one embodiment of the personal guidance unit 10 the tactile elements are configurable to the preferences of the user to ensure the utility of the personal guidance unit 10 is maximised.
In a further embodiment, the output pattern provided by the tactile elements 76ac can be independently programmed by the user thus enabling the personal guidance devi.ce to provide an output pattern which is tailored to provide the directional information in a manner personal to the user which enables effective and accurate navigation. in a yet further embodiment, the tactually detectable output pattern can be activated such that a predetermined specific output corresponds to predefined directional information types. The provision of an output pattern which corresponds to predefined directional information enables provision of complex directional information to be imparted to a user in simple, recognisabie manner.
In use, the personal guidance unit 10 will, by the provision of a tactually detectable output pattern, facilitate th.e navigation of a user such as a person with visual impairment. The tactile elemen.ts 76a-c must he in contact wi.th the user which requires the tactile unit 1 8b to he carried or worn against the users skin, for example in their hand. The remainder of the personal guidance unit 10, such as the input unit 18a can be kept in a pocket of a jacket which the user is wearing, or can be carried by the user in a bag or backpack. The tactually detectable output generated by the tactile elements 76a-c provides unequivocal directionality by giving an indication of the direction of motion required for a location to be reached.
In use, the tactile elements 76a-c can be presented to the user in a variety of ways.
For example, the tactile elements 76a-c may be incorporated in a small handheld device such as a mobile phone. However, the tactile elements 76a-c can alternatively be incorporated in a ergonomic unit such as a glove, in which the finger tip pads will each be in direct contact with one of the tactile elements 76-c, alternatively, the tactile elements could be arranged in the glove to come into contact with any desired part of the hand, In a further embodiment two such glove ergonomic units may be employed within the personal guidance apparatus which allows the tactually detectable output pattern to be delivered to a preferred glove to indicate specific direction information, such as delivery of the tactually detectable pattern. For example, a specific tactually detectable pattern may be delivered to the right hand glove when indicating an immediate right turn should be taken.
The tactile elements could be provided i.n a mouldable unit which can be applied to an aid which the user carries, such as a walking stick, so that the elements will be in contact with the user's hand. For example, the mouldable element could be applied to the top of the walking stick handle where it will become shaped to correspond to the user's hand, for example by heat or pressure deformation, to provide a high level of comfort whilst also enabling the user to take advantage of the navigation aid. Similarly, the mouldable tactile element could be applied to the handle or arm of a wheelchair such that it conforms to a users hand and is comfortable whilst the tactile elements are in contact wi.th the users hand to facilitate navigation.
Although aspects of the invention have been described with reference to the embodiment shown in the accompanying drawings, it is to be understood that the invention is not limited to the precise embodiment shown and that various changes and modifications may be effected without further inventive skill and effort, for example, it will be appreciated that the present invention may also provide advantages to hearing-impaired users who cannot usefully interact with a guidance device that provides audible signals only, or to literary-disadvantaged individuals who cannot usefully interact with visual text-based input or output data requirements. In addition it will be clearly understood that whilst the microprocessor 34 of the input unit 1 8a has been described as producing a suitable instruction signal coded in to a suitable transmission protocol for wireless module 44, the instruction signal could be coded to be transmitted through for example, a wireless transceiver, such as, for example, Bluetooth or Zigbee module or any other suitable module utilising a suitable communication protocol.
The microprocessor 34 has also been detailed as determining which direction the input unit 1 8a is pointing by means of the data received from the digital compass 46. However, it will be un.derstood that the input unit 1. 8a may also include an n-axis accelerometer (not shown) and data from the accelerometer may also be used in helping to determine the di.rection in which the input unit I 8a i.s facing and determining changes in pace, resulting in changes in the output signal.
The signal from microprocessor 64 has been described as being sent to a suitable driver which drives the tactile element, However, it will be understood that the signal can, in an alternative embodiment, be sent directly to a motor or equivalent tactile element which will then generate a tactually detectable output pattern. In the embodiments detailed above the tactile elements 76a-c generate a pulsed vibration which acts as a tactually detectable output however the tactually detectable output may be vibration without pulsing or may be any other suitable physical stimulus which provides a user with a signal indicative of directional information.
The tactile unit 1 8b has been described above as being provided with an array of three tactile elements 78a, b and c however it will be clearly understood that any suitable number of tactile elements can be used to provide the tactually detectable output pattern. It will also be understood that the tactile unit may be provided as an array of tactile elements whereby the tactile element array may optionally be segregated into zones. The navigation aid can communicate with the tactile unit when it is in contact with the hand of a user and the user will be able to select, by pre-programming, the tactile elements or zones which become activated when specific directional instructions are output during navigation, It will be further understood that whilst a tactile unit has been described as being in contact with a users hand, the tactile unit could be put in contact with any chosen body part on a personal preference basis, for example, it may be in contact with a forearm or a foot.
It will be clearly understood that whilst the personal guidance apparatus 10 has been described in the above embodiment as comprising a input uni.t 1 8a and a tactile unit 1 8b which are in. communication wirelessly, the i.nput unit 1 8a and tactile unit 1 8b could he in communication. by means of a cable (not shown) which will necessitate the use of a communication protocol suitable for wired communication, such as for example, a serial link. It will be further understood that whilst the input unit I 8a and tactile unit 1 8b have been described as separate units, they could be incorporated within a single body (not shown). In addition, if the input unit and tactile unit are combined in a single body, the function of microprocessor 34 an.d microprocessor 64 can be implemented by a singl.e microprocessor (not shown) which carries out the signal reception and position establishment as well as driving the tactile elements 76a-c either directly or indirectly through the drivers 78ac.

Claims (9)

  1. CLAIMS1. Personal guidance apparatus for directing a user from a first location to a predefined destination, the apparatus comprising: an input unit operable to receive first location data and input destination data; a controller in communication with the input unit and operable to receive first location data and input destination data from the input unit and generate output direction data; and an output unit provided with an array of tactile zones, the output unit operable to receive output direction data from the controller and generate a tactually-detectable output pattern at the array to communicate directional information to the user.
  2. 2. Personal guidance apparatus as claimed in claim 1, wherein the tactile zones within the array are operable to selectively produce a tactually-detectable output relative to one another.
  3. 3. Personal guidance apparatus as claimed in any preceding claim, wherein the output pattern is independently programmable by the user.
  4. 4. Personal guidance apparatus as claimed in any preceding claim wherein the output pattern corresponds to predefined directional information types.
  5. 5. Personal guidance apparatus as claimed in any preceding claim, wherein at least one of the tactile zones is operable to produce a vibratory signal.
  6. 6. Personal guidance apparatus as claimed in claim 5, wherein all of the tactile zones are operable to produce a vibratory signal.
  7. 7. Personal guidance apparatus as claimed in claim 5 or claim 6, wherein the vibratory signal is a pulsed vibratory signal.
  8. 8. Personal guidance apparatus as claimed in any preceding claim, wherein the output unit is remote from the input unit.
  9. 9. Personal guidance apparatus substantially as hereinbefore described and with reference to the accompanying drawings.
GB201011874A 2010-07-15 2010-07-15 Tactile navigation unit for the visually impaired Withdrawn GB2489906A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
GB201011874A GB2489906A (en) 2010-07-15 2010-07-15 Tactile navigation unit for the visually impaired

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
GB201011874A GB2489906A (en) 2010-07-15 2010-07-15 Tactile navigation unit for the visually impaired

Publications (2)

Publication Number Publication Date
GB201011874D0 GB201011874D0 (en) 2010-09-01
GB2489906A true GB2489906A (en) 2012-10-17

Family

ID=42734957

Family Applications (1)

Application Number Title Priority Date Filing Date
GB201011874A Withdrawn GB2489906A (en) 2010-07-15 2010-07-15 Tactile navigation unit for the visually impaired

Country Status (1)

Country Link
GB (1) GB2489906A (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2014170539A1 (en) * 2013-04-17 2014-10-23 Nokia Corporation Haptic device for pedestrian navigation
US11079249B1 (en) 2020-12-14 2021-08-03 International Business Machines Corporation Haptic navigation device

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110068354B (en) * 2019-05-30 2023-12-29 漳州卫生职业学院 Navigation reminding device and navigation method for blind person

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
GB2287535A (en) * 1994-03-17 1995-09-20 Univ Surrey Personal navigation system
US6502032B1 (en) * 2001-06-25 2002-12-31 The United States Of America As Represented By The Secretary Of The Air Force GPS urban navigation system for the blind
JP2007248478A (en) * 2007-06-21 2007-09-27 National Institute Of Advanced Industrial & Technology Haptic type human navigation system
US20080120029A1 (en) * 2006-02-16 2008-05-22 Zelek John S Wearable tactile navigation system
US7706212B1 (en) * 2007-01-30 2010-04-27 Campbell Terry L Mobility director device and cane for the visually impaired

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
GB2287535A (en) * 1994-03-17 1995-09-20 Univ Surrey Personal navigation system
US6502032B1 (en) * 2001-06-25 2002-12-31 The United States Of America As Represented By The Secretary Of The Air Force GPS urban navigation system for the blind
US20080120029A1 (en) * 2006-02-16 2008-05-22 Zelek John S Wearable tactile navigation system
US7706212B1 (en) * 2007-01-30 2010-04-27 Campbell Terry L Mobility director device and cane for the visually impaired
JP2007248478A (en) * 2007-06-21 2007-09-27 National Institute Of Advanced Industrial & Technology Haptic type human navigation system

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2014170539A1 (en) * 2013-04-17 2014-10-23 Nokia Corporation Haptic device for pedestrian navigation
US10466787B2 (en) 2013-04-17 2019-11-05 Provenance Asset Group Llc Haptic device for pedestrian navigation
US11079249B1 (en) 2020-12-14 2021-08-03 International Business Machines Corporation Haptic navigation device

Also Published As

Publication number Publication date
GB201011874D0 (en) 2010-09-01

Similar Documents

Publication Publication Date Title
JP6514222B2 (en) Method and system for transmitting haptic instructions to the human body
US10371544B2 (en) Vibrating haptic device for the blind
CN105748265B (en) Navigation device and method
RU2708973C2 (en) Navigation devices and methods of navigation
US8629766B2 (en) Portable information display apparatus
US20080120029A1 (en) Wearable tactile navigation system
US20140077945A1 (en) Biological information notifying apparatus, biological information notifying method, and computer-readable storage medium having biological information notifying program stored thereon
JP4612928B2 (en) Body motion sensing device
EP1625841A4 (en) Device and method of applying skin sensory stimulation
KR101715472B1 (en) Smart walking assistance device for the blind and Smart walking assistance system using the same
EP3197309A1 (en) Signaling sole for shoes, shoe provided with said sole and kit comprising at least one of said sole
GB2489906A (en) Tactile navigation unit for the visually impaired
KR20180013353A (en) Wearable information providing device for blind
US11684537B2 (en) Human-interface device and a guiding apparatus for a visually impaired user including such human-interface device
WO2017153708A1 (en) Cyclist accessory system
EP2084494A1 (en) Method and apparatus for communicating navigational instructions to a user
KR20060073111A (en) Guide equipment for the sight disabled
KR20160098566A (en) Information providing system for changing features for a body parts by using wearable device and the method thereof
KR102097795B1 (en) The smart braille cane for blind person
JP2015219361A (en) Glove for information transmission and information transmission system
KR101706770B1 (en) Smart glove for bicycle
KR102238029B1 (en) Tactile information supply devide
KR20180033484A (en) Wearable information providing device for blind
KR101628599B1 (en) Multi function smart fashion belt
KR20170035742A (en) The Electric Walking Assistive Device

Legal Events

Date Code Title Description
AT Applications terminated before publication under section 16(1)
S20A Reinstatement of application (sect. 20a/patents act 1977)

Free format text: REQUEST FOR REINSTATEMENT FILED

Effective date: 20120621

Free format text: REQUEST FOR REINSTATEMENT ALLOWED

Effective date: 20120704

WAP Application withdrawn, taken to be withdrawn or refused ** after publication under section 16(1)