CN109602585B - Blind guiding glasses and anti-collision early warning method thereof - Google Patents
Blind guiding glasses and anti-collision early warning method thereof Download PDFInfo
- Publication number
- CN109602585B CN109602585B CN201811451997.7A CN201811451997A CN109602585B CN 109602585 B CN109602585 B CN 109602585B CN 201811451997 A CN201811451997 A CN 201811451997A CN 109602585 B CN109602585 B CN 109602585B
- Authority
- CN
- China
- Prior art keywords
- blind
- image
- glasses
- obstacle
- frame
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Images
Classifications
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61H—PHYSICAL THERAPY APPARATUS, e.g. DEVICES FOR LOCATING OR STIMULATING REFLEX POINTS IN THE BODY; ARTIFICIAL RESPIRATION; MASSAGE; BATHING DEVICES FOR SPECIAL THERAPEUTIC OR HYGIENIC PURPOSES OR SPECIFIC PARTS OF THE BODY
- A61H3/00—Appliances for aiding patients or disabled persons to walk about
- A61H3/06—Walking aids for blind persons
- A61H3/061—Walking aids for blind persons with electronic detecting or guiding means
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/56—Extraction of image or video features relating to colour
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/50—Context or environment of the image
- G06V20/56—Context or environment of the image exterior to a vehicle by using sensors mounted on the vehicle
- G06V20/58—Recognition of moving objects or obstacles, e.g. vehicles or pedestrians; Recognition of traffic objects, e.g. traffic signs, traffic lights or roads
-
- G—PHYSICS
- G08—SIGNALLING
- G08B—SIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
- G08B21/00—Alarms responsive to a single specified undesired or abnormal condition and not otherwise provided for
- G08B21/18—Status alarms
- G08B21/24—Reminder alarms, e.g. anti-loss alarms
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61H—PHYSICAL THERAPY APPARATUS, e.g. DEVICES FOR LOCATING OR STIMULATING REFLEX POINTS IN THE BODY; ARTIFICIAL RESPIRATION; MASSAGE; BATHING DEVICES FOR SPECIAL THERAPEUTIC OR HYGIENIC PURPOSES OR SPECIFIC PARTS OF THE BODY
- A61H2201/00—Characteristics of apparatus not provided for in the preceding codes
- A61H2201/01—Constructive details
- A61H2201/0173—Means for preventing injuries
- A61H2201/0184—Means for preventing injuries by raising an alarm
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61H—PHYSICAL THERAPY APPARATUS, e.g. DEVICES FOR LOCATING OR STIMULATING REFLEX POINTS IN THE BODY; ARTIFICIAL RESPIRATION; MASSAGE; BATHING DEVICES FOR SPECIAL THERAPEUTIC OR HYGIENIC PURPOSES OR SPECIFIC PARTS OF THE BODY
- A61H2201/00—Characteristics of apparatus not provided for in the preceding codes
- A61H2201/16—Physical interface with patient
- A61H2201/1602—Physical interface with patient kind of interface, e.g. head rest, knee support or lumbar support
- A61H2201/165—Wearable interfaces
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Multimedia (AREA)
- Health & Medical Sciences (AREA)
- Rehabilitation Therapy (AREA)
- General Health & Medical Sciences (AREA)
- Public Health (AREA)
- Veterinary Medicine (AREA)
- Business, Economics & Management (AREA)
- Emergency Management (AREA)
- Animal Behavior & Ethology (AREA)
- Life Sciences & Earth Sciences (AREA)
- Physical Education & Sports Medicine (AREA)
- Pain & Pain Management (AREA)
- Epidemiology (AREA)
- Rehabilitation Tools (AREA)
- Image Analysis (AREA)
Abstract
The invention discloses a pair of blind guiding glasses, which comprises a glasses frame and lenses arranged on the glasses frame, wherein the front side of the glasses frame is respectively provided with a camera, a lighting lamp and a light sensor, the end part of one supporting leg of the glasses frame is provided with an earphone, the outer side of the end part of the other supporting leg of the glasses frame is provided with a control key, and the glasses also comprise a control box. The invention discloses an anti-collision early warning method of blind-guiding glasses, which is implemented according to the following steps; step 1, shooting a plurality of continuous images; step 2, preprocessing a plurality of continuous images; step 3, identifying the blind sidewalk, if the blind sidewalk cannot be identified, re-shooting the picture, returning to the step 1, and restarting; step 4, recognizing the obstacles in the blind road area, and if the obstacles cannot be recognized, re-shooting the picture and returning to the step 1 to restart; and 5, calculating the distance between the barrier and the blind-guiding glasses user, and giving an early warning to the blind-guiding glasses user. The blind guiding glasses can be carried about, are convenient to control, and can effectively remind a user of the distance between the user and an obstacle through an anti-collision early warning method.
Description
Technical Field
The invention belongs to the technical field of wearable equipment, and particularly relates to blind guiding glasses and an anti-collision early warning method thereof.
Background
The visual system is one of the most important systems for human perception of the objective world, and the human obtains more than 80% of the external information. Therefore, the lack or decline of visual information makes it difficult for visually impaired people to perceive the surrounding environment information, and the free travel is hindered, and the aspects of life self-care and the like are also limited, which brings much inconvenience and trouble to the daily life of the group and is a main factor causing the reduction of the quality of life. The world health organization predicts that the number of global vision-impaired people will increase to 7500 thousands of people in 2020, the sixth census of China and the second disabled people survey data statistics of China show that the number of the vision-impaired people in China reaches 877 thousands of people, and the world is the most country. And as the aging population of China continuously increases, the number of the visually impaired people also increases sharply, so that the problem of safe trip of the visually impaired people is solved, the life quality of the visually impaired people is improved, and the visually impaired people are attracted by social attention.
At present, the travel of the visually impaired people mainly depends on the blind road and the handheld blind stick, but the limitation of the blind road area and the space information amount provided by the blind stick can not enable the visually impaired people to safely travel. In addition, only a few visually-impaired people use the guide dog to lead a trip, but the guide dog training learning time is long, the cost is high, and the actual requirements of visually-impaired people are difficult to meet. The blind guiding robot integrates various sensors on the movable carrier, so that the blind guiding robot can sense environmental information and guide visually impaired people to safely go out. However, the blind guiding robot has low capability of adapting to the environment and limited application range. The blind guiding and walking aid device in the prior art is often provided with a movable carrier, or cannot be carried about due to large volume, or brings unnecessary trouble to the use of the visually impaired people due to troublesome operation and the like.
Disclosure of Invention
The invention aims to provide blind guiding glasses which can be carried about, are convenient to operate and control and effectively remind a user of the distance between the user and an obstacle.
The technical scheme adopted by the invention is as follows: a pair of blind guiding glasses comprises a glasses frame and lenses arranged on the glasses frame, wherein the front side of the glasses frame is respectively provided with a camera, a lighting lamp and a light sensor, the camera is positioned at the center of the front side of the glasses frame, the end part of one supporting leg of the glasses frame is provided with an earphone, the outer side of the end part of the other supporting leg of the glasses frame is provided with a control key,
also comprises a control box, a main controller, an image processor, a battery, a vibration module, a voice module, a Bluetooth communication module and a WiFi communication module which are connected with the main controller are arranged in the control box,
the camera is connected with an image processor in the control box through a video line, the control keys and the earphone are respectively connected with the main controller and the voice module in the control box through data lines, the illuminating lamp and the light sensor are both connected with the main controller in the control box through the data lines, and the control box controls the opening and closing of the illuminating lamp according to light intensity signals transmitted by the light sensor.
The present invention is also characterized in that,
different key areas are divided on the control keys, and braille characters are arranged on each key area, so that the operation of the blind is facilitated.
The spectacle frame is made of hard aluminum alloy materials, so that the spectacle frame is convenient to be attached to the face without falling off.
The second technical scheme adopted by the invention is that the anti-collision early warning method for the blind guiding glasses is implemented according to the following steps:
step 1, shooting to obtain a plurality of continuous images of the traveling direction of a user of the blind guiding glasses;
step 2, preprocessing a plurality of continuous images, increasing the image contrast and reducing the image noise;
step 4, identifying obstacles in the blind road area, determining static obstacles and dynamic obstacles, if the static obstacles and the dynamic obstacles cannot be determined, re-shooting the picture, returning to the step 1 to start, and otherwise, performing the step 5;
and 5, calculating the distance between the obstacle and the blind guiding glasses user according to the static obstacle and the dynamic obstacle determined in the step 4, and giving an early warning to the blind guiding glasses user.
The preprocessing process in the step 2 is to process a plurality of continuous images by adopting a histogram equalization and homomorphic filtering algorithm.
step 3.1, converting the RGB model of the preprocessed image into a Lab color space model by adopting a color space conversion model to obtain b-component characteristics of the blind road image in the Lab color space model, wherein the color space conversion model is as follows:
wherein X, Y, Z are the three stimulus values of the image, X0、Y0、Z0Three stimulus values, L, for CTE standard illumination*Representing the brightness of the image, a*、b*Representing image chromaticity;
step 3.2, adopting a gray level co-occurrence matrix, and selecting four direction angles of 0 degree, 45 degrees, 90 degrees and 135 degrees to perform statistics on the image to obtain four gray level co-occurrence matrices of P0 degrees, P45 degrees, P90 degrees and P135 degrees respectively;
step 3.3, combining the b-component characteristic of the blind road image obtained in the step 3.1 and the four gray level co-occurrence matrixes respectively of P0 degrees, P45 degrees, P90 degrees and P135 degrees obtained in the step 3.2, and segmenting the image by adopting a maximum inter-class variance method to preliminarily obtain a blind road region;
and 3.4, extracting blind channel edge characteristics from the blind channel region obtained primarily in the step 3.3 by adopting a Canny operator, and extracting the blind channel edge characteristics through Hough transformation to obtain a final complete blind channel region.
Step 4 is specifically implemented according to the following steps:
step 4.1, determining static obstacles
Step 4.1.1, neglecting the upper area of the blind road area in each image, and extracting the blind road area;
step 4.1.2, calculating the gray values of the blind road region obtained in the step 4.1.1 in the vertical direction and the horizontal direction, calculating the entropy of the blind road region obtained in the step 4.1.1 to judge whether a static obstacle exists in the blind road region, and judging whether the static obstacle exists when the gray values of the blind road region in the vertical direction and the horizontal direction are both smaller than the entropy of the blind road region;
step 4.2, determining dynamic obstacle
Step 4.2.1, detecting every three adjacent images by adopting a three-frame difference method, wherein the formula of the three-frame difference method is as follows
Wherein, Fk-1(x,y)、Fk(x,y)、Fk+1(x, y) is the k-1, k +1 frame image in succession, G1(x,y)、G2(x, y) is the frame-differenced image;
step 4.2.2, image G after frame difference1(x,y)、G2(x, y) selecting a threshold value T for binarization processing, wherein the binarization processing process comprises the following steps:
wherein T is a threshold value and D1(x,y)、D2(x, y) is the binarized image,
wherein the threshold T is based on the frame-corrupted image G1(x,y)、G2(x, y) selecting the distribution of gray values;
step 4.2.3, the binarized image D obtained in the step 4.2.21(x,y)、D2(x, y) performing a logical AND operation:
in the formula, R (x, y) is an image obtained after AND operation;
and 4.2.4, performing gray projection on the image obtained after the AND operation in the horizontal and vertical directions by adopting a gray projection algorithm to obtain two curves of the image obtained after the AND operation in the two directions, performing cross-correlation calculation on the two curves by using a cross-correlation function to obtain cross-correlation curves of the two curves, and determining the dynamic obstacle by using a peak point on the cross-correlation curves.
Step 5 is specifically implemented according to the following steps:
step 5.1, substituting the static obstacle and dynamic obstacle information obtained in the step 4 into a distance measurement formula to calculate the distance between the static obstacle and the dynamic obstacle and a user of the blind guiding glasses, wherein the distance measurement formula is as follows:
h is the installation height of the camera, H is the height of a static obstacle or a dynamic obstacle, f is the focal length of the camera, u, v, u0,v0,ax,ayThe internal parameters of the camera are taken as the parameters;
step 5.2, according to the distance between the static barrier and the dynamic barrier calculated in the step 5.1 and the blind guiding glasses user, giving an early warning to the blind guiding glasses user: when the distance is less than 3 meters, the blind guiding glasses user is in an unsafe area, the control box sends out a warning about the existence of the obstacle to the blind guiding glasses user through the voice module and the vibration module, when the distance is greater than 7 meters, the control box sends out a safe voice signal, and when the distance is greater than 3 meters and less than 7 meters, the system sends out a warning signal about the existence of the obstacle.
The invention has the beneficial effects that: the blind guiding glasses are simple in preparation process and low in cost, are suitable for large-scale production and popularization, can be carried about, are convenient to operate and control, can effectively remind a user of the distance between the user and an obstacle through an anti-collision early warning method, avoid the user from colliding with the obstacle on a blind road, and bring great convenience to the user in traveling.
Drawings
FIG. 1 is a view of the structure of the appearance of a pair of blind-guiding spectacles according to the present invention;
fig. 2 is a flow chart of an anti-collision early warning method of blind guide glasses according to the present invention.
The device comprises an earphone 1, an illuminating lamp 2, a camera 3, a lens 4, an eyeglass frame 5, a control box 6, a control button 7 and a light sensor 8.
Detailed Description
The following detailed description of the embodiments of the invention refers to the accompanying drawings and accompanying claims.
The invention provides a pair of blind guiding glasses, as shown in figure 1, comprising a glasses frame 5 and a lens 4 arranged on the glasses frame 5, wherein the front side of the glasses frame 5 is respectively provided with a camera 3, a lighting lamp 2 and a light sensor 8, the camera 3 is positioned at the center of the front side of the glasses frame 5, the end part of one supporting leg of the glasses frame 5 is provided with an earphone 1, the outer side of the end part of the other supporting leg of the glasses frame 5 is provided with a control key 7,
also comprises a control box 6, a main controller, an image processor, a battery, a vibration module, a voice module, a Bluetooth communication module and a WiFi communication module which are connected with the main controller are arranged in the control box 6,
Furthermore, different key areas are divided on the control keys 7, and braille characters are arranged on each key area, so that the operation of the blind is facilitated.
Furthermore, the spectacle frame 5 is made of hard aluminum alloy materials, so that the spectacle frame 5 is convenient to be attached to the face without falling off.
An anti-collision early warning method based on the blind-guiding glasses is specifically implemented according to the following steps as shown in fig. 2:
step 1, shooting to obtain a plurality of continuous images of the traveling direction of a user of the blind guiding glasses;
step 2, preprocessing a plurality of continuous images, increasing the image contrast and reducing the image noise;
step 4, identifying obstacles in the blind road area, determining static obstacles and dynamic obstacles, if the static obstacles and the dynamic obstacles cannot be determined, re-shooting the picture, returning to the step 1 to start, and otherwise, performing the step 5;
and 5, calculating the distance between the obstacle and the blind guiding glasses user according to the static obstacle and the dynamic obstacle determined in the step 4, and giving an early warning to the blind guiding glasses user.
Further, the preprocessing process in step 2 is to process a plurality of continuous images by using histogram equalization and homomorphic filtering algorithm.
step 3.1, converting the RGB model of the preprocessed image into a Lab color space model by adopting a color space conversion model to obtain b-component characteristics of the blind road image in the Lab color space model, wherein the color space conversion model is as follows:
wherein X, Y, Z are the three stimulus values of the image, X0、Y0、Z0Three stimulus values, L, for CTE standard illumination*Representing the brightness of the image, a*、b*Representing image chromaticity;
step 3.2, adopting a gray level co-occurrence matrix, and selecting four direction angles of 0 degree, 45 degrees, 90 degrees and 135 degrees to perform statistics on the image to obtain four gray level co-occurrence matrices of P0 degrees, P45 degrees, P90 degrees and P135 degrees respectively;
step 3.3, combining the b-component characteristic of the blind road image obtained in the step 3.1 and the four gray level co-occurrence matrixes respectively of P0 degrees, P45 degrees, P90 degrees and P135 degrees obtained in the step 3.2, and segmenting the image by adopting a maximum inter-class variance method to preliminarily obtain a blind road region;
and 3.4, extracting blind channel edge characteristics from the blind channel region obtained primarily in the step 3.3 by adopting a Canny operator, and extracting the blind channel edge characteristics through Hough transformation to obtain a final complete blind channel region.
Step 4 is specifically implemented according to the following steps:
step 4.1, determining static obstacles
Step 4.1.1, neglecting the upper area of the blind road area in each image, and extracting the blind road area;
step 4.1.2, calculating the gray values of the blind road region obtained in the step 4.1.1 in the vertical direction and the horizontal direction, calculating the entropy of the blind road region obtained in the step 4.1.1 to judge whether a static obstacle exists in the blind road region, and judging whether the static obstacle exists when the gray values of the blind road region in the vertical direction and the horizontal direction are both smaller than the entropy of the blind road region;
step 4.2, determining dynamic obstacle
Step 4.2.1, detecting every three adjacent images by adopting a three-frame difference method, wherein the formula of the three-frame difference method is as follows
G1(x,y)=|Fk(x,y)-Fk-1(x,y)| (2)
G2(x,y)=|Fk+1(x,y)-Fk(x,y)|
Wherein, Fk-1(x,y)、Fk(x,y)、Fk+1(x, y) is the k-1, k +1 frame image in succession, G1(x,y)、G2(x, y) is the frame-differenced image;
step 4.2.2, image G after frame difference1(x,y)、G2(x, y) selecting a threshold value T for binarization processing, wherein the binarization processing process comprises the following steps:
in the formula (I), the compound is shown in the specification,t is a threshold value, D1(x,y)、D2(x, y) is the binarized image,
wherein the threshold T is based on the frame-corrupted image G1(x,y)、G2(x, y) selecting the distribution of gray values;
step 4.2.3, the binarized image D obtained in the step 4.2.21(x,y)、D2(x, y) performing a logical AND operation:
in the formula, R (x, y) is an image obtained after AND operation;
and 4.2.4, performing gray projection on the image obtained after the AND operation in the horizontal and vertical directions by adopting a gray projection algorithm to obtain two curves of the image obtained after the AND operation in the two directions, performing cross-correlation calculation on the two curves by using a cross-correlation function to obtain cross-correlation curves of the two curves, and determining the dynamic obstacle by using a peak point on the cross-correlation curves.
Step 5 is specifically implemented according to the following steps:
step 5.1, substituting the static obstacle and dynamic obstacle information obtained in the step 4 into a distance measurement formula to calculate the distance between the static obstacle and the dynamic obstacle and a user of the blind guiding glasses, wherein the distance measurement formula is as follows:
h is the installation height of the camera, H is the height of a static obstacle or a dynamic obstacle, f is the focal length of the camera, u, v, u0,v0,ax,ayThe internal parameters of the camera are taken as the parameters;
step 5.2, according to the distance between the static barrier and the dynamic barrier calculated in the step 5.1 and the blind guiding glasses user, giving an early warning to the blind guiding glasses user: when the distance is less than 3 meters, the blind guiding glasses user is in an unsafe area, the control box 6 sends out a warning about the existence of the obstacle to the blind guiding glasses user through the voice module and the vibration module, when the distance is greater than 7 meters, the control box 6 sends out a safe voice signal, and when the distance is greater than 3 meters and less than 7 meters, the system sends out a warning signal about the existence of the obstacle.
The blind guiding glasses are simple in preparation process and low in cost, are suitable for large-scale production and popularization, can be carried about, are convenient to operate and control, can effectively remind a user of the distance between the user and an obstacle through an anti-collision early warning method, avoid the user from colliding with the obstacle on a blind road, and bring great convenience to the user in traveling.
The invention discloses an anti-collision early warning method for blind guiding glasses, which is a blind road identification detection algorithm integrating color features and texture features. Extracting yellow b component of the blind road in Lab color space, segmenting the blind road image by a maximum inter-class variance method, preliminarily extracting edge points of the blind road by adopting a Canny operator through texture features of the blind road, further optimizing the edge points of the blind road through Hough transformation, and simultaneously fusing segmented blind road regions to effectively identify the blind road.
Because the environment along the blind road is complex, and the appearance of the obstacle has unpredictability and uncertainty, the method adopts the single-frame-based image to process the obstacle right ahead, establishes an interested area for the acquired image, namely ignores the area above the blind road area in each image, extracts the blind road area to reduce the search range of the obstacle, further reduces the area of the obstacle by respectively carrying out edge detection in two directions on the area, and finally calculates the entropy of the image to judge whether the area really has the obstacle.
The invention adopts sequence-based images to process the moving obstacles suddenly appearing on both sides of the blind road, and combines a gray projection algorithm to detect the moving obstacles on the basis of a three-frame difference method.
On the basis of realizing the detection and positioning of the cataract obstacle in the blind road area, the invention designs early warning reminding, determines the geometric relation between a world coordinate system and an image coordinate system through the calibration of a vision system on the premise of assuming that the road surface of the blind road is horizontal, estimates the actual distance between the visually impaired and the obstacle through distance measurement, and sends out a safe voice signal, an early warning signal and a warning signal in 3 different forms according to the actual distance, thereby effectively ensuring the personal safety of visually impaired people.
Claims (1)
1. The anti-collision early warning method of the blind guiding glasses is characterized in that the blind guiding glasses are adopted and comprise a glasses frame (5) and lenses (4) arranged on the glasses frame (5), a camera (3), a lighting lamp (2) and a light sensor (8) are respectively arranged on the front side of the glasses frame (5), the camera (3) is positioned at the center of the front side of the glasses frame (5), an earphone (1) is arranged at the end part of one supporting leg of the glasses frame (5), a control key (7) is arranged on the outer side of the end part of the other supporting leg of the glasses frame (5),
also comprises a control box (6), a main controller, an image processor, a battery, a vibration module, a voice module, a Bluetooth communication module and a WiFi communication module which are connected with the main controller are arranged in the control box (6),
the camera (3) is connected with an image processor in the control box (6) through a video line, the control key (7) and the earphone (1) are respectively connected with a main controller and a voice module in the control box (6) through data lines, the illuminating lamp (2) and the light sensor (8) are both connected with the main controller in the control box (6) through data lines, and the control box (6) controls the opening and closing of the illuminating lamp (2) according to light intensity signals transmitted by the light sensor (8);
different key areas are divided on the control keys (7), and braille characters are arranged on each key area, so that the operation of the blind is facilitated;
the spectacle frame (5) is made of hard aluminum alloy materials, so that the spectacle frame (5) is attached to the face without falling off;
the method is implemented according to the following steps:
step 1, shooting to obtain a plurality of continuous images of the traveling direction of a user of the blind guiding glasses;
step 2, preprocessing a plurality of continuous images, increasing the image contrast and reducing the image noise;
the preprocessing process comprises the steps of processing a plurality of continuous images by adopting histogram equalization and homomorphic filtering algorithms;
step 3, extracting blind road characteristics according to the color characteristics and the texture characteristics of the preprocessed image, identifying the blind road, re-shooting the picture if the blind road characteristics cannot be extracted, returning to the step 1, and otherwise, performing the step 4;
step 3.1, converting the RGB model of the preprocessed image into a Lab color space model by adopting a color space conversion model to obtain b-component characteristics of the blind road image in the Lab color space model, wherein the color space conversion model is as follows:
wherein X, Y, Z are the three stimulus values of the image, X0、Y0、Z0Three stimulus values, L, for CTE standard illumination*Representing the brightness of the image, a* 、 b*Representing image chromaticity;
step 3.2, adopting a gray level co-occurrence matrix, and selecting four direction angles of 0 degree, 45 degrees, 90 degrees and 135 degrees to perform statistics on the image to obtain four gray level co-occurrence matrices of P0 degrees, P45 degrees, P90 degrees and P135 degrees respectively;
step 3.3, combining the b-component characteristic of the blind road image obtained in the step 3.1 and the four gray level co-occurrence matrixes respectively of P0 degrees, P45 degrees, P90 degrees and P135 degrees obtained in the step 3.2, and segmenting the image by adopting a maximum inter-class variance method to preliminarily obtain a blind road region;
step 3.4, extracting blind channel edge characteristics from the blind channel region obtained primarily in the step 3.3 by adopting a Canny operator, and extracting the blind channel edge characteristics by Hough transformation to obtain a final complete blind channel region;
step 4, identifying obstacles in the blind road area, determining static obstacles and dynamic obstacles, if the static obstacles and the dynamic obstacles cannot be determined, re-shooting the picture, returning to the step 1 to start, and otherwise, performing the step 5;
step 4.1, determining static obstacles
Step 4.1.1, neglecting the upper area of the blind road area in each image, and extracting the blind road area;
step 4.1.2, calculating the gray values of the blind road region obtained in the step 4.1.1 in the vertical direction and the horizontal direction, calculating the entropy of the blind road region obtained in the step 4.1.1 to judge whether a static obstacle exists in the blind road region, and judging whether the static obstacle exists when the gray values of the blind road region in the vertical direction and the horizontal direction are both smaller than the entropy of the blind road region;
step 4.2, determining dynamic obstacle
Step 4.2.1, detecting every three adjacent images by adopting a three-frame difference method, wherein the formula of the three-frame difference method is as follows
Wherein, Fk-1(x,y)、Fk(x,y)、Fk+1(x, y) is the k-1, k +1 frame image in succession, G1(x,y)、G2(x, y) is the frame-differenced image;
step 4.2.2, image G after frame difference1(x,y)、G2(x, y) selecting a threshold value T for binarization processing, wherein the binarization processing process comprises the following steps:
wherein T is a threshold value and D1(x,y)、D2(x, y) is the binarized image,
wherein the threshold T is based on the frame-corrupted image G1(x,y)、G2Division of the (x, y) grey valueSelecting cloth;
step 4.2.3, the binarized image D obtained in the step 4.2.21(x,y)、D2(x, y) performing a logical AND operation:
in the formula, R (x, y) is an image obtained after AND operation;
4.2.4, performing gray projection on the image obtained after the AND operation in the horizontal and vertical directions by adopting a gray projection algorithm to obtain two curves of the image obtained after the AND operation in the two directions, performing cross-correlation calculation on the two curves by using a cross-correlation function to obtain cross-correlation curves of the two curves, and determining the dynamic barrier by using a peak point on the cross-correlation curves;
step 5, calculating the distance between the obstacle and the blind guiding glasses user according to the static obstacle and the dynamic obstacle determined in the step 4, and giving an early warning prompt to the blind guiding glasses user;
step 5.1, substituting the static obstacle and dynamic obstacle information obtained in the step 4 into a distance measurement formula to calculate the distance between the static obstacle and the dynamic obstacle and a user of the blind guiding glasses, wherein the distance measurement formula is as follows:
h is the installation height of the camera, H is the height of a static obstacle or a dynamic obstacle, f is the focal length of the camera, u, v, u0,v0,ax,ayThe internal parameters of the camera are taken as the parameters;
step 5.2, according to the distance between the static barrier and the dynamic barrier calculated in the step 5.1 and the blind guiding glasses user, giving an early warning to the blind guiding glasses user: when the distance is less than 3 meters, the blind guiding glasses user is in an unsafe area, the control box (6) sends out a warning about the existence of the obstacle to the blind guiding glasses user through the voice module and the vibration module, when the distance is greater than 7 meters, the control box (6) sends out a safe voice signal, and when the distance is greater than 3 meters and less than 7 meters, the system sends out a warning signal about the existence of the obstacle.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201811451997.7A CN109602585B (en) | 2018-11-30 | 2018-11-30 | Blind guiding glasses and anti-collision early warning method thereof |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201811451997.7A CN109602585B (en) | 2018-11-30 | 2018-11-30 | Blind guiding glasses and anti-collision early warning method thereof |
Publications (2)
Publication Number | Publication Date |
---|---|
CN109602585A CN109602585A (en) | 2019-04-12 |
CN109602585B true CN109602585B (en) | 2022-02-22 |
Family
ID=66006495
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201811451997.7A Active CN109602585B (en) | 2018-11-30 | 2018-11-30 | Blind guiding glasses and anti-collision early warning method thereof |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN109602585B (en) |
Families Citing this family (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110623820A (en) * | 2019-07-15 | 2019-12-31 | 电子科技大学 | Blind device is led to wearable intelligence |
CN113080550A (en) * | 2021-05-12 | 2021-07-09 | 新疆爱华盈通信息技术有限公司 | Headband for recognizing obstacle and method for recognizing obstacle |
Citations (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101336856A (en) * | 2008-08-08 | 2009-01-07 | 西安电子科技大学 | Information acquisition and transfer method of auxiliary vision system |
WO2011083963A2 (en) * | 2010-01-06 | 2011-07-14 | (주) 부성 리싸이클링 | System and method for guiding the walking direction of the visually impaired using rfid blocks |
CN102184550A (en) * | 2011-05-04 | 2011-09-14 | 华中科技大学 | Mobile platform ground movement object detection method |
CN103637900A (en) * | 2013-12-20 | 2014-03-19 | 北京航空航天大学 | Intelligent blind guiding stick based on image identification |
CN203825313U (en) * | 2013-12-16 | 2014-09-10 | 智博锐视(北京)科技有限公司 | Blind navigation glasses |
CN104173174A (en) * | 2014-08-22 | 2014-12-03 | 河海大学常州校区 | Portable walking aid blind guide device and method |
CN104778723A (en) * | 2015-04-09 | 2015-07-15 | 海视英科光电(苏州)有限公司 | Method for performing motion detection on infrared image with three-frame difference method |
CN106074098A (en) * | 2016-06-13 | 2016-11-09 | 金英子 | A kind of navigation system for visually impaired person |
CN106709518A (en) * | 2016-12-20 | 2017-05-24 | 西南大学 | Android platform-based blind way recognition system |
RO132133A2 (en) * | 2016-03-03 | 2017-09-29 | Gafencu Daniela În Calitate De Tutore A Minorei Gafencu Miruna-Alexandra | Individual portable device and method of orientation in motion, meant to be used by persons suffering from severe sight deficiencies |
CN107456363A (en) * | 2017-08-10 | 2017-12-12 | 李海强 | Multifunctional guide walking stick for blind people |
CN108524209A (en) * | 2018-03-30 | 2018-09-14 | 江西科技师范大学 | Blind-guiding method, system, readable storage medium storing program for executing and mobile terminal |
-
2018
- 2018-11-30 CN CN201811451997.7A patent/CN109602585B/en active Active
Patent Citations (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101336856A (en) * | 2008-08-08 | 2009-01-07 | 西安电子科技大学 | Information acquisition and transfer method of auxiliary vision system |
WO2011083963A2 (en) * | 2010-01-06 | 2011-07-14 | (주) 부성 리싸이클링 | System and method for guiding the walking direction of the visually impaired using rfid blocks |
CN102184550A (en) * | 2011-05-04 | 2011-09-14 | 华中科技大学 | Mobile platform ground movement object detection method |
CN203825313U (en) * | 2013-12-16 | 2014-09-10 | 智博锐视(北京)科技有限公司 | Blind navigation glasses |
CN103637900A (en) * | 2013-12-20 | 2014-03-19 | 北京航空航天大学 | Intelligent blind guiding stick based on image identification |
CN104173174A (en) * | 2014-08-22 | 2014-12-03 | 河海大学常州校区 | Portable walking aid blind guide device and method |
CN104778723A (en) * | 2015-04-09 | 2015-07-15 | 海视英科光电(苏州)有限公司 | Method for performing motion detection on infrared image with three-frame difference method |
RO132133A2 (en) * | 2016-03-03 | 2017-09-29 | Gafencu Daniela În Calitate De Tutore A Minorei Gafencu Miruna-Alexandra | Individual portable device and method of orientation in motion, meant to be used by persons suffering from severe sight deficiencies |
CN106074098A (en) * | 2016-06-13 | 2016-11-09 | 金英子 | A kind of navigation system for visually impaired person |
CN106709518A (en) * | 2016-12-20 | 2017-05-24 | 西南大学 | Android platform-based blind way recognition system |
CN107456363A (en) * | 2017-08-10 | 2017-12-12 | 李海强 | Multifunctional guide walking stick for blind people |
CN108524209A (en) * | 2018-03-30 | 2018-09-14 | 江西科技师范大学 | Blind-guiding method, system, readable storage medium storing program for executing and mobile terminal |
Non-Patent Citations (1)
Title |
---|
导盲机器人的交通标志视觉识别技术研究;汤智超;《计算机技术与发展》;20140930;第24卷(第9期);第23-27页 * |
Also Published As
Publication number | Publication date |
---|---|
CN109602585A (en) | 2019-04-12 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Jie et al. | A new traffic light detection and recognition algorithm for electronic travel aid | |
CN106682603B (en) | Real-time driver fatigue early warning system based on multi-source information fusion | |
US9704060B2 (en) | Method for detecting traffic violation | |
CN101847206B (en) | Pedestrian traffic statistical method and system based on traffic monitoring facilities | |
CN105631831B (en) | Video image enhancing method under the conditions of a kind of haze | |
CN110155274B (en) | Underwater intelligent lifesaving system for swimming pool | |
CN109602585B (en) | Blind guiding glasses and anti-collision early warning method thereof | |
CN101999972A (en) | Stereoscopic vision based auxiliary walking device for blindmen and auxiliary method thereof | |
CN111553214B (en) | Method and system for detecting smoking behavior of driver | |
CN108737785B (en) | Indoor automatic detection system that tumbles based on TOF 3D camera | |
CN103034843A (en) | Method for detecting vehicle at night based on monocular vision | |
CN102609724A (en) | Method for prompting ambient environment information by using two cameras | |
CN113101155A (en) | Traffic light intersection blind guiding method and blind guiding device based on machine vision | |
CN109948433A (en) | A kind of embedded human face tracing method and device | |
CN107122732A (en) | The quick license plate locating method of high robust under a kind of monitoring scene | |
CN106248635B (en) | A kind of visibility detecting method based on digital camera | |
CN111428680A (en) | Pupil positioning method based on deep learning | |
CN107480629A (en) | A kind of method for detecting fatigue driving and device based on depth information | |
CN105380590B (en) | A kind of equipment and its implementation with eye position detection function | |
CN107432793A (en) | Safety-type Two-wheeled wheelchair | |
CN107049717B (en) | One kind is based on RGB-D camera and stereosonic visually impaired people's zebra stripes detection system and method | |
CN201870808U (en) | Guide device for blind people based on stereoscopic vision | |
CN102542241A (en) | Embedded image processing and identifying system and method | |
CN107432794A (en) | A kind of Intelligent Recognition and the method driven safely | |
Andalibi et al. | Automatic glare detection via photometric, geometric, and global positioning information |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |