CN114159281A - Device and method capable of realizing stair detection - Google Patents
Device and method capable of realizing stair detection Download PDFInfo
- Publication number
- CN114159281A CN114159281A CN202210007685.7A CN202210007685A CN114159281A CN 114159281 A CN114159281 A CN 114159281A CN 202210007685 A CN202210007685 A CN 202210007685A CN 114159281 A CN114159281 A CN 114159281A
- Authority
- CN
- China
- Prior art keywords
- detection
- stairs
- threshold value
- stair
- ultrasonic sensor
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000000034 method Methods 0.000 title claims abstract description 35
- 238000001514 detection method Methods 0.000 title claims abstract description 32
- 241001166076 Diapheromera femorata Species 0.000 claims abstract description 20
- 230000001771 impaired effect Effects 0.000 claims abstract description 12
- 230000006870 function Effects 0.000 claims abstract description 10
- 238000013528 artificial neural network Methods 0.000 claims abstract description 8
- 238000013136 deep learning model Methods 0.000 claims abstract 2
- 238000011897 real-time detection Methods 0.000 claims abstract 2
- 238000012549 training Methods 0.000 claims description 22
- 240000007651 Rubus glaucus Species 0.000 claims description 9
- 235000011034 Rubus glaucus Nutrition 0.000 claims description 9
- 235000009122 Rubus idaeus Nutrition 0.000 claims description 9
- 238000002372 labelling Methods 0.000 claims description 9
- 238000005259 measurement Methods 0.000 claims description 8
- 238000012360 testing method Methods 0.000 claims description 8
- 238000012545 processing Methods 0.000 claims description 7
- 230000008569 process Effects 0.000 claims description 6
- 238000013527 convolutional neural network Methods 0.000 claims description 5
- 230000000977 initiatory effect Effects 0.000 claims description 3
- 238000010606 normalization Methods 0.000 claims description 3
- 239000011664 nicotinic acid Substances 0.000 description 3
- 238000011176 pooling Methods 0.000 description 3
- 238000004364 calculation method Methods 0.000 description 2
- 238000011161 development Methods 0.000 description 2
- 238000000605 extraction Methods 0.000 description 2
- 238000001914 filtration Methods 0.000 description 2
- 230000005764 inhibitory process Effects 0.000 description 2
- 230000001629 suppression Effects 0.000 description 2
- 201000004569 Blindness Diseases 0.000 description 1
- 208000007014 Retinitis pigmentosa Diseases 0.000 description 1
- 206010047571 Visual impairment Diseases 0.000 description 1
- 238000013459 approach Methods 0.000 description 1
- 238000012937 correction Methods 0.000 description 1
- 230000002354 daily effect Effects 0.000 description 1
- 230000007423 decrease Effects 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 238000010586 diagram Methods 0.000 description 1
- 230000008030 elimination Effects 0.000 description 1
- 238000003379 elimination reaction Methods 0.000 description 1
- 230000003203 everyday effect Effects 0.000 description 1
- 230000004927 fusion Effects 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 238000012417 linear regression Methods 0.000 description 1
- 238000013507 mapping Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000008092 positive effect Effects 0.000 description 1
- 238000011160 research Methods 0.000 description 1
- 230000000717 retained effect Effects 0.000 description 1
- 238000005070 sampling Methods 0.000 description 1
- 238000012216 screening Methods 0.000 description 1
- 230000009466 transformation Effects 0.000 description 1
- 238000013519 translation Methods 0.000 description 1
- 208000029257 vision disease Diseases 0.000 description 1
- 230000004393 visual impairment Effects 0.000 description 1
Images
Classifications
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61H—PHYSICAL THERAPY APPARATUS, e.g. DEVICES FOR LOCATING OR STIMULATING REFLEX POINTS IN THE BODY; ARTIFICIAL RESPIRATION; MASSAGE; BATHING DEVICES FOR SPECIAL THERAPEUTIC OR HYGIENIC PURPOSES OR SPECIFIC PARTS OF THE BODY
- A61H3/00—Appliances for aiding patients or disabled persons to walk about
- A61H3/06—Walking aids for blind persons
- A61H3/068—Sticks for blind persons
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/241—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
- G06F18/2413—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on distances to training or reference patterns
- G06F18/24133—Distances to prototypes
- G06F18/24137—Distances to cluster centroïds
- G06F18/2414—Smoothing the distance, e.g. radial basis function networks [RBFN]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/241—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
- G06F18/2415—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on parametric or probabilistic models, e.g. based on likelihood ratio or false acceptance rate versus a false rejection rate
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/25—Fusion techniques
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61H—PHYSICAL THERAPY APPARATUS, e.g. DEVICES FOR LOCATING OR STIMULATING REFLEX POINTS IN THE BODY; ARTIFICIAL RESPIRATION; MASSAGE; BATHING DEVICES FOR SPECIAL THERAPEUTIC OR HYGIENIC PURPOSES OR SPECIFIC PARTS OF THE BODY
- A61H3/00—Appliances for aiding patients or disabled persons to walk about
- A61H3/06—Walking aids for blind persons
- A61H3/061—Walking aids for blind persons with electronic detecting or guiding means
- A61H2003/063—Walking aids for blind persons with electronic detecting or guiding means with tactile perception
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- Life Sciences & Earth Sciences (AREA)
- General Engineering & Computer Science (AREA)
- Artificial Intelligence (AREA)
- General Physics & Mathematics (AREA)
- Evolutionary Computation (AREA)
- Health & Medical Sciences (AREA)
- Evolutionary Biology (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- General Health & Medical Sciences (AREA)
- Bioinformatics & Computational Biology (AREA)
- Biomedical Technology (AREA)
- Software Systems (AREA)
- Biophysics (AREA)
- Mathematical Physics (AREA)
- Computational Linguistics (AREA)
- Computing Systems (AREA)
- Molecular Biology (AREA)
- Probability & Statistics with Applications (AREA)
- Epidemiology (AREA)
- Pain & Pain Management (AREA)
- Physical Education & Sports Medicine (AREA)
- Rehabilitation Therapy (AREA)
- Animal Behavior & Ethology (AREA)
- Public Health (AREA)
- Veterinary Medicine (AREA)
- Image Analysis (AREA)
Abstract
The invention relates to the technical field of computer vision, and discloses a device and a method capable of realizing stair detection. The RGB-D camera is used for acquiring a three-dimensional environment image, inputting a real-time image into a trained deep learning model fast-rcnn-initiation-v 2-coco model, performing real-time detection, and judging whether the user goes upstairs or downstairs; the ultrasonic sensor is used for measuring the distance from the ground, presetting a threshold value, and judging whether the user goes upstairs or downstairs by comparing the measured height with the threshold value; the system realizes stair detection and identification by adopting a method of fusing the two, and the vibration sensor is used for identifying the stair category and giving a user warning. The blind guiding walking stick for stair detection by the hybrid method provided by the invention combines the image recognition of the deep neural network and the recognition function of the ultrasonic sensor, and provides effective assistance for visually impaired people.
Description
Technical Field
The invention relates to the technical field of computer vision, in particular to a stair detection method based on machine vision.
Background
With the rapid development of world science and technology, new systems are being developed every day to make daily life more comfortable. Those with physical disabilities, however, require more assistance than the average. Technological advances provide solutions that allow them to live in society like other normal people. Nowadays, the development of visual impairment aids has become a prominent research area. Over the past decade, much work has been done in this area to help visually impaired people in different directions. To solve the travel problem of visually impaired people, we developed a stair detection framework using a hybrid approach, in which stair detection is accomplished by a pre-trained model and ultrasonic sensors.
The existing stair detection method has three-dimensional reconstruction based on binocular vision, and although the three-dimensional reconstruction method has high detection accuracy, the method has high requirements on equipment and relatively high cost; the bionic eye can help the visually impaired people to recover partial vision, but the bionic eye can only help the visually impaired people to see the low-resolution gray image, is difficult to distinguish stairs from other scenes, and is only suitable for blind people blindness caused by retinitis pigmentosa; the traditional interaction mode for the visually impaired people is mainly voice prompt and touch vibration. Semantic prompts generally broadcast short information, require a certain time to broadcast, cause delay and accident risk, and have a small amount of transmittable information. The touch vibration is hardware through vibrating a waistband or a vibrating vest to some azimuth information is prompted through vibration, the delay problem can be solved, burden is brought to the visually impaired, and wearing feeling of different people is different.
In view of the above-mentioned shortcomings in the prior art, the present invention provides a device and method for detecting stairs, which mainly combines image detection and ultrasonic sensor to detect and identify stairs.
The difficulty of solving the above-mentioned prior art problem: the three-dimensional reconstruction method has higher requirements on equipment and relatively higher cost; the bionic eye can only help the visually impaired people to see the low-resolution gray level image, and the stairs and other scenes are difficult to distinguish; semantic prompts generally broadcast short information, require a certain time to broadcast, cause delay and accident risk, and have a small amount of transmittable information. The touch vibration brings burden to the visually impaired people, and different people have different wearing feelings.
The significance of solving the above-mentioned prior art problems: the machine vision technology and the ultrasonic sensor are combined to realize the positioning of the stairs, so that the detection is more accurate; the detection and identification of going upstairs or downstairs are realized; the vibration sensor is arranged at the handle part of the walking stick, and reminds a user by the vibration of the walking stick body, so that the conventional auxiliary hardware equipment for vibrating a waistband or a vibration vest is avoided; the method has the advantages of reliability, strong real-time performance, low cost, high precision and the like.
Disclosure of Invention
In order to solve the problems, the invention provides a device and a method capable of realizing stair detection.
The invention is realized in this way, a device for detecting stairs, which comprises: comprises a blind guiding stick body, an ultrasonic sensor, an R-GBD camera, a processor, a vibration sensor and a buzzer, wherein the ultrasonic sensor, the R-GBD camera, the processor, the vibration sensor and the buzzer are connected with the blind guiding stick body;
the RGB-D camera is used for collecting three-dimensional environment images, the ultrasonic sensor is used for measuring the distance from the ground, a threshold value is set, and whether the user ascends or descends stairs is judged by comparing the measured distance with the threshold value;
the ultrasonic sensor is an HC-SR04 module, is arranged on the stick body and is arranged at a fixed position of the stick;
the camera is placed at the top of the walking stick, so that scene information can be conveniently acquired;
the processor adopts a raspberry pi main board;
the vibration sensor is arranged in the handle of the walking stick.
Furthermore, the device further comprises a blind guiding stick body, and an R-GBD camera, a processor, a vibration sensor and a buzzer which are connected with the blind guiding stick body;
the camera is placed at the top of the walking stick, so that scene information can be conveniently acquired;
the processor adopts a raspberry pi main board;
the vibration sensor is arranged in the handle of the walking stick.
Furthermore, the walking stick combines the image recognition of the deep neural network and the recognition function of the ultrasonic sensor, so that effective auxiliary information is provided for visually impaired people, and the condition of missed detection is reduced; and (5) carrying out boundary labeling by adopting a frame regression method, and marking out the confidence coefficient of the recognition.
Further, the walking stick combines the image recognition function of the depth neural network and the recognition function of the ultrasonic sensor, and provides effective assistance for visually impaired people.
Further, the deep neural network uses a deep convolutional neural network, the neural network is trained by adopting a data set with labels, and an initiation network with deeper layers and stronger expression capability replaces a VGG-16 network used for extracting image features in FasterRCNN to obtain a convolutional feature map; inputting the convolution characteristic graph into a regional network RPN; and meanwhile, a redundant anchor frame is optimized on the interested area layer by utilizing a soft-nms algorithm, the position information of the object to be detected is adjusted, and the condition of missing detection is reduced. And then, carrying out boundary labeling by adopting a frame regression method, and marking out the confidence coefficient of the recognition.
Another object of the present invention is to provide a method for enabling detection of stairs, which uses the device for enabling detection of stairs, and the method for enabling detection of stairs comprises the following steps:
(1) collecting images of stairs (including upstairs and downstairs) with different sizes and categories to form a data set;
(2) labeling the data set by adopting a labeling tool, wherein the data set is labeled as upstair and downstair;
(3) dividing the data set marked in the step 2 into a training set and a testing set;
(4) carrying out normalization processing on the data set;
(5) training the training set by adopting a fast-rcnn-initiation-v 2-coco-model to obtain an identification model, wherein in the training process, the initial learning rate is set to be 0.0002, and the maximum iteration number is set to be 200000;
(6) and inputting the test set into the recognition model, outputting the stair category in the test image, and the accuracy of the prediction result.
(7) The accuracy of the prediction result is to position the identified stairs, then to label the boundary by frame regression method, and to mark the confidence of the identification;
(8) the ultrasonic sensor is used for assisting detection and identification, the ground distance is measured, the fixed measurement value is set as a threshold value, the measurement data rule sets that 1000 data are taken per second, the average value is taken, then the average value is compared with the threshold value, if the measurement average value is lower than the threshold value, the user is determined to go upstairs, and if the measurement average value is higher than the threshold value, the user is determined to go downstairs.
Another object of the present invention is to provide a storage medium for storing a program for implementing the above method.
Another object of the present invention is to provide a data processing terminal, which executes the above-mentioned detection method.
The invention has the advantages and positive effects that:
(1) the machine vision technology and the ultrasonic sensor are combined to realize the positioning of the stairs, so that the detection is more accurate;
(2) the detection and identification of going upstairs or downstairs are realized;
(3) the vibration sensor is arranged at the handle part of the walking stick, and reminds a user by the vibration of the walking stick body, so that the conventional auxiliary hardware equipment for vibrating a waistband or a vibration vest is avoided;
(4) the method has the advantages of reliability, strong real-time performance, low cost, high precision and the like.
Drawings
FIG. 1 is a schematic structural view of the present invention;
FIG. 2 is a block diagram of the hardware module connections of the present invention;
FIG. 3 is a flow chart of the fast-RCNN-Iception-V2-COCO model training of the present invention;
FIG. 4 is a flow chart of fasterRCNN according to an embodiment of the present invention;
FIG. 5 is a graph of loss functions during a model training phase according to an embodiment of the present invention;
FIG. 6 is a graph of test results of an example of the present invention.
In fig. 1: 1. a cane body; 2. a handle; 3. an RGB-D camera; 4. a vibration sensor; 5. a wire slot 1; 6. an ultrasonic sensor; 7. a raspberry pi processor; 8. a wire slot 2; 9. a switch; 10. and a reminding module.
Detailed Description
In order to make the objects, technical solutions and advantages of the present invention more apparent, the present invention is further described in detail with reference to the following embodiments. It should be understood that the specific embodiments described herein are merely illustrative of the invention and are not intended to limit the invention.
The application of the principles of the present invention will be further described with reference to the accompanying drawings and specific embodiments.
As shown in fig. 1, the device comprises a walking stick body 1 and a handle 2, wherein an RGB-D camera 3 is arranged in the handle 2, a vibration inductor 4 is connected in the handle 2, and the walking stick body 1 is provided with a transparent wire slot 5 or 8, an ultrasonic sensor 6 and a raspberry sending processor 7; the ultrasonic sensor 6, the RGB-D camera 3 and the vibration sensor 4 are connected with the raspberry pi sensor 7 through the wire slot 5, and the detection result is transmitted to the vibration sensor 4 through the raspberry pi processor 7 and then transmitted to a user.
As shown in fig. 2, in an actual environment, images around a user are collected in real time through an RGB-D camera 3, and a trained model is used for detection and stair recognition; and then data fusion is carried out by combining the measured value of the ultrasonic sensor 6 through the raspberry pi processor 7, and the fused processing result is sent to the user through the reminding module 10.
As shown in fig. 3, the data set processing procedure:
data set
Collecting 510 stairs with different sizes and categories, 300 stairs from different buildings in a real environment, and 210 stairs from a website, wherein training sets and test sets are divided according to a ratio of 7: 3;
the data is normalized, and the resolution is uniformly adjusted to 720x960 pixels.
As shown in fig. 4, the method is run by fast-rcnn and comprises the following specific steps:
(1) inputting the image into a convolutional neural network, and extracting features to obtain a feature map, wherein the convolutional neural network adopts an inceptionV2 network architecture;
the inclusion v2 is a series of improvements made by Google team on the basis of inclusion v1 in month 2 of 2015, and the network is composed of 5 convolution layers of 3 × 3 convolution kernels and 2 pooling layers of 3 × 3 and 8 × 8, and finally, linear layers and classification layers are added to further process the features. The inclusion v2 introduced Batch-Normalization, which replaced the original 5 × 5 convolutional layer with a convolution of two layers 3 × 3, and has better feature extraction function and fewer parameters under the condition of unchanged sensing field of view. In the training stage, a deep convolutional neural network initiation is used for processing a part of the data set serving as a training set, and the length is converted into a feature vector with fixed degree of normality. And (3) convolution calculation process:
for input stair image data 1(w, h) with the size of w × h, the convolution kernel is K (m, n), the convolution result is S (w, h), and the corresponding convolution formula is
Wherein, w image length, h image width, m convolution kernel length, n convolution kernel width
Inputting the convolution characteristic graph into a regional network RPN, screening candidate frames by using a softening non-maximum suppression algorithm and a sampling technology, and selecting candidate frame characteristic information anchor with higher confidence coefficient; meanwhile, the feature map is transmitted to a region of interest pooling (ROIpooling) layer to be subjected to pooling operation, and a candidate region feature map with a fixed size is generated;
generating 9 Anchoroboxes for each pixel on the feature map 45 x 60 through convolution feature extraction, and filtering and marking the generated Anchoroboxes; the filtering and labeling rules are as follows:
pixel-by-pixel pair Anchors classification labels for feature information
Removing anchors beyond 720x960 original border
If the IoU value of anchorbox and grountruth is maximum, the sample is marked as positive, and label is 1
If IoU of achorbox and grountruth is more than 0.7, the label is positive sample, and label is 1
If IoU of achorbox and grountruth is less than 0.3, the label is negative, and label is 0
The rest are neither negative nor positive samples and are not used for final training, label-1
The IOU defines the overlapping degree of two bounding boxes. IoU is the ratio of the overlapping area of rectangular box A, B to the area of the union of rectangular boxes A, B;
pixel-by-pixel regression correction
In addition to tagging, the offset between the anchor and the grountruth is calculated
Set the center store position coordinate (x) of the grountruth calibration frame*,y*) And width and height w*,h*
Coordinate of center point of Anchorbox is xa,yaWidth and height are wa,ha
Offset calculation
Δx=(x*-xa)/wa
Δy=(y*-ya)/ha
Δw=log(w*/wa)
Δh=log(h*/ha)
And learning is carried out through the difference between the two frames, and the cross border elimination and the inhibition of a softening non-maximum value inhibition algorithm are further carried out on the prediction frame, so that the overlapped frame is eliminated.
Performing combined training on classification probability and frame regression by passing the feature map of the generated candidate region through a full connection layer and a linear regression layer to realize classification identification of going upstairs and downstairs and a stair target detection frame;
frame regression method
The input original window P is mapped to obtain a regression window which is closer to the real window G through translation and scale transformation
Given (P)x,Py,Pw,Ph) Find a mapping f such that
Softening non-maximum suppression algorithm (soft-NMS)
In target detection, the classifier will calculate a class score for each bounding box (bbox), that is, the probability that this bbox belongs to each class, and the NMS performs the main process according to these values:
for each class, first set a decay function for all scores < threshold's bbox score, giving a low score;
sorting all bboxes according to scores, selecting the highest score and the corresponding bboxes, traversing the rest bboxes, deleting the bboxes if the overlapping area (IoU) of the selected bboxes and the current highest score bb is larger than a certain threshold value, continuously selecting one highest-score bb from the unprocessed bboxes, and repeating the process; until all remaining bbox is found;
the final prediction was drawn from all classscore and classsolor that retained bbox.
As shown in fig. 5, the loss function used in training is divided into two parts, i.e., classification loss (clslos) and regression loss (bboxregressingloss). During model training, loss occurs at each training step, usually the loss starts from high and decreases as training progresses. The design loss continues to drop below 0.05, taking approximately 600 steps, and training is stopped.
The loss function equation is as follows:
wherein L isclsIs the classification loss and LregIs the regression loss, i is the index of the chosen anchor, and p is the predicted probability of i.
LclsLogarithmic loss, N, is calculated for each anchorclsIs the total number of anchors, here 256, piTo predict the target probability for the anchor,is the logarithmic loss of two classes (stairs and non-stairs), tiIs a vector, which is the predicted offset in the training phase of the RPN;is and tiThe same dimensional vector, representing the anchor, the actual offset of the RPN training phase,
λ is 1, so that the classification and regression weights are the same.
Both loss terms are normalized by Ncls and Nreg, which are the number of samples classified and regressed, respectively.
The working principle of the invention is as follows: open the switch, start the camera, gather the image in real time, the system detects whether there is the stair and whether go up or down the stair through the identification model earlier, then opens ultrasonic sensor 6, carries out the secondary through the distance difference and judges, compares twice result and realizes accurate detection, gives raspberry group treater 7 after fusing with the two measuring result, and treater 7 sends vibration signal to the user. If the walking stick goes upstairs, the walking stick vibration sensor 4 vibrates at intervals; if the walking stick goes down the stairs, the walking stick vibration sensor 4 accelerates continuous vibration; if no staircase is detected, the vibration sensor 4 does not react.
The above description is only for the purpose of illustrating the preferred embodiments of the present invention and is not to be construed as limiting the invention, and any modifications, equivalents and improvements made within the spirit and principle of the present invention are intended to be included within the scope of the present invention.
Claims (7)
1. A device capable of realizing stair detection is characterized by comprising:
the RGB-D camera is used for acquiring a three-dimensional environment image, inputting a real-time image into a trained deep learning model fast-rcnn-initiation-v 2-coco model, performing real-time detection, and judging whether the user goes upstairs or downstairs;
the ultrasonic sensor is used for measuring the distance from the ground, presetting a threshold value, and judging whether the user goes upstairs or downstairs by comparing the measured height with the threshold value; the ultrasonic sensor is arranged on the stick body and is arranged at a fixed position of the stick;
the vibration sensor is used for identifying the stair type and giving a user warning.
2. The device for detecting stairs of claim 1, further comprising a blind guiding stick body and an R-GBD camera, a processor and a vibration sensor and a buzzer connected thereto;
the camera is placed at the top of the walking stick, so that scene information can be conveniently acquired;
the processor adopts a raspberry pi main board;
the vibration sensor is arranged in the handle of the walking stick.
3. The device for detecting the stairs according to claim 1, wherein the walking stick combines the image recognition of the deep neural network and the recognition function of the ultrasonic sensor, so as to provide effective auxiliary information for visually impaired people and reduce the condition of missing detection; and then, carrying out boundary labeling by adopting a frame regression method, and marking out the confidence coefficient of the recognition.
4. A method for detecting stairs by using the device for detecting stairs of claim 1, wherein the method for detecting stairs comprises the following steps:
(1) collecting images of upstairs and downstairs stairs with different sizes and types to form a data set;
(2) labeling the data set by adopting a labeling tool, wherein the data set is labeled as upstair and downstair;
(3) dividing the data set labeled in the step (2) into a training set and a testing set;
(4) carrying out normalization processing on the data set;
(5) training the training set by adopting a fast-rcnn-initiation-v 2-coco-model to obtain an identification model, wherein in the training process, the initial learning rate is set to be 0.0002, and the maximum iteration number is set to be 200000;
(6) inputting the test set into an identification model, outputting the stair category in the test image and the accuracy of the prediction result;
(7) the accuracy of the prediction result is to position the identified stairs, then to label the boundary by frame regression method, and to mark the confidence of the identification;
(8) the ultrasonic sensor is used for assisting detection and identification, the ground distance is measured, the fixed measurement value is set as a threshold value, the measurement data rule sets that 1000 data are taken per second, the average value is taken, then the average value is compared with the threshold value, if the measurement average value is lower than the threshold value, the user is determined to go upstairs, and if the measurement average value is higher than the threshold value, the user is determined to go downstairs.
5. The method for detecting the staircase according to claim 2, wherein the deep neural network uses a deep convolutional neural network, and the neural network is trained by using a labeled data set, and an initiation network with a deeper layer and a stronger expression ability replaces a VGG-16 network for extracting image features in fastern to obtain a convolutional feature map; inputting the convolution characteristic graph into a regional network RPN; and meanwhile, a redundant anchor frame is optimized on the interested area layer by utilizing a soft-nms algorithm, the position information of the object to be detected is adjusted, and the condition of missing detection is reduced. And then, carrying out boundary labeling by adopting a frame regression method, and marking out the confidence coefficient of the recognition.
6. A storage medium for storing a program for implementing the method of claim 4.
7. A data processing terminal, which executes the method of claim 4.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202210007685.7A CN114159281A (en) | 2022-01-06 | 2022-01-06 | Device and method capable of realizing stair detection |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202210007685.7A CN114159281A (en) | 2022-01-06 | 2022-01-06 | Device and method capable of realizing stair detection |
Publications (1)
Publication Number | Publication Date |
---|---|
CN114159281A true CN114159281A (en) | 2022-03-11 |
Family
ID=80489013
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202210007685.7A Pending CN114159281A (en) | 2022-01-06 | 2022-01-06 | Device and method capable of realizing stair detection |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN114159281A (en) |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2004174176A (en) * | 2002-11-24 | 2004-06-24 | Shohei Sato | Visually challenging person and healthy person guide device |
CN110269787A (en) * | 2019-06-26 | 2019-09-24 | 青岛海尔科技有限公司 | A kind of blind-guiding method, device and cap |
US20190325628A1 (en) * | 2018-04-23 | 2019-10-24 | Accenture Global Solutions Limited | Ai-driven design platform |
CN110378422A (en) * | 2019-07-22 | 2019-10-25 | 福州大学 | A kind of weapon recognition methods based on Faster R-CNN |
CN210078040U (en) * | 2019-02-28 | 2020-02-18 | 上海工程技术大学 | Intelligent blind guiding device |
CN211092197U (en) * | 2019-12-05 | 2020-07-28 | 浙江师范大学 | Intelligent indoor navigation anti-lost walking stick device |
-
2022
- 2022-01-06 CN CN202210007685.7A patent/CN114159281A/en active Pending
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2004174176A (en) * | 2002-11-24 | 2004-06-24 | Shohei Sato | Visually challenging person and healthy person guide device |
US20190325628A1 (en) * | 2018-04-23 | 2019-10-24 | Accenture Global Solutions Limited | Ai-driven design platform |
CN210078040U (en) * | 2019-02-28 | 2020-02-18 | 上海工程技术大学 | Intelligent blind guiding device |
CN110269787A (en) * | 2019-06-26 | 2019-09-24 | 青岛海尔科技有限公司 | A kind of blind-guiding method, device and cap |
CN110378422A (en) * | 2019-07-22 | 2019-10-25 | 福州大学 | A kind of weapon recognition methods based on Faster R-CNN |
CN211092197U (en) * | 2019-12-05 | 2020-07-28 | 浙江师范大学 | Intelligent indoor navigation anti-lost walking stick device |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN108898047B (en) | Pedestrian detection method and system based on blocking and shielding perception | |
CN106548182B (en) | Pavement crack detection method and device based on deep learning and main cause analysis | |
CN113065578B (en) | Image visual semantic segmentation method based on double-path region attention coding and decoding | |
CN110287960A (en) | The detection recognition method of curve text in natural scene image | |
CN108960124B (en) | Image processing method and device for pedestrian re-identification | |
CN108389220B (en) | Remote sensing video image motion target real-time intelligent cognitive method and its device | |
CN111160249A (en) | Multi-class target detection method of optical remote sensing image based on cross-scale feature fusion | |
CN111027481B (en) | Behavior analysis method and device based on human body key point detection | |
CN114494192B (en) | Thoracolumbar fracture identification segmentation and detection positioning method based on deep learning | |
CN110705558A (en) | Image instance segmentation method and device | |
CN115861772A (en) | Multi-scale single-stage target detection method based on RetinaNet | |
CN115138059B (en) | Pull-up standard counting method, pull-up standard counting system and storage medium of pull-up standard counting system | |
CN113065431B (en) | Human body violation prediction method based on hidden Markov model and recurrent neural network | |
CN116129129B (en) | Character interaction detection model and detection method | |
CN114972316A (en) | Battery case end surface defect real-time detection method based on improved YOLOv5 | |
CN112861785A (en) | Shielded pedestrian re-identification method based on example segmentation and image restoration | |
CN113870254B (en) | Target object detection method and device, electronic equipment and storage medium | |
CN115861715A (en) | Knowledge representation enhancement-based image target relation recognition algorithm | |
CN111339967A (en) | Pedestrian detection method based on multi-view graph convolution network | |
CN113191352A (en) | Water meter pointer reading identification method based on target detection and binary image detection | |
CN117011932A (en) | Running behavior detection method, electronic device and storage medium | |
CN114159281A (en) | Device and method capable of realizing stair detection | |
CN115953387A (en) | Radiographic image weld defect detection method based on deep learning | |
CN114067186B (en) | Pedestrian detection method and device, electronic equipment and storage medium | |
CN115393351A (en) | Method and device for judging cornea immune state based on Langerhans cells |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination |