CN111084711A - Terrain detection method of blind guiding stick based on active visual guidance - Google Patents

Terrain detection method of blind guiding stick based on active visual guidance Download PDF

Info

Publication number
CN111084711A
CN111084711A CN201911355769.4A CN201911355769A CN111084711A CN 111084711 A CN111084711 A CN 111084711A CN 201911355769 A CN201911355769 A CN 201911355769A CN 111084711 A CN111084711 A CN 111084711A
Authority
CN
China
Prior art keywords
ground
blind
audio
image
discriminator
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201911355769.4A
Other languages
Chinese (zh)
Other versions
CN111084711B (en
Inventor
刘华平
李尧尧
赵怀林
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Tsinghua University
Original Assignee
Tsinghua University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Tsinghua University filed Critical Tsinghua University
Priority to CN201911355769.4A priority Critical patent/CN111084711B/en
Publication of CN111084711A publication Critical patent/CN111084711A/en
Application granted granted Critical
Publication of CN111084711B publication Critical patent/CN111084711B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61HPHYSICAL THERAPY APPARATUS, e.g. DEVICES FOR LOCATING OR STIMULATING REFLEX POINTS IN THE BODY; ARTIFICIAL RESPIRATION; MASSAGE; BATHING DEVICES FOR SPECIAL THERAPEUTIC OR HYGIENIC PURPOSES OR SPECIFIC PARTS OF THE BODY
    • A61H3/00Appliances for aiding patients or disabled persons to walk about
    • A61H3/06Walking aids for blind persons
    • A61H3/061Walking aids for blind persons with electronic detecting or guiding means
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61HPHYSICAL THERAPY APPARATUS, e.g. DEVICES FOR LOCATING OR STIMULATING REFLEX POINTS IN THE BODY; ARTIFICIAL RESPIRATION; MASSAGE; BATHING DEVICES FOR SPECIAL THERAPEUTIC OR HYGIENIC PURPOSES OR SPECIFIC PARTS OF THE BODY
    • A61H3/00Appliances for aiding patients or disabled persons to walk about
    • A61H3/06Walking aids for blind persons
    • A61H3/068Sticks for blind persons
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61HPHYSICAL THERAPY APPARATUS, e.g. DEVICES FOR LOCATING OR STIMULATING REFLEX POINTS IN THE BODY; ARTIFICIAL RESPIRATION; MASSAGE; BATHING DEVICES FOR SPECIAL THERAPEUTIC OR HYGIENIC PURPOSES OR SPECIFIC PARTS OF THE BODY
    • A61H3/00Appliances for aiding patients or disabled persons to walk about
    • A61H3/06Walking aids for blind persons
    • A61H3/061Walking aids for blind persons with electronic detecting or guiding means
    • A61H2003/063Walking aids for blind persons with electronic detecting or guiding means with tactile perception

Landscapes

  • Health & Medical Sciences (AREA)
  • Epidemiology (AREA)
  • Pain & Pain Management (AREA)
  • Physical Education & Sports Medicine (AREA)
  • Rehabilitation Therapy (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Animal Behavior & Ethology (AREA)
  • General Health & Medical Sciences (AREA)
  • Public Health (AREA)
  • Veterinary Medicine (AREA)
  • Rehabilitation Tools (AREA)

Abstract

The invention relates to a terrain detection method of a blind guiding stick based on active visual guidance, and belongs to the technical field of active guidance and deep learning. The method of the invention is in the field of disabled person service equipment, utilizes the existing GAN to generate the tactile signal, is used in blind person auxiliary equipment, namely a blind guiding stick, introduces active visual guidance, and can help the blind person to better sense ground information through vibration tactile sensation. Under the condition that the blind people can sense the ground information, the blind people can be prevented from detecting the ground without purpose, but the ground is detected based on active visual guidance, so that the visual handicapped people can go out conveniently. Compared with the traditional blind guiding device, the blind guiding device does not need to purposefully detect through the blind guiding stick, but carries out ground abnormity detection firstly, achieves targeted detection of ground areas, and is more beneficial to the safety, reliability and convenience of the visually impaired people in going out.

Description

Terrain detection method of blind guiding stick based on active visual guidance
Technical Field
The invention relates to a terrain detection method of a blind guiding stick based on active visual guidance, and belongs to the technical field of active guidance and deep learning.
Background
In recent years, with the popularization of electronic products and the change of life styles of people, the number of people with impaired vision or blindness is increasing. According to 2019, the world health organization reports that at least 22 hundred million people worldwide have impaired vision or blindness. Even more alarming is the increasing number of blind people worldwide every year, and the blind people become a group which is not neglected globally.
The touch sense is one of 5 kinds of sense channels of human beings, is a basic channel for human beings to interact information with the outside world, and important information such as hardness, temperature, shape, surface texture and the like from the outside can be sensed through the touch sense channel, and particularly for blind people, the touch sense channel is one of main ways for the blind people to sense external things and compensate visual defects. Therefore, the technology of replacing visual sense with tactile sense has been a research focus, and technologists use the technology in blind-person auxiliary equipment to develop a series of auxiliary equipment for replacing visual sense with tactile sense, wherein the first effective tool for replacing visual sense with tactile sense is a walking stick, and the blind person can obtain surrounding ground information such as material, height and slope by touching the ground through the walking stick. And then the walking stick is provided with a laser, sonar and other obstacle detectors, so that the blind can obtain the direction and distance of the obstacle. Walking sticks are widely used due to the simplicity of operation, but research on converting ground information into a tactile sensation which is an intuitive feeling for the blind is rare, and especially, the research on guiding the blind based on active visual guidance is not seen at present.
The deep learning model has a multi-level structure, and can automatically extract characteristic information from a bottom layer to a high layer in an image. In the process of learning data, the model automatically generates the feature expression of the image without considering the problem of artificially constructing features, so that deep learning is widely applied, and a countermeasure network (GANs) is generated as a generating network model in the deep learning, and is widely applied to the field of computer vision, such as image synthesis, text-to-image synthesis, style migration, image super-resolution, image domain conversion, image restoration and the like. Despite their great success in computer vision, there has been little progress in audio modeling using GAN and little research into generating haptic signals using GAN.
Disclosure of Invention
The invention aims to provide a terrain detection method of a blind guiding stick based on active visual guidance, which overcomes the defects of the prior art, detects ground abnormity based on extracted image characteristics, generates tactile vibration based on a generated confrontation network, converts visual information into tactile information and provides reliable information for a blind person to detect the ground.
The invention provides a terrain detection method of a blind guiding stick based on active visual guidance, which comprises the following steps:
(1) acquiring a ground image by using a camera on the glasses for the blind;
(2) dividing the ground image in the step (1) by adopting a uniform partitioning method to obtain a plurality of image blocks;
(3) and (3) performing feature extraction on the plurality of image blocks in the step (2), wherein the feature extraction method comprises the following steps:
(3-1) color histogram h of each image blockij,cComprises the following steps:
Figure BDA0002335875320000021
wherein M, N respectively represent the length and width of the image block, fmnRepresenting color values at pixel points (m, n), C representing one color in the image blocks, wherein the color set contained in each image block is C, delta represents an activation function, and a color histogram is used as the image feature of each image block;
(3-2) calculating color histograms h of all image blocksij,cAverage value of (d):
Figure BDA0002335875320000022
wherein, I and J respectively represent the positions of the image blocks after the ground image is divided, and the color histogram of the image block with the position of (I, J) is hij,c
(3-3) setting a confidence threshold value sigma according to the color histogram h of each image blockij,cAnd a color histogram hij,cMean value of
Figure BDA0002335875320000023
The state of the ground is judged, if so
Figure BDA0002335875320000024
Then it is determined that the ground state is noneAbnormal condition, the blind walking normally, if
Figure BDA0002335875320000025
Judging that the ground state is abnormal, and performing the step (4);
(4) the method for the tactile representation of the blind guide stick comprises the following steps:
(4-1) acquiring an acceleration signal of the abnormal ground in the step (3-3) through an acceleration sensor of the blind guiding stick, and performing short-time Fourier transform on the acceleration signal to obtain a spectrogram of the ground image corresponding to the acceleration signal;
(4-2) training a generative warfare network (MelGAN) using the LJ Speech dataset, the generative warfare network consisting of a generator and an arbiter, wherein an objective function of the generator is:
Figure BDA0002335875320000031
the objective function of the discriminator is:
Figure BDA0002335875320000036
wherein x represents a real audio and is acquired from the LJ Speech data set, and s represents an input spectrogram of the generator and is acquired from the LJ Speech data set; z denotes the Gaussian noise vector, k denotes the kth discriminator in the generative countermeasure network, λ is the weight of the feature matching penalty, T denotes the number of layers of the discrimination network, NiThe number of unit neurons of the i-th discrimination layer is represented, and the parameters are set according to the training precision;
Figure BDA0002335875320000032
a feature map representing the output of the ith discrimination layer of the kth discriminator, G (s, z) representing the audio generated by the generator,
Figure BDA0002335875320000033
the mathematical expectation of the input spectrogram representing the generator and the gaussian noise vector,
Figure BDA0002335875320000034
mathematical expectations of input spectrograms representing real audio and generators, Dk(x) Representing the probability that the kth discriminator discriminates the sample as true audio,
Figure BDA0002335875320000035
a mathematical expectation representing real audio;
the training process comprises the following steps:
(4-2-1) sampling audio in the LJ Speech data set to obtain a sampling signal, and performing short-time Fourier transform on the sampling signal to obtain a corresponding spectrogram;
(4-2-2) inputting the spectrogram of (4-2-1) to a generator generating the countermeasure network, outputting audio;
(4-2-3) taking the audio output by the generator and the original audio in the LJ Speech data set as the input of a discriminator for generating the countermeasure network, and outputting a discrimination result by the discriminator to obtain the probability that the input audio of the discriminator is the real audio;
(4-2-4) training a generation countermeasure network consisting of (4-2-2) and (4-2-3) according to the objective functions of the generator and the discriminator to obtain the weight of the generation countermeasure network;
(4-3) inputting the spectrogram of the ground image in the step (4-1) into the trained generation countermeasure network in the step (4-2), and outputting an audio corresponding to the acceleration signal in the step (4-1);
and (4-4) outputting the audio frequency obtained in the step (4-3) through a power amplifier to generate touch vibration, and prompting the ground information of the blind according to different ground touch vibration to realize the terrain detection of the blind guiding stick.
Compared with the prior art, the terrain detection method of the blind guiding stick based on the active visual guidance has the advantages that:
the invention discloses a terrain detection method of a blind guiding stick based on active visual guidance, which utilizes the existing GAN to generate a tactile signal in the field of service equipment for disabled people, but in the prior art, GAN is utilized to simulate time sequence data distribution, a vibration tactile signal is converted into an image, and finally the vibration tactile signal is generated according to a texture image or texture characteristics. The method is indirect, end-to-end processing cannot be achieved, and vibration information is inevitably lost in the middle. The method of the invention uses MelGAN in vibrotactile signal generation, realizes the direct conversion from image to vibration, and is an end-to-end process. The method is used for blind auxiliary equipment, namely a blind guiding stick, active visual guidance is introduced, and the blind can be helped to better sense ground information through vibration and touch. Under the condition that the blind people can sense the ground information, the blind people can be prevented from detecting the ground without purpose, but the ground is detected based on active visual guidance, so that the visual handicapped people can go out conveniently. Compared with the traditional blind guiding device, the blind guiding device does not need to purposefully detect through the blind guiding stick, but carries out ground abnormity detection firstly, achieves targeted detection of ground areas, and is more beneficial to the safety, reliability and convenience of the visually impaired people in going out.
Drawings
FIG. 1 is a block flow diagram of the method of the present invention.
FIG. 2 is a terrain detection device of an active visual guide blind guiding stick according to the present invention
In fig. 1, 1 is a camera on the glasses for the blind, 2 is an integrated chip, 3 is an earphone, 4 is a vibrating mass, 5 is a power amplifier, and 6 is an acceleration sensor.
Detailed Description
The invention provides a terrain detection method of a blind guiding stick based on active visual guidance, which has a flow chart shown in figure 1 and comprises the following steps:
(1) a camera 1 on the glasses for the blind is used for acquiring a ground image, as shown in figure 2;
(2) processing the ground image by an integrated chip 2 on the blind glasses, and dividing the ground image in the step (1) by adopting a uniform partitioning method to obtain a plurality of image blocks;
in the step, the image is segmented by using a block method, and common block methods include a uniform block method, a super-pixel segmentation method and the like. The local blocks segmented by the superpixel method are different in size, and the proportion of each feature point in the image information amount cannot be guaranteed to be consistent. In order to avoid introducing more interference parameters, a uniform blocking method is adopted to segment the image in the step.
(3) And (3) performing feature extraction on the plurality of image blocks in the step (2), wherein the feature extraction method comprises the following steps:
the image information of the area where the image block is located is represented by extracting the features of the image block, and meanwhile, noise interference caused by few feature points on the image is reduced. There are many ways to describe the characteristics of objects, and color features are most widely used in image retrieval. The main reason is that the color tends to be quite correlated with the objects or scenes contained in the image. In addition, compared with other visual features, the color features have smaller dependence on the size, direction and visual angle of the image, so that the robustness is higher. The color feature is an intuitive feature based on the pixel points, and comprises a color histogram, a color set, a color cluster, a color correlation diagram and the like. The most common color feature expression method is a color histogram method, which has the advantages that the normalized color feature expression method is not influenced by image rotation, translation and scale change, and the common color histogram feature matching method comprises a distance method, a histogram cumulative method and the like.
(3-1) color histogram h of each image blockij,cComprises the following steps:
Figure BDA0002335875320000051
wherein M, N respectively represent the length and width of the image block, fmnRepresenting color values at pixel points (m, n), C representing one color in the image blocks, wherein the color set contained in each image block is C, delta represents an activation function, and a color histogram is used as the image feature of each image block;
(3-2) calculating color histograms h of all image blocksij,cAverage value of (d):
Figure BDA0002335875320000052
wherein, I and J respectively represent the positions of the image blocks after the ground image is divided, and the color of the image block at the position (I, J)Histogram is hij,c
In one embodiment of the present invention, I ═ J ═ 20, the average color histogram of the image block
Figure BDA0002335875320000057
Figure BDA0002335875320000053
(3-3) setting a confidence threshold value sigma according to the color histogram h of each image blockij,cAnd a color histogram hij,cMean value of
Figure BDA0002335875320000054
The state of the ground is judged, if so
Figure BDA0002335875320000055
Judging that the ground state is not abnormal, and the blind person normally walks if the ground state is abnormal
Figure BDA0002335875320000056
Judging that the ground state is abnormal, and performing the step (4); the blind person can walk normally under the condition that the ground is not abnormal. If the abnormal position is detected, stopping detection and detecting the abnormal position.
(4) The blind guiding stick is used for tactile representation, the structure of the blind guiding stick is shown in figure 2, 1 is a camera on blind glasses, 2 is an integrated chip, 3 is an earphone, 4 is a vibrating block, 5 is a power amplifier, and 6 is an acceleration sensor.
The haptic rendering comprises the steps of:
(4-1) acquiring the acceleration signal of the abnormal ground in the step (3-3) through an acceleration sensor of the blind guiding stick, and performing short-time Fourier transform (STFT) on the acceleration signal, wherein the STFT is a mathematical transform related to the Fourier transform and determines a local area sine wave of the frequency and phase time-varying signal. The STFT processed signal has locality in the time and frequency domains. The acceleration signal is used as a representation of vibrotactile stimulation, the acceleration signal can be obtained through an acceleration sensor arranged on the blind guiding stick, and a spectrogram of the ground image corresponding to the acceleration signal is obtained through STFT;
(4-2) training a generative warfare network (MelGAN) using the LJ Speech dataset, the generative warfare network consisting of a generator and an arbiter, wherein an objective function of the generator is:
Figure BDA0002335875320000061
the objective function of the discriminator is:
Figure BDA0002335875320000062
wherein x represents a real audio and is acquired from the LJ Speech data set, and s represents an input spectrogram of the generator and is acquired from the LJ Speech data set; z denotes the Gaussian noise vector, k denotes the kth discriminator in the generative countermeasure network, λ is the weight of the feature matching penalty, T denotes the number of layers of the discrimination network, NiThe number of unit neurons of the i-th discrimination layer is represented, and the parameters are set according to the training precision;
Figure BDA0002335875320000063
a feature map representing the output of the ith discrimination layer of the kth discriminator, G (s, z) representing the audio generated by the generator,
Figure BDA0002335875320000064
the mathematical expectation of the input spectrogram representing the generator and the gaussian noise vector,
Figure BDA0002335875320000065
mathematical expectations of input spectrograms representing real audio and generators, Dk(x) Representing the probability that the kth discriminator discriminates the sample as true audio,
Figure BDA0002335875320000066
a mathematical expectation representing real audio;
the generation of the countermeasure network (MelGAN) for training is an autoregressive forward convolution structure. MelGAN enables the generation of audio waveforms in GAN. This is the first to successfully train GANs to generate raw audio without the need for additional perceptual loss functions, while still producing high quality audio generation models. The training process comprises the following steps:
(4-2-1) the LJ Speech dataset used for training is a public domain Speech dataset consisting of 13100 short audio clips from a speaker reading paragraphs in 7 non-novel books. There is one transcription per fragment. The length of the clips varies from 1 second to 10 seconds, and the total length is about 24 hours. Is a common audio data set for training models. Sampling audio in the LJ Speech data set to obtain a sampling signal, and performing short-time Fourier transform on the sampling signal to obtain a corresponding spectrogram;
(4-2-2) inputting the spectrogram of (4-2-1) to a generator generating the countermeasure network, outputting audio;
inputting a spectrogram, passing through a convolutional layer, then sending the spectrogram to an upsampling stage, sequentially passing through 8 times of upsampling and 2 times of upsampling twice, sending the spectrogram to a residual error module with cavity convolution after each upsampling, and finally obtaining audio output through a convolutional layer. The residual block is mainly composed of 3 cavity convolution blocks, and each cavity convolution block is composed of two layers of convolution layers with different expansion rates and an activation function. The hole convolution is chosen to enhance the remote correlation between time steps in the audio generation process. The sensing field of the cavity convolution layer increases exponentially with the increase of the number of layers, and the sensing field of each output time step can be effectively increased. There is a large overlap in the sensing fields at longer time steps, resulting in better remote correlation.
(4-2-3) taking the audio output by the generator and the original audio in the LJ Speech data set as the input of a discriminator for generating the countermeasure network, and outputting a discrimination result by the discriminator to obtain the probability that the input audio of the discriminator is the real audio;
the discriminator adopts a multi-scale architecture, namely, the original audio is discriminated, the original audio is subjected to frequency reduction processing and then fed into the next discriminator for discrimination, the frequency reduction mode adopts an average pooling method, 2 times of frequency reduction processing are carried out totally, and the discriminator corresponds to 3 scales. The inner module design of the discriminator mainly comprises a convolution layer and a down-sampling layer.
(4-2-4) training a generation countermeasure network consisting of (4-2-2) and (4-2-3) according to the objective functions of the generator and the discriminator to obtain the weight of the generation countermeasure network;
(4-3) inputting the spectrogram of the ground image in the step (4-1) into the trained generation countermeasure network in the step (4-2), and outputting an audio corresponding to the acceleration signal in the step (4-1);
and (4-4) outputting the audio frequency obtained in the step (4-3) through a power amplifier to generate touch vibration, and prompting the ground information of the blind according to different ground touch vibration to realize the terrain detection of the blind guiding stick.

Claims (1)

1. A terrain detection method of a blind guiding stick based on active visual guidance is characterized by comprising the following steps:
(1) acquiring a ground image by using a camera on the glasses for the blind;
(2) dividing the ground image in the step (1) by adopting a uniform partitioning method to obtain a plurality of image blocks;
(3) and (3) performing feature extraction on the plurality of image blocks in the step (2), wherein the feature extraction method comprises the following steps:
(3-1) color histogram h of each image blockij,cComprises the following steps:
Figure FDA0002335875310000011
wherein M, N respectively represent the length and width of the image block, fmnRepresenting color values at pixel points (m, n), C representing one color in the image blocks, wherein the color set contained in each image block is C, delta represents an activation function, and a color histogram is used as the image feature of each image block;
(3-2) calculating color histograms h of all image blocksij,cAverage value of (d):
Figure FDA0002335875310000012
wherein, I and J respectively represent the positions of the image blocks after the ground image is divided, and the color histogram of the image block with the position of (I, J) is hij,c
(3-3) setting a confidence threshold value sigma according to the color histogram h of each image blockij,cAnd a color histogram hij,cMean value of
Figure FDA0002335875310000013
The state of the ground is judged, if so
Figure FDA0002335875310000014
Judging that the ground state is not abnormal, and the blind person normally walks if the ground state is abnormal
Figure FDA0002335875310000015
Judging that the ground state is abnormal, and performing the step (4);
(4) the method for the tactile representation of the blind guide stick comprises the following steps:
(4-1) acquiring an acceleration signal of the abnormal ground in the step (3-3) through an acceleration sensor of the blind guiding stick, and performing short-time Fourier transform on the acceleration signal to obtain a spectrogram of the ground image corresponding to the acceleration signal;
(4-2) training a generative warfare network (MelGAN) using the LJ Speech dataset, the generative warfare network consisting of a generator and an arbiter, wherein an objective function of the generator is:
Figure FDA0002335875310000016
the objective function of the discriminator is:
Figure FDA0002335875310000021
wherein x represents a real audio and is obtained from an LJ Speech data set, and s represents an input spectrogram of the generator and is obtained from an LJSpeech data set; z denotes the Gaussian noise vector, k denotes the kth discriminator in the generative countermeasure network, λ is the weight of the feature matching penalty, T denotes the number of layers of the discrimination network, NiThe number of unit neurons of the i-th discrimination layer is represented, and the parameters are set according to the training precision;
Figure FDA0002335875310000022
a feature map representing the output of the ith discrimination layer of the kth discriminator, G (s, z) representing the audio generated by the generator,
Figure FDA0002335875310000023
the mathematical expectation of the input spectrogram representing the generator and the gaussian noise vector,
Figure FDA0002335875310000024
mathematical expectations of input spectrograms representing real audio and generators, Dk(x) Representing the probability that the kth discriminator discriminates the sample as true audio,
Figure FDA0002335875310000025
a mathematical expectation representing real audio;
the training process comprises the following steps:
(4-2-1) sampling audio in the LJ Speech data set to obtain a sampling signal, and performing short-time Fourier transform on the sampling signal to obtain a corresponding spectrogram;
(4-2-2) inputting the spectrogram of (4-2-1) to a generator generating the countermeasure network, outputting audio;
(4-2-3) taking the audio output by the generator and the original audio in the LJ Speech data set as the input of a discriminator for generating the countermeasure network, and outputting a discrimination result by the discriminator to obtain the probability that the input audio of the discriminator is the real audio;
(4-2-4) training a generation countermeasure network consisting of (4-2-2) and (4-2-3) according to the objective functions of the generator and the discriminator to obtain the weight of the generation countermeasure network;
(4-3) inputting the spectrogram of the ground image in the step (4-1) into the trained generation countermeasure network in the step (4-2), and outputting an audio corresponding to the acceleration signal in the step (4-1);
and (4-4) outputting the audio frequency obtained in the step (4-3) through a power amplifier to generate touch vibration, and prompting the ground information of the blind according to different ground touch vibration to realize the terrain detection of the blind guiding stick.
CN201911355769.4A 2019-12-25 2019-12-25 Terrain detection method of blind guiding stick based on active visual guidance Active CN111084711B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201911355769.4A CN111084711B (en) 2019-12-25 2019-12-25 Terrain detection method of blind guiding stick based on active visual guidance

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201911355769.4A CN111084711B (en) 2019-12-25 2019-12-25 Terrain detection method of blind guiding stick based on active visual guidance

Publications (2)

Publication Number Publication Date
CN111084711A true CN111084711A (en) 2020-05-01
CN111084711B CN111084711B (en) 2020-12-11

Family

ID=70397122

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201911355769.4A Active CN111084711B (en) 2019-12-25 2019-12-25 Terrain detection method of blind guiding stick based on active visual guidance

Country Status (1)

Country Link
CN (1) CN111084711B (en)

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5973618A (en) * 1996-09-25 1999-10-26 Ellis; Christ G. Intelligent walking stick
CN103645480A (en) * 2013-12-04 2014-03-19 北京理工大学 Geographic and geomorphic characteristic construction method based on laser radar and image data fusion
CN103839238A (en) * 2014-02-28 2014-06-04 西安电子科技大学 SAR image super-resolution method based on marginal information and deconvolution
CN107146221A (en) * 2017-04-18 2017-09-08 重庆金山医疗器械有限公司 The localization method on main landform border in the WCE color videos for color and vein description that view-based access control model is perceived
WO2018166747A1 (en) * 2017-03-15 2018-09-20 Jaguar Land Rover Limited Improvements in vehicle control
CN108960287A (en) * 2018-05-29 2018-12-07 杭州视氪科技有限公司 A kind of blind person's auxiliary eyeglasses for realizing landform and target detection
CN110147780A (en) * 2019-05-28 2019-08-20 山东大学 The landform recognition methods of real-time field robot and system based on level landform

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5973618A (en) * 1996-09-25 1999-10-26 Ellis; Christ G. Intelligent walking stick
CN103645480A (en) * 2013-12-04 2014-03-19 北京理工大学 Geographic and geomorphic characteristic construction method based on laser radar and image data fusion
CN103839238A (en) * 2014-02-28 2014-06-04 西安电子科技大学 SAR image super-resolution method based on marginal information and deconvolution
WO2018166747A1 (en) * 2017-03-15 2018-09-20 Jaguar Land Rover Limited Improvements in vehicle control
CN107146221A (en) * 2017-04-18 2017-09-08 重庆金山医疗器械有限公司 The localization method on main landform border in the WCE color videos for color and vein description that view-based access control model is perceived
CN108960287A (en) * 2018-05-29 2018-12-07 杭州视氪科技有限公司 A kind of blind person's auxiliary eyeglasses for realizing landform and target detection
CN110147780A (en) * 2019-05-28 2019-08-20 山东大学 The landform recognition methods of real-time field robot and system based on level landform

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
夏轩等: "基于DSP的主动视觉系统", 《机器人》 *

Also Published As

Publication number Publication date
CN111084711B (en) 2020-12-11

Similar Documents

Publication Publication Date Title
CN107578775B (en) Multi-classification voice method based on deep neural network
US20180061439A1 (en) Automatic audio captioning
CN107492382B (en) Voiceprint information extraction method and device based on neural network
US9020822B2 (en) Emotion recognition using auditory attention cues extracted from users voice
US20200218806A1 (en) Artificial intelligence based malware detection system and method
JP6189970B2 (en) Combination of auditory attention cue and phoneme posterior probability score for sound / vowel / syllable boundary detection
EP2695160B1 (en) Speech syllable/vowel/phone boundary detection using auditory attention cues
CN109935243A (en) Speech-emotion recognition method based on the enhancing of VTLP data and multiple dimensioned time-frequency domain cavity convolution model
CN110428364B (en) Method and device for expanding Parkinson voiceprint spectrogram sample and computer storage medium
CN109559758B (en) Method for converting texture image into tactile signal based on deep learning
CN111915618A (en) Example segmentation algorithm and computing device based on peak response enhancement
Shinde et al. Real time two way communication approach for hearing impaired and dumb person based on image processing
CN116129129B (en) Character interaction detection model and detection method
CN111126280A (en) Gesture recognition fusion-based aphasia patient auxiliary rehabilitation training system and method
CN111084711B (en) Terrain detection method of blind guiding stick based on active visual guidance
Rodríguez-Hidalgo et al. Echoic log-surprise: A multi-scale scheme for acoustic saliency detection
EP3847646B1 (en) An audio processing apparatus and method for audio scene classification
Seong et al. A review of audio-visual speech recognition
Kanisha et al. Speech recognition with advanced feature extraction methods using adaptive particle swarm optimization
Wang et al. A hierarchical birdsong feature extraction architecture combining static and dynamic modeling
Siva Teja et al. Identification of sloshing noises using convolutional neural network
Mavaddati Voice-based age, gender, and language recognition based on ResNet deep model and transfer learning in spectro-temporal domain
Rajab et al. An automatic lip reading for short sentences using deep learning nets.
CN114723049A (en) Class activation mapping method and device based on gradient optimization
Xie et al. Image processing and classification procedure for the analysis of australian frog vocalisations

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant