CN111084711B - Terrain detection method of blind guiding stick based on active visual guidance - Google Patents
Terrain detection method of blind guiding stick based on active visual guidance Download PDFInfo
- Publication number
- CN111084711B CN111084711B CN201911355769.4A CN201911355769A CN111084711B CN 111084711 B CN111084711 B CN 111084711B CN 201911355769 A CN201911355769 A CN 201911355769A CN 111084711 B CN111084711 B CN 111084711B
- Authority
- CN
- China
- Prior art keywords
- ground
- blind
- image
- audio
- discriminator
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 230000000007 visual effect Effects 0.000 title claims abstract description 25
- 238000001514 detection method Methods 0.000 title claims abstract description 19
- 238000000034 method Methods 0.000 claims abstract description 29
- 230000001133 acceleration Effects 0.000 claims description 20
- 238000012549 training Methods 0.000 claims description 15
- 230000006870 function Effects 0.000 claims description 14
- 230000002159 abnormal effect Effects 0.000 claims description 11
- 238000005070 sampling Methods 0.000 claims description 10
- 230000005283 ground state Effects 0.000 claims description 8
- 238000000605 extraction Methods 0.000 claims description 6
- 239000011521 glass Substances 0.000 claims description 6
- 230000008569 process Effects 0.000 claims description 6
- 230000004913 activation Effects 0.000 claims description 4
- 210000002569 neuron Anatomy 0.000 claims description 3
- 238000000638 solvent extraction Methods 0.000 claims description 3
- 238000013135 deep learning Methods 0.000 abstract description 4
- 230000001771 impaired effect Effects 0.000 abstract description 4
- 230000009286 beneficial effect Effects 0.000 abstract description 2
- 230000035807 sensation Effects 0.000 abstract description 2
- 208000009119 Giant Axonal Neuropathy Diseases 0.000 description 7
- 201000003382 giant axonal neuropathy 1 Diseases 0.000 description 7
- 230000000875 corresponding effect Effects 0.000 description 6
- 238000012545 processing Methods 0.000 description 4
- 238000011160 research Methods 0.000 description 4
- 241001166076 Diapheromera femorata Species 0.000 description 3
- 230000009467 reduction Effects 0.000 description 3
- 230000015541 sensory perception of touch Effects 0.000 description 3
- 201000004569 Blindness Diseases 0.000 description 2
- 241000282414 Homo sapiens Species 0.000 description 2
- 230000008901 benefit Effects 0.000 description 2
- 230000015572 biosynthetic process Effects 0.000 description 2
- 230000008859 change Effects 0.000 description 2
- 238000006243 chemical reaction Methods 0.000 description 2
- 230000007547 defect Effects 0.000 description 2
- 238000010586 diagram Methods 0.000 description 2
- 238000003786 synthesis reaction Methods 0.000 description 2
- 208000037170 Delayed Emergence from Anesthesia Diseases 0.000 description 1
- 241000253999 Phasmatodea Species 0.000 description 1
- 230000005856 abnormality Effects 0.000 description 1
- 230000000903 blocking effect Effects 0.000 description 1
- 230000002596 correlated effect Effects 0.000 description 1
- 230000001186 cumulative effect Effects 0.000 description 1
- 238000013136 deep learning model Methods 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 239000012634 fragment Substances 0.000 description 1
- 230000007274 generation of a signal involved in cell-cell signaling Effects 0.000 description 1
- 230000036541 health Effects 0.000 description 1
- 239000000463 material Substances 0.000 description 1
- 230000005012 migration Effects 0.000 description 1
- 238000013508 migration Methods 0.000 description 1
- 230000008520 organization Effects 0.000 description 1
- 238000011176 pooling Methods 0.000 description 1
- 238000009877 rendering Methods 0.000 description 1
- 230000011218 segmentation Effects 0.000 description 1
- 230000000638 stimulation Effects 0.000 description 1
- 238000013518 transcription Methods 0.000 description 1
- 230000035897 transcription Effects 0.000 description 1
- 238000013519 translation Methods 0.000 description 1
Images
Classifications
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61H—PHYSICAL THERAPY APPARATUS, e.g. DEVICES FOR LOCATING OR STIMULATING REFLEX POINTS IN THE BODY; ARTIFICIAL RESPIRATION; MASSAGE; BATHING DEVICES FOR SPECIAL THERAPEUTIC OR HYGIENIC PURPOSES OR SPECIFIC PARTS OF THE BODY
- A61H3/00—Appliances for aiding patients or disabled persons to walk about
- A61H3/06—Walking aids for blind persons
- A61H3/061—Walking aids for blind persons with electronic detecting or guiding means
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61H—PHYSICAL THERAPY APPARATUS, e.g. DEVICES FOR LOCATING OR STIMULATING REFLEX POINTS IN THE BODY; ARTIFICIAL RESPIRATION; MASSAGE; BATHING DEVICES FOR SPECIAL THERAPEUTIC OR HYGIENIC PURPOSES OR SPECIFIC PARTS OF THE BODY
- A61H3/00—Appliances for aiding patients or disabled persons to walk about
- A61H3/06—Walking aids for blind persons
- A61H3/068—Sticks for blind persons
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61H—PHYSICAL THERAPY APPARATUS, e.g. DEVICES FOR LOCATING OR STIMULATING REFLEX POINTS IN THE BODY; ARTIFICIAL RESPIRATION; MASSAGE; BATHING DEVICES FOR SPECIAL THERAPEUTIC OR HYGIENIC PURPOSES OR SPECIFIC PARTS OF THE BODY
- A61H3/00—Appliances for aiding patients or disabled persons to walk about
- A61H3/06—Walking aids for blind persons
- A61H3/061—Walking aids for blind persons with electronic detecting or guiding means
- A61H2003/063—Walking aids for blind persons with electronic detecting or guiding means with tactile perception
Landscapes
- Health & Medical Sciences (AREA)
- Epidemiology (AREA)
- Pain & Pain Management (AREA)
- Physical Education & Sports Medicine (AREA)
- Rehabilitation Therapy (AREA)
- Life Sciences & Earth Sciences (AREA)
- Animal Behavior & Ethology (AREA)
- General Health & Medical Sciences (AREA)
- Public Health (AREA)
- Veterinary Medicine (AREA)
- Rehabilitation Tools (AREA)
Abstract
The invention relates to a terrain detection method of a blind guiding stick based on active visual guidance, and belongs to the technical field of active guidance and deep learning. The method of the invention is in the field of disabled person service equipment, utilizes the existing GAN to generate the tactile signal, is used in blind person auxiliary equipment, namely a blind guiding stick, introduces active visual guidance, and can help the blind person to better sense ground information through vibration tactile sensation. Under the condition that the blind people can sense the ground information, the blind people can be prevented from detecting the ground without purpose, but the ground is detected based on active visual guidance, so that the visual handicapped people can go out conveniently. Compared with the traditional blind guiding device, the blind guiding device does not need to purposefully detect through the blind guiding stick, but carries out ground abnormity detection firstly, achieves targeted detection of ground areas, and is more beneficial to the safety, reliability and convenience of the visually impaired people in going out.
Description
Technical Field
The invention relates to a terrain detection method of a blind guiding stick based on active visual guidance, and belongs to the technical field of active guidance and deep learning.
Background
In recent years, with the popularization of electronic products and the change of life styles of people, the number of people with impaired vision or blindness is increasing. According to 2019, the world health organization reports that at least 22 hundred million people worldwide have impaired vision or blindness. Even more alarming is the increasing number of blind people worldwide every year, and the blind people become a group which is not neglected globally.
The touch sense is one of 5 kinds of sense channels of human beings, is a basic channel for human beings to interact information with the outside world, and important information such as hardness, temperature, shape, surface texture and the like from the outside can be sensed through the touch sense channel, and particularly for blind people, the touch sense channel is one of main ways for the blind people to sense external things and compensate visual defects. Therefore, the technology of replacing visual sense with tactile sense has been a research focus, and technologists use the technology in blind-person auxiliary equipment to develop a series of auxiliary equipment for replacing visual sense with tactile sense, wherein the first effective tool for replacing visual sense with tactile sense is a walking stick, and the blind person can obtain surrounding ground information such as material, height and slope by touching the ground through the walking stick. And then the walking stick is provided with a laser, sonar and other obstacle detectors, so that the blind can obtain the direction and distance of the obstacle. Walking sticks are widely used due to the simplicity of operation, but research on converting ground information into a tactile sensation which is an intuitive feeling for the blind is rare, and especially, the research on guiding the blind based on active visual guidance is not seen at present.
The deep learning model has a multi-level structure, and can automatically extract characteristic information from a bottom layer to a high layer in an image. In the process of learning data, the model automatically generates the feature expression of the image without considering the problem of artificially constructing features, so that deep learning is widely applied, and a countermeasure network (GANs) is generated as a generating network model in the deep learning, and is widely applied to the field of computer vision, such as image synthesis, text-to-image synthesis, style migration, image super-resolution, image domain conversion, image restoration and the like. Despite their great success in computer vision, there has been little progress in audio modeling using GAN and little research into generating haptic signals using GAN.
Disclosure of Invention
The invention aims to provide a terrain detection method of a blind guiding stick based on active visual guidance, which overcomes the defects of the prior art, detects ground abnormity based on extracted image characteristics, generates tactile vibration based on a generated confrontation network, converts visual information into tactile information and provides reliable information for a blind person to detect the ground.
The invention provides a terrain detection method of a blind guiding stick based on active visual guidance, which comprises the following steps:
(1) acquiring a ground image by using a camera on the glasses for the blind;
(2) dividing the ground image in the step (1) by adopting a uniform partitioning method to obtain a plurality of image blocks;
(3) and (3) performing feature extraction on the plurality of image blocks in the step (2), wherein the feature extraction method comprises the following steps:
(3-1) color histogram h of each image blockij,cComprises the following steps:
wherein M, N respectively represent the length and width of the image block, fmnRepresenting color values at pixel points (m, n), C representing one color in the image blocks, wherein the color set contained in each image block is C, representing an activation function, and taking a color histogram as each image blockThe image characteristics of (1);
(3-2) calculating color histograms h of all image blocksij,cAverage value of (d):
wherein, I and J respectively represent the positions of the image blocks after the ground image is divided, and the color histogram of the image block with the position of (I, J) is hij,c;
(3-3) setting a confidence threshold value sigma according to the color histogram h of each image blockij,cAnd a color histogram hij,cMean value ofThe state of the ground is judged, if soJudging that the ground state is not abnormal, and the blind person normally walks if the ground state is abnormalJudging that the ground state is abnormal, and performing the step (4);
(4) the method for the tactile representation of the blind guide stick comprises the following steps:
(4-1) acquiring an acceleration signal of the abnormal ground in the step (3-3) through an acceleration sensor of the blind guiding stick, and performing short-time Fourier transform on the acceleration signal to obtain a spectrogram of the ground image corresponding to the acceleration signal;
(4-2) training a generative warfare network (MelGAN) with an LJ Speech dataset, the generative warfare network consisting of a generator and an arbiter, wherein an objective function of the generator is:
the objective function of the discriminator is:
wherein x represents a real audio and is acquired from the LJ Speech data set, and s represents an input spectrogram of the generator and is acquired from the LJ Speech data set; z denotes the Gaussian noise vector, k denotes the kth discriminator in the generative countermeasure network, λ is the weight of the feature matching penalty, T denotes the number of layers of the discrimination network, NiThe number of unit neurons of the i-th discrimination layer is represented, and the parameters are set according to the training precision;a feature map representing the output of the ith discrimination layer of the kth discriminator, G (s, z) representing the audio generated by the generator,the mathematical expectation of the input spectrogram representing the generator and the gaussian noise vector,mathematical expectations of input spectrograms representing real audio and generators, Dk(x) Representing the probability that the kth discriminator discriminates the sample as true audio,a mathematical expectation representing real audio;
the training process comprises the following steps:
(4-2-1) sampling audio in the LJ Speech data set to obtain a sampling signal, and performing short-time Fourier transform on the sampling signal to obtain a corresponding spectrogram;
(4-2-2) inputting the spectrogram of the step (4-2-1) to a generator for generating a countermeasure network, and outputting audio;
(4-2-3) taking the audio output by the generator and the original audio in the LJ Speech data set as the input of a discriminator for generating the countermeasure network, and outputting a discrimination result by the discriminator to obtain the probability that the input audio of the discriminator is the real audio;
(4-2-4) training a generation countermeasure network consisting of (4-2-2) and (4-2-3) according to the objective functions of the generator and the discriminator to obtain the weight of the generation countermeasure network;
(4-3) inputting the spectrogram of the ground image in the step (4-1) into the trained generation countermeasure network in the step (4-2), and outputting an audio corresponding to the acceleration signal in the step (4-1);
and (4-4) outputting the audio frequency obtained in the step (4-3) through a power amplifier to generate touch vibration, and prompting the ground information of the blind according to different ground touch vibration to realize the terrain detection of the blind guiding stick.
Compared with the prior art, the terrain detection method of the blind guiding stick based on the active visual guidance has the advantages that:
the invention discloses a terrain detection method of a blind guiding stick based on active visual guidance, which utilizes the existing GAN to generate a tactile signal in the field of service equipment for disabled people, but in the prior art, GAN is utilized to simulate time sequence data distribution, a vibration tactile signal is converted into an image, and finally the vibration tactile signal is generated according to a texture image or texture characteristics. The method is indirect, end-to-end processing cannot be achieved, and vibration information is inevitably lost in the middle. The method of the invention uses MelGAN in vibrotactile signal generation, realizes the direct conversion from image to vibration, and is an end-to-end process. The method is used for blind auxiliary equipment, namely a blind guiding stick, active visual guidance is introduced, and the blind can be helped to better sense ground information through vibration and touch. Under the condition that the blind people can sense the ground information, the blind people can be prevented from detecting the ground without purpose, but the ground is detected based on active visual guidance, so that the visual handicapped people can go out conveniently. Compared with the traditional blind guiding device, the blind guiding device does not need to purposefully detect through the blind guiding stick, but carries out ground abnormity detection firstly, achieves targeted detection of ground areas, and is more beneficial to the safety, reliability and convenience of the visually impaired people in going out.
Drawings
FIG. 1 is a block flow diagram of the method of the present invention.
FIG. 2 is a terrain detection device of an active visual guide blind guiding stick according to the present invention
In fig. 1, 1 is a camera on the glasses for the blind, 2 is an integrated chip, 3 is an earphone, 4 is a vibrating mass, 5 is a power amplifier, and 6 is an acceleration sensor.
Detailed Description
The invention provides a terrain detection method of a blind guiding stick based on active visual guidance, which has a flow chart shown in figure 1 and comprises the following steps:
(1) a camera 1 on the glasses for the blind is used for acquiring a ground image, as shown in figure 2;
(2) processing the ground image by an integrated chip 2 on the blind glasses, and dividing the ground image in the step (1) by adopting a uniform partitioning method to obtain a plurality of image blocks;
in the step, the image is segmented by using a block method, and common block methods include a uniform block method, a super-pixel segmentation method and the like. The local blocks segmented by the superpixel method are different in size, and the proportion of each feature point in the image information amount cannot be guaranteed to be consistent. In order to avoid introducing more interference parameters, a uniform blocking method is adopted to segment the image in the step.
(3) And (3) performing feature extraction on the plurality of image blocks in the step (2), wherein the feature extraction method comprises the following steps:
the image information of the area where the image block is located is represented by extracting the features of the image block, and meanwhile, noise interference caused by few feature points on the image is reduced. There are many ways to describe the characteristics of objects, and color features are most widely used in image retrieval. The main reason is that the color tends to be quite correlated with the objects or scenes contained in the image. In addition, compared with other visual features, the color features have smaller dependence on the size, direction and visual angle of the image, so that the robustness is higher. The color feature is an intuitive feature based on the pixel points, and comprises a color histogram, a color set, a color cluster, a color correlation diagram and the like. The most common color feature expression method is a color histogram method, which has the advantages that the normalized color feature expression method is not influenced by image rotation, translation and scale change, and the common color histogram feature matching method comprises a distance method, a histogram cumulative method and the like.
(3-1) color histogram h of each image blockij,cComprises the following steps:
wherein M, N respectively represent the length and width of the image block, fmnRepresenting color values at pixel points (m, n), C represents one color in the image blocks, the color set contained in each image block is C, representing an activation function, and taking a color histogram as the image feature of each image block;
(3-2) calculating color histograms h of all image blocksij,cAverage value of (d):
wherein, I and J respectively represent the positions of the image blocks after the ground image is divided, and the color histogram of the image block with the position of (I, J) is hij,c;
In one embodiment of the present invention, I ═ J ═ 20, the average color histogram of the image block
(3-3) setting a confidence threshold value sigma according to the color histogram h of each image blockij,cAnd a color histogram hij,cMean value ofThe state of the ground is judged, if soIt is determined that there is no abnormality in the ground state,if the blind is walking normallyJudging that the ground state is abnormal, and performing the step (4); the blind person can walk normally under the condition that the ground is not abnormal. If the abnormal position is detected, stopping detection and detecting the abnormal position.
(4) The blind guiding stick is used for tactile representation, the structure of the blind guiding stick is shown in figure 2, 1 is a camera on blind glasses, 2 is an integrated chip, 3 is an earphone, 4 is a vibrating block, 5 is a power amplifier, and 6 is an acceleration sensor. The haptic rendering comprises the steps of:
(4-1) acquiring the acceleration signal of the abnormal ground in the step (3-3) through an acceleration sensor of the blind guiding stick, and performing short-time Fourier transform (STFT) on the acceleration signal, wherein the STFT is a mathematical transform related to the Fourier transform and determines a local area sine wave of the frequency and phase time-varying signal. The STFT processed signal has locality in the time and frequency domains. The acceleration signal is used as a representation of vibrotactile stimulation, the acceleration signal can be obtained through an acceleration sensor arranged on the blind guiding stick, and a spectrogram of the ground image corresponding to the acceleration signal is obtained through STFT;
(4-2) training a generative warfare network (MelGAN) with an LJ Speech dataset, the generative warfare network consisting of a generator and an arbiter, wherein an objective function of the generator is:
the objective function of the discriminator is:
wherein x represents a real audio and is acquired from the LJ Speech data set, and s represents an input spectrogram of the generator and is acquired from the LJ Speech data set; z represents a Gaussian noise vector, k represents the kth arbiter in the generative countermeasure networkλ is the weight of the feature matching loss, T represents the number of layers of the discrimination network, NiThe number of unit neurons of the i-th discrimination layer is represented, and the parameters are set according to the training precision;a feature map representing the output of the ith discrimination layer of the kth discriminator, G (s, z) representing the audio generated by the generator,the mathematical expectation of the input spectrogram representing the generator and the gaussian noise vector,mathematical expectations of input spectrograms representing real audio and generators, Dk(x) Representing the probability that the kth discriminator discriminates the sample as true audio,a mathematical expectation representing real audio;
the generation of the countermeasure network (MelGAN) for training is an autoregressive forward convolution structure. MelGAN enables the generation of audio waveforms in GAN. This is the first to successfully train GANs to generate raw audio without the need for additional perceptual loss functions, while still producing high quality audio generation models. The training process comprises the following steps:
(4-2-1) the LJ Speech dataset used for training is a public domain Speech dataset consisting of 13100 short audio clips from a speaker reading paragraphs in 7 non-novel books. There is one transcription per fragment. The length of the clips varies from 1 second to 10 seconds, and the total length is about 24 hours. Is a common audio data set for training models. Sampling audio in the LJ Speech data set to obtain a sampling signal, and performing short-time Fourier transform on the sampling signal to obtain a corresponding spectrogram;
(4-2-2) inputting the spectrogram of the step (4-2-1) to a generator for generating a countermeasure network, and outputting audio;
inputting a spectrogram, passing through a convolutional layer, then sending the spectrogram to an upsampling stage, sequentially passing through 8 times of upsampling and 2 times of upsampling twice, sending the spectrogram to a residual error module with cavity convolution after each upsampling, and finally obtaining audio output through a convolutional layer. The residual block is mainly composed of 3 cavity convolution blocks, and each cavity convolution block is composed of two layers of convolution layers with different expansion rates and an activation function. The hole convolution is chosen to enhance the remote correlation between time steps in the audio generation process. The sensing field of the cavity convolution layer increases exponentially with the increase of the number of layers, and the sensing field of each output time step can be effectively increased. There is a large overlap in the sensing fields at longer time steps, resulting in better remote correlation.
(4-2-3) taking the audio output by the generator and the original audio in the LJ Speech data set as the input of a discriminator for generating the countermeasure network, and outputting a discrimination result by the discriminator to obtain the probability that the input audio of the discriminator is the real audio;
the discriminator adopts a multi-scale architecture, namely, the original audio is discriminated, the original audio is subjected to frequency reduction processing and then fed into the next discriminator for discrimination, the frequency reduction mode adopts an average pooling method, 2 times of frequency reduction processing are carried out totally, and the discriminator corresponds to 3 scales. The inner module design of the discriminator mainly comprises a convolution layer and a down-sampling layer.
(4-2-4) training a generation countermeasure network consisting of (4-2-2) and (4-2-3) according to the objective functions of the generator and the discriminator to obtain the weight of the generation countermeasure network;
(4-3) inputting the spectrogram of the ground image in the step (4-1) into the trained generation countermeasure network in the step (4-2), and outputting an audio corresponding to the acceleration signal in the step (4-1);
and (4-4) outputting the audio frequency obtained in the step (4-3) through a power amplifier to generate touch vibration, and prompting the ground information of the blind according to different ground touch vibration to realize the terrain detection of the blind guiding stick.
Claims (1)
1. A terrain detection method of a blind guiding stick based on active visual guidance is characterized by comprising the following steps:
(1) acquiring a ground image by using a camera on the glasses for the blind;
(2) dividing the ground image in the step (1) by adopting a uniform partitioning method to obtain a plurality of image blocks;
(3) and (3) performing feature extraction on the plurality of image blocks in the step (2), wherein the feature extraction method comprises the following steps:
(3-1) color histogram h of each image blockij,cComprises the following steps:
wherein M, N respectively represent the length and width of the image block, fmnRepresenting color values at pixel points (m, n), C represents one color in the image blocks, the color set contained in each image block is C, representing an activation function, and taking a color histogram as the image feature of each image block;
(3-2) calculating color histograms h of all image blocksij,cAverage value of (d):
wherein, I and J respectively represent the positions of the image blocks after the ground image is divided, and the color histogram of the image block with the position of (I, J) is hij,c;
(3-3) setting a confidence threshold value sigma according to the color histogram h of each image blockij,cAnd a color histogram hij,cMean value ofThe state of the ground is judged, if soJudging that the ground state is not abnormal, and the blind person normally walks if the ground state is abnormalJudging that the ground state is abnormal, and performing the step (4);
(4) the method for the tactile representation of the blind guide stick comprises the following steps:
(4-1) acquiring an acceleration signal of the abnormal ground in the step (3-3) through an acceleration sensor of the blind guiding stick, and performing short-time Fourier transform on the acceleration signal to obtain a spectrogram of the ground image corresponding to the acceleration signal;
(4-2) training a generative warfare network (MelGAN) with an LJ Speech dataset, the generative warfare network consisting of a generator and an arbiter, wherein an objective function of the generator is:
the objective function of the discriminator is:
wherein x represents a real audio and is acquired from the LJ Speech data set, and s represents an input spectrogram of the generator and is acquired from the LJ Speech data set; z denotes the Gaussian noise vector, k denotes the kth discriminator in the generative countermeasure network, λ is the weight of the feature matching penalty, T denotes the number of layers of the discrimination network, NiThe number of unit neurons of the i-th discrimination layer is represented, and the parameters are set according to the training precision;a feature map representing the output of the ith discrimination layer of the kth discriminator, G (s, z) representing the audio generated by the generator,input spectrogram representing a generator and Gaussian noiseThe mathematical expectation of the vector is that,mathematical expectations of input spectrograms representing real audio and generators, Dk(x) Representing the probability that the kth discriminator discriminates the sample as true audio,a mathematical expectation representing real audio;
the training process comprises the following steps:
(4-2-1) sampling audio in the LJ Speech data set to obtain a sampling signal, and performing short-time Fourier transform on the sampling signal to obtain a corresponding spectrogram;
(4-2-2) inputting the spectrogram of the step (4-2-1) to a generator for generating a countermeasure network, and outputting audio;
(4-2-3) taking the audio output by the generator and the original audio in the LJ Speech data set as the input of a discriminator for generating the countermeasure network, and outputting a discrimination result by the discriminator to obtain the probability that the input audio of the discriminator is the real audio;
(4-2-4) training a generation countermeasure network consisting of (4-2-2) and (4-2-3) according to the objective functions of the generator and the discriminator to obtain the weight of the generation countermeasure network;
(4-3) inputting the spectrogram of the ground image in the step (4-1) into the trained generation countermeasure network in the step (4-2), and outputting an audio corresponding to the acceleration signal in the step (4-1);
and (4-4) outputting the audio frequency obtained in the step (4-3) through a power amplifier to generate touch vibration, and prompting the ground information of the blind according to different ground touch vibration to realize the terrain detection of the blind guiding stick.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201911355769.4A CN111084711B (en) | 2019-12-25 | 2019-12-25 | Terrain detection method of blind guiding stick based on active visual guidance |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201911355769.4A CN111084711B (en) | 2019-12-25 | 2019-12-25 | Terrain detection method of blind guiding stick based on active visual guidance |
Publications (2)
Publication Number | Publication Date |
---|---|
CN111084711A CN111084711A (en) | 2020-05-01 |
CN111084711B true CN111084711B (en) | 2020-12-11 |
Family
ID=70397122
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201911355769.4A Active CN111084711B (en) | 2019-12-25 | 2019-12-25 | Terrain detection method of blind guiding stick based on active visual guidance |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN111084711B (en) |
Family Cites Families (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5973618A (en) * | 1996-09-25 | 1999-10-26 | Ellis; Christ G. | Intelligent walking stick |
CN103645480B (en) * | 2013-12-04 | 2015-11-18 | 北京理工大学 | Based on the topography and landform character construction method of laser radar and fusing image data |
CN103839238B (en) * | 2014-02-28 | 2017-02-15 | 西安电子科技大学 | SAR image super-resolution method based on marginal information and deconvolution |
GB2563198B (en) * | 2017-03-15 | 2021-05-26 | Jaguar Land Rover Ltd | A system for identifying water ahead of a vehicle |
CN107146221B (en) * | 2017-04-18 | 2020-04-21 | 重庆金山医疗器械有限公司 | Method for positioning main terrain boundary in WCE color video based on color texture descriptor of visual perception |
CN108960287B (en) * | 2018-05-29 | 2022-08-05 | 杭州视氪科技有限公司 | Blind person auxiliary glasses capable of realizing terrain and target detection |
CN110147780B (en) * | 2019-05-28 | 2021-01-01 | 山东大学 | Real-time field robot terrain identification method and system based on hierarchical terrain |
-
2019
- 2019-12-25 CN CN201911355769.4A patent/CN111084711B/en active Active
Also Published As
Publication number | Publication date |
---|---|
CN111084711A (en) | 2020-05-01 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Atila et al. | Attention guided 3D CNN-LSTM model for accurate speech based emotion recognition | |
CN107578775B (en) | Multi-classification voice method based on deep neural network | |
US20180061439A1 (en) | Automatic audio captioning | |
CN105740823B (en) | Dynamic gesture track recognizing method based on depth convolutional neural networks | |
CN107492382B (en) | Voiceprint information extraction method and device based on neural network | |
Ghoraani et al. | Time–frequency matrix feature extraction and classification of environmental audio signals | |
US9020822B2 (en) | Emotion recognition using auditory attention cues extracted from users voice | |
EP2695160B1 (en) | Speech syllable/vowel/phone boundary detection using auditory attention cues | |
JP6189970B2 (en) | Combination of auditory attention cue and phoneme posterior probability score for sound / vowel / syllable boundary detection | |
CN110428364B (en) | Method and device for expanding Parkinson voiceprint spectrogram sample and computer storage medium | |
CN111126280B (en) | Gesture recognition fusion-based aphasia patient auxiliary rehabilitation training system and method | |
KR20160072676A (en) | Apparatus and method for detecting object in image, apparatus and method for computer aided diagnosis | |
CN116129129B (en) | Character interaction detection model and detection method | |
CN111915618A (en) | Example segmentation algorithm and computing device based on peak response enhancement | |
CN117116290B (en) | Method and related equipment for positioning defects of numerical control machine tool parts based on multidimensional characteristics | |
CN116884648A (en) | Voice interaction optimization method, device, equipment and medium based on traditional Chinese medicine inquiry | |
Mavaddati | Voice-based age, gender, and language recognition based on ResNet deep model and transfer learning in spectro-temporal domain | |
Wang et al. | A hierarchical birdsong feature extraction architecture combining static and dynamic modeling | |
CN111084711B (en) | Terrain detection method of blind guiding stick based on active visual guidance | |
Kostopoulos et al. | Haptic access to conventional 2D maps for the visually impaired | |
Rodríguez-Hidalgo et al. | Echoic log-surprise: A multi-scale scheme for acoustic saliency detection | |
CN116363700A (en) | Method for re-identifying blocked pedestrians based on blocking perception and feature restoration | |
Siva Teja et al. | Identification of sloshing noises using convolutional neural network | |
Bhukhya et al. | Virtual Assistant and Navigation for Visually Impaired using Deep Neural Network and Image Processing | |
Ige et al. | State-of-the-Art in 1D Convolutional Neural Networks: A Survey |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |