CN110503634A - Visibility measurement method based on image automatic identification - Google Patents

Visibility measurement method based on image automatic identification Download PDF

Info

Publication number
CN110503634A
CN110503634A CN201910690633.2A CN201910690633A CN110503634A CN 110503634 A CN110503634 A CN 110503634A CN 201910690633 A CN201910690633 A CN 201910690633A CN 110503634 A CN110503634 A CN 110503634A
Authority
CN
China
Prior art keywords
image
visibility
formula
mass center
puppet
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201910690633.2A
Other languages
Chinese (zh)
Other versions
CN110503634B (en
Inventor
胡辽林
荆霄
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Xian University of Technology
Original Assignee
Xian University of Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Xian University of Technology filed Critical Xian University of Technology
Priority to CN201910690633.2A priority Critical patent/CN110503634B/en
Publication of CN110503634A publication Critical patent/CN110503634A/en
Application granted granted Critical
Publication of CN110503634B publication Critical patent/CN110503634B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • G06T5/73
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/136Segmentation; Edge detection involving thresholding
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/155Segmentation; Edge detection involving morphological operators

Abstract

The present invention discloses the visibility measurement method based on image automatic identification, acquires original image and test image containing target;Maximum kind variance method is respectively adopted, Threshold segmentation is carried out to original image, test image;The test image after Threshold segmentation is continuously expanded using diamond shape probe, obtains expanding image;It is counter to expanding image to be coloured, using centroid principle, " puppet " the mass center calibration of more connected regions after searching continuous expansion;By " puppet " mass center as the quasi- center for selecting multiple target region, multiple target region to be selected is marked;In the dark effect picture of original image after singulation, find the maximum central point of " puppet " mass center light intensity in corresponding label multiple target region to be selected, the final goal region that region where the central point is calculated as visibility, selection suitably remove fog coefficient ω, carry out the calculating of visibility.The image for the different scenes that can be shot to picture pick-up device, selection is most suitable to go fog coefficient, calculates visibility.

Description

Visibility measurement method based on image automatic identification
Technical field
The invention belongs to technical field of image processing, and in particular to the visibility measurement method based on image automatic identification.
Background technique
With the continuous propulsion of reform and opening-up process, economy rapid development and urban population increase, and consequent is energy The consumption in source rapidly increases, and atmosphere pollution is caused to aggravate, and visibility level decline is brought to daily life etc. Serious influence, especially aerospace, in terms of it is very harmful.Though in nationwide by every behave into Row control, but difficulty is larger, still can see in recent years, various regions haze weather takes place frequently.
There are three types of common visibility detecting methods: artificial ocular estimate, device survey method and digital camera method.Ocular estimate error Greatly, it is affected by observer personal considerations;Device survey method is expensive, and site requirements is high, cannot be widely used;Digital camera method Influence practical, at low cost, that adverse weather can be overcome, therefore studied extensively by domestic and international experts and scholars.
Summary of the invention
The object of the present invention is to provide the visibility measurement methods based on image automatic identification, can shoot to picture pick-up device Different scenes image, choose it is most suitable goes fog coefficient, calculate visibility.
The technical solution adopted by the present invention is that the visibility measurement method based on image automatic identification, specifically according to following Step is implemented:
Step 1, original image of the acquisition containing target, same date, identical place, same shooting angle do not acquire not on the same day The target image of gas situation is as test image;
Maximum kind variance method is respectively adopted to original image, test image progress Threshold segmentation in step 2;
Step 3 continuously expands the test image after Threshold segmentation using diamond shape probe, obtains expanding image;
Step 4, it is counter to expanding image coloured, using centroid principle, more connected regions after searching continuous expansion The calibration of " puppet " mass center;
Step 5, the center by " puppet " mass center as quasi- selection multiple target region, mark multiple target region to be selected;
Step 6, original image after singulation dark effect picture in, find corresponding label multiple target area to be selected " puppet " mass center light intensity I in domainc(y) maximum central point, the final goal that the region where the central point is calculated as visibility Region, selection suitably remove fog coefficient ω, carry out the calculating of visibility.
The features of the present invention also characterized in that:
Step 2 detailed process are as follows: the Threshold segmentation of original image, test image is carried out using maximum kind variance method;
It is g (x, y) after Threshold segmentation if the pixel of original image or test image is f (x, y), then:
If the size of original image or test image is M × N, segmentation threshold is denoted as T.
The calculating process of segmentation threshold are as follows: target image vegetarian refreshments is in the ratio of entire imageImage averaging gray scale is μ0;Background pixel point is in the ratio of entire imageImage averaging gray scale is μ1;The gray value of pixel is less than threshold value in image The number of pixels of T is denoted as N0, gray scale is denoted as N greater than the number of pixels of threshold value T1;The overall average gray scale of image is denoted as μ, inter-class variance It is denoted as g;Following formula then can be obtained:
The threshold value T for making the maximum threshold value of inter-class variance g, as acquiring is obtained using traversal.
Step 3 detailed process are as follows:
" probe ", which is done, using structural element collects target image information;The condition about origin is first carried out to structural element S Reflection obtains SV, then in test image after segmentation by SVTranslate x, those translation after at least one non-zero of test image X When element intersects, corresponding origin position set is dilation operation result;
Then expansion results are write as formula (3):
Structural element S is diamond shape probe;
Structural element S continuously expands repeatedly, obtains expanding image.
Step 4 detailed process are as follows:
Step 4.1 searches the pixel of the image after continuous expansion one by one, and the part all the points that pixel is 0 are assigned Value is 1, and all the points that pixel is 1 are assigned a value of 0, obtains anti-rendered image;
Step 4.2 utilizes bwlabel function to anti-rendered image, finds all connected domains separated on anti-rendered image, The principle for recycling mass center to search searches " puppet " mass center, and mass center searches principle such as formula (5):
Wherein, M10It is the summation of all the points x coordinate on object, M01It is the summation of all the points y-coordinate on object, M00It is object The area of body;
Indicate the position of " puppet " mass center.
Step 5 detailed process are as follows: by " puppet " mass center as the quasi- center for selecting multiple target region, utilize Rectangle square Shape function marks multiple target region to be selected, the rectangular area that setting area size is 6 × 6;
‘Position’[x,y,w,h] (6)
In formula (6), x, y are lower-left angular coordinate, and w, h respectively represent wide and high.
Step 6 detailed process are as follows: Cauchy's Mead daytime visibility measuring principle, such as formula (7):
Wherein VmetIndicate that atmospheric visibility, β indicate that extinction coefficient, ε indicate to select 0.05 than sense threshold value;
Then formula (7) is rewritten as formula (8):
The dark effect picture of original image after being divided by dark principle, wherein dark channel prior principle, such as Formula (9):
WhereinIndicate transmissivity, Ic(y) haze image luminous intensity, A are indicatedcIndicate global atmosphere luminous intensity, ω is indicated Fog coefficient is gone, value is 0 < ω < 1;
Meet Lambert-Beer's law between transmissive portions light splitting and incident intensity;
Meet the relationship such as formula (10) between atmospheric extinction coefficient and transmissivity:
T (x)=e-βd(x) (10)
Wherein, β is extinction coefficient, and d is object at a distance from receiving point, can be calculated by marking on map;
According to formula (9), I is chosenc(y) the final goal area that the region where maximum central point is calculated as visibility Domain, selection suitably remove fog coefficient ω, go the presence of fog coefficient and visibility to choose rule, convolution (8), (9), (10) can It is finally inversed by atmospheric visibility.
The presence of fog coefficient and visibility is gone to choose rule as shown in table 1:
Table 1
The beneficial effects of the present invention are:
The present invention is based on the visibility measurement methods of image automatic identification, can be to the different scenes that picture pick-up device is shot Image, selection is most suitable to go fog coefficient, carries out the automatic selection and visibility measuring and calculating of target area;Measurement method of the invention Also have the characteristics that it is practical, at low cost, influenced by Changes in weather it is small.
Detailed description of the invention
Fig. 1 is the grayscale image for shooting original image;
Fig. 2 is the effect picture of different threshold segmentation methods;
Wherein, (a) is Local threshold segmentation method;It (b) is iteration global threshold split plot design;(c) complete for maximum kind variance method Office's thresholding method;
Fig. 3 is expansion process schematic diagram;
Wherein, (a) image;(b) structural element;(c) conditioned reflex;(d) expansion results;
Fig. 4 is expansion process figure;
Wherein, (a) is 1 expansion effect figure;It (b) is 2 expansion effect figures;It (c) is 3 expansion effect figures;It (d) is 4 Secondary expansion effect figure;It (e) is 5 expansion effect figures;It (f) is 6 expansion effect figures;It (g) is 7 expansion effect figures;It (h) is 8 Secondary expansion effect figure;
Fig. 5 is to negate effect picture after repeatedly expanding;
Wherein, (a) is multiple expansion plans, (b) is anti-color-patch map;
Fig. 6 is that more connected regions " puppet " mass center demarcates effect picture;
Fig. 7 is the quasi- selected target area schematic using 6 × 6 size of rectangular function label;
Fig. 8 is original shooting image target area marker effect picture;
Fig. 9 is the dark effect picture of original image;
Figure 10 is several hundred days visibility comparison diagrams;
Figure 11 is that visibility measuring and calculating in several hundred days is recorded a demerit and laser light scattering instrument (CJY-1G) true value overall linear regression analysis Figure;
Figure 12 is the linear regression analysis figure for calculating visibility V < 2000m;
Figure 13 is measuring and calculating 2000≤V of visibility≤10000m linear regression analysis figure;
Figure 14 is the linear regression analysis figure for calculating visibility V > 10000m.
Specific embodiment
The following describes the present invention in detail with reference to the accompanying drawings and specific embodiments.
The present invention is based on the visibility measurement methods of image automatic identification, are specifically implemented according to the following steps:
Step 1, original image of the acquisition containing target, same date, identical place, same shooting angle do not acquire not on the same day The target image of gas situation is as test image;
Maximum kind variance method is respectively adopted to original image, test image progress Threshold segmentation in step 2;
Detailed process are as follows: the Threshold segmentation of original image, test image is carried out using maximum kind variance method;
It is g (x, y) after Threshold segmentation if the pixel of original image or test image is f (x, y), then:
If the size of original image or test image is M × N, segmentation threshold is denoted as T.
The calculating process of segmentation threshold are as follows: target image vegetarian refreshments is in the ratio of entire imageImage averaging gray scale is μ0;Background pixel point is in the ratio of entire imageImage averaging gray scale is μ1;The gray value of pixel is less than threshold value in image The number of pixels of T is denoted as N0, gray scale is denoted as N greater than the number of pixels of threshold value T1;The overall average gray scale of image is denoted as μ, inter-class variance It is denoted as g;Following formula then can be obtained:
The threshold value T for making the maximum threshold value of inter-class variance g, as acquiring is obtained using traversal.
We carry out gray proces to the original image of shooting and obtain result such as Fig. 1, recycle three kinds of different threshold value sides Method processing, obtains different treatment effects such as Fig. 2.Finally there are obvious bimodal image background and target to histogram because of Da-Jin algorithm Separating effect is preferable, and mistake divides rate minimum, therefore selects OTSU Da-Jin algorithm.
Step 3 continuously expands the test image after Threshold segmentation using diamond shape probe, obtains expanding image;
Detailed process are as follows:
" probe ", which is done, using structural element collects target image information;The condition about origin is first carried out to structural element S Reflection obtains SV, then in test image after segmentation by SVTranslate x, those translation after at least one non-zero of test image X When element intersects, corresponding origin position set is dilation operation result;The density bullet of (a) represents target figure in Fig. 3 As X, (b) representative structure element S, figure (c) is shown in its reflection, and expansion results are figure (d).
Then expansion results are write as formula (3):
Structural element S is diamond shape probe;
Structural element S continuously expands repeatedly, obtains expanding image.Effect is as shown in Figure 4.
Step 4, it is counter to expanding image coloured, using centroid principle, more connected regions after searching continuous expansion The calibration of " puppet " mass center;
Detailed process are as follows:
Step 4.1 searches the pixel of the image after continuous expansion one by one, and the part all the points that pixel is 0 are assigned Value is 1, and all the points that pixel is 1 are assigned a value of 0, obtains anti-rendered image, as shown in Figure 5.
Step 4.2 utilizes bwlabel function to anti-rendered image, finds all connected domains separated on anti-rendered image, The principle for recycling mass center to search searches " puppet " mass center, and mass center searches principle such as formula (5):
Wherein, M10It is the summation of all the points x coordinate on object, M01It is the summation of all the points y-coordinate on object, M00It is object The area of body;According to (6) formula, all connected domains " puppet " mass center is found, is marked with " * ", effect such as Fig. 6.
Indicate the position of " puppet " mass center.
Step 5, the center by " puppet " mass center as quasi- selection multiple target region, mark multiple target region to be selected;
Detailed process are as follows: by " puppet " mass center as the quasi- center for selecting multiple target region, utilize Rectangle rectangular function Mark multiple target region to be selected, the rectangular area that setting area size is 6 × 6;Marking serial numbers are easily distinguishable with that can select Color regular font indicates, selects black, signal such as Fig. 7 herein.
‘Position’[x,y,w,h] (6)
In formula (6), x, y are lower-left angular coordinate, and w, h respectively represent wide and high.Final label result is reflected in the original of shooting Such as Fig. 8 on beginning image.
Step 6, original image after singulation dark effect picture in, find corresponding label multiple target area to be selected " puppet " mass center light intensity I in domainc(y) maximum central point, the final goal that the region where the central point is calculated as visibility Region, selection suitably remove fog coefficient ω, carry out the calculating of visibility;
Detailed process are as follows: Cauchy's Mead daytime visibility measuring principle, such as formula (7):
Wherein VmetIndicate that atmospheric visibility, β indicate that extinction coefficient, ε indicate to select 0.05 than sense threshold value;
Then formula (7) is rewritten as formula (8):
The dark effect picture of original image is as shown in Figure 9 after being divided by dark principle, wherein dark is first Principle is tested, such as formula (9):
WhereinIndicate transmissivity, Ic(y) haze image luminous intensity, A are indicatedcIndicate global atmosphere luminous intensity, ω is indicated Fog coefficient is gone, value is 0 < ω < 1;
When light is propagated in the medium, a part can be converted into thermal energy release by Absorption of Medium, and another part is deposited in medium Particle scattering, deviate the original direction of propagation, lose information, remaining part just passes through Jie by the original direction of propagation Matter, into video camera.Therefore, meet Lambert-Beer's law between transmissive portions light splitting and incident intensity.
Meet the relationship such as formula (10) between atmospheric extinction coefficient and transmissivity:
T (x)=e-βd(x) (10)
Wherein, β is extinction coefficient, and d is object at a distance from receiving point, can be calculated by marking on map;
According to formula (9), I is chosenc(y) the final goal area that the region where maximum central point is calculated as visibility Domain, selection suitably remove fog coefficient ω, go the presence of fog coefficient and visibility to choose rule, convolution (8), (9), (10) can It is finally inversed by atmospheric visibility.
The presence of fog coefficient and visibility is gone to choose rule as shown in table 1:
Table 1
Embodiment
Several hundred days images for calculating shooting, using six building, certain teaching as object, to visibility meter pair before the CJY-1G of roof Than verification.Obtain comparison diagram such as Figure 10.From the point of view of the overall trend of figure, the visibility value and instrument value basic trend one that calculate It causes, thinks that the goodness of fit is higher.But at the 45th, 54,55,56,57,84 day, it is seen that two curves have apparent differentiation.This 6 days because of energy View degree instrument failure, obtained data are 0, and therefore, this 6 days data can not verify, and need to reject.Finally, by measuring and calculating, have 8 days Data error be more than allow worst error, therefore the accuracy rate of context of methods be 92.30%.
Measuring and calculating value and CJY-1G true value are subjected to whole and piecewise linear regression analysis, establish visibility measurement model, X Axis indicates that measuring and calculating value, Y-axis indicate CJY-1G true value.It is obtained by fitting such as Figure 11, obtaining unary linear regression equation is formula (12):
Y=1.11473x+431.48036 (12)
Its fitting coefficient R2=0.96296.It is 80.77% that model accuracy rate is obtained after measuring and calculating comparison.
Its reason is analyzed, is the scatter plot as shown in Figure 11, in low visibility and number of days phase when compared with high-visibility To less, therefore this partial data may being affected for linear regression fit.Therefore, we attempt the several hundred of measuring and calculating Its visibility carries out regression analysis stage by stage.If Figure 12, Figure 13, Figure 14 are linear visibility linear regression stage by stage.By quasi- It is formula (13) that conjunction, which obtains regression model:
It is compared through measuring and calculating, obtaining segmented model accuracy rate is 91.35%.
By the above-mentioned means, can be clapped picture pick-up device the present invention is based on the visibility measurement method of image automatic identification The image for the different scenes taken the photograph, selection is most suitable to go fog coefficient, carries out the automatic selection and visibility measuring and calculating of target area;This The measurement method of invention also have the characteristics that it is practical, at low cost, influenced by Changes in weather it is small;Experiment proves that accuracy It is higher.

Claims (8)

1. the visibility measurement method based on image automatic identification, which is characterized in that be specifically implemented according to the following steps:
Step 1, original image of the acquisition containing target, same date, identical place, same shooting angle do not acquire different weather feelings The target image of condition is as test image;
Maximum kind variance method is respectively adopted to original image, test image progress Threshold segmentation in step 2;
Step 3 continuously expands the test image after Threshold segmentation using diamond shape probe, obtains expanding image;
Step 4, it is counter to expanding image coloured, utilize centroid principle, " puppet " of more connected regions after searching continuous expansion Mass center calibration;
Step 5, the center by " puppet " mass center as quasi- selection multiple target region, mark multiple target region to be selected;
Step 6, original image after singulation dark effect picture in, find in corresponding label multiple target region to be selected " puppet " mass center light intensity Ic(y) maximum central point, the final goal region that the region where the central point is calculated as visibility, Selection suitably removes fog coefficient ω, carries out the calculating of visibility.
2. according to claim 1 based on the visibility measurement method of image automatic identification, which is characterized in that step 2 is specific Process are as follows: the Threshold segmentation of original image, test image is carried out using maximum kind variance method;
It is g (x, y) after Threshold segmentation if the pixel of original image or test image is f (x, y), then:
If the size of original image or test image is M × N, segmentation threshold is denoted as T.
3. according to claim 2 based on the visibility measurement method of image automatic identification, which is characterized in that the segmentation threshold The calculating process of value are as follows: target image vegetarian refreshments is in the ratio of entire imageImage averaging gray scale is μ0;Background pixel point exists The ratio of entire image isImage averaging gray scale is μ1;Number of pixels of the gray value of pixel less than threshold value T is denoted as in image N0, gray scale is denoted as N greater than the number of pixels of threshold value T1;The overall average gray scale of image is denoted as μ, and inter-class variance is denoted as g;Then it can be obtained Following formula:
The threshold value T for making the maximum threshold value of inter-class variance g, as acquiring is obtained using traversal.
4. according to claim 1 based on the visibility measurement method of image automatic identification, which is characterized in that step 3 is specific Process are as follows:
" probe ", which is done, using structural element collects target image information;The conditioned reflex about origin is first carried out to structural element S Obtain SV, then in test image after segmentation by SVTranslate x, those translation after at least one nonzero element of test image X When intersection, corresponding origin position set is dilation operation result;
Then expansion results are write as formula (3):
Structural element S is diamond shape probe;
Structural element S continuously expands repeatedly, obtains expanding image.
5. according to claim 1 based on the visibility measurement method of image automatic identification, which is characterized in that step 4 is specific Process are as follows:
Step 4.1 searches the pixel of the image after continuous expansion one by one, and the part all the points that pixel is 0 are assigned a value of 1, all the points that pixel is 1 are assigned a value of 0, obtain anti-rendered image;
Step 4.2 utilizes bwlabel function to anti-rendered image, finds all connected domains separated on anti-rendered image, then benefit " puppet " mass center is searched with the principle that mass center is searched, mass center searches principle such as formula (5):
Wherein, M10It is the summation of all the points x coordinate on object, M01It is the summation of all the points y-coordinate on object, M00It is object Area;
Indicate the position of " puppet " mass center.
6. according to claim 1 based on the visibility measurement method of image automatic identification, which is characterized in that step 5 is specific Process are as follows: by " puppet " mass center as the quasi- center for selecting multiple target region, mark more mesh to be selected using Rectangle rectangular function Mark region, the rectangular area that setting area size is 6 × 6;
‘Position’[x,y,w,h] (6)
In formula (6), x, y are lower-left angular coordinate, and w, h respectively represent wide and high.
7. according to claim 1 based on the visibility measurement method of image automatic identification, which is characterized in that step 6 is specific Process are as follows: Cauchy's Mead daytime visibility measuring principle, such as formula (7):
Wherein VmetIndicate that atmospheric visibility, β indicate that extinction coefficient, ε indicate to select 0.05 than sense threshold value;
Then formula (7) is rewritten as formula (8):
The dark effect picture of original image after being divided by dark principle, wherein dark channel prior principle, such as formula (9):
WhereinIndicate transmissivity, Ic(y) haze image luminous intensity, A are indicatedcIndicate global atmosphere luminous intensity, ω indicates defogging Coefficient, value are 0 < ω < 1;
Meet Lambert-Beer's law between transmissive portions light splitting and incident intensity;
Meet the relationship such as formula (10) between atmospheric extinction coefficient and transmissivity:
T (x)=e-βd(x) (10)
Wherein, β is extinction coefficient, and d is object at a distance from receiving point, can be calculated by marking on map;
According to formula (9), I is chosenc(y) the final goal region that the region where maximum central point is calculated as visibility is chosen Fog coefficient ω suitably is removed, the presence of fog coefficient and visibility is gone to choose rule, convolution (8), (9), (10) can be finally inversed by big Gas visibility.
8. according to claim 1 based on the visibility measurement method of image automatic identification, which is characterized in that the defogging system Several and visibility presence chooses rule as shown in table 1:
Table 1
CN201910690633.2A 2019-07-29 2019-07-29 Visibility measuring method based on automatic image identification Active CN110503634B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910690633.2A CN110503634B (en) 2019-07-29 2019-07-29 Visibility measuring method based on automatic image identification

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910690633.2A CN110503634B (en) 2019-07-29 2019-07-29 Visibility measuring method based on automatic image identification

Publications (2)

Publication Number Publication Date
CN110503634A true CN110503634A (en) 2019-11-26
CN110503634B CN110503634B (en) 2022-07-15

Family

ID=68587662

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910690633.2A Active CN110503634B (en) 2019-07-29 2019-07-29 Visibility measuring method based on automatic image identification

Country Status (1)

Country Link
CN (1) CN110503634B (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111192275A (en) * 2019-12-30 2020-05-22 西安金路交通工程科技发展有限责任公司 Highway fog visibility identification method based on dark channel prior theory
CN114280056A (en) * 2021-12-20 2022-04-05 北京普测时空科技有限公司 Visibility measurement system
CN114720425A (en) * 2022-04-24 2022-07-08 安徽气象信息有限公司 Visibility monitoring system and method based on image recognition

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101901477A (en) * 2010-07-27 2010-12-01 中国农业大学 Method and system for extracting field image edges of plant leaves
CN105931220A (en) * 2016-04-13 2016-09-07 南京邮电大学 Dark channel experience and minimal image entropy based traffic smog visibility detection method
CN107784669A (en) * 2017-10-27 2018-03-09 东南大学 A kind of method that hot spot extraction and its barycenter determine

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101901477A (en) * 2010-07-27 2010-12-01 中国农业大学 Method and system for extracting field image edges of plant leaves
CN105931220A (en) * 2016-04-13 2016-09-07 南京邮电大学 Dark channel experience and minimal image entropy based traffic smog visibility detection method
CN107784669A (en) * 2017-10-27 2018-03-09 东南大学 A kind of method that hot spot extraction and its barycenter determine

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
HE KAIMING,ET AL: "《Single Image Haze Removal Using Dark Channel Prior》", 《IEEE TRANSACTION ON PATTERN ANALYSIS MACHINE INTELLIGENCE》 *
PAN_JINQUAN: "《Matlab形态学图像处理:二值图像分割 标记连通区域和重心位置 删除连通区域》", 《HTTP://BLOG.CSDN.NET/GUYUEALIAN/ARTICLE/DETAILS/71440949》 *
赵瑞: "《基于机器视觉的大气能见度测量方法研究》", 《中国优秀硕士学位论文全文数据库 基础科学辑》 *
赵瑞等: "《基于暗原色先验理论的大气能见度测量》", 《激光与光电子学进展》 *

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111192275A (en) * 2019-12-30 2020-05-22 西安金路交通工程科技发展有限责任公司 Highway fog visibility identification method based on dark channel prior theory
CN114280056A (en) * 2021-12-20 2022-04-05 北京普测时空科技有限公司 Visibility measurement system
CN114720425A (en) * 2022-04-24 2022-07-08 安徽气象信息有限公司 Visibility monitoring system and method based on image recognition
CN114720425B (en) * 2022-04-24 2023-02-21 安徽气象信息有限公司 Visibility monitoring system and method based on image recognition

Also Published As

Publication number Publication date
CN110503634B (en) 2022-07-15

Similar Documents

Publication Publication Date Title
Zhang et al. CCTSDB 2021: a more comprehensive traffic sign detection benchmark
Heinzler et al. Cnn-based lidar point cloud de-noising in adverse weather
CN109816024A (en) A kind of real-time automobile logo detection method based on multi-scale feature fusion and DCNN
CN107516077B (en) Traffic sign information extraction method based on fusion of laser point cloud and image data
CN105373794B (en) A kind of licence plate recognition method
CN110503634A (en) Visibility measurement method based on image automatic identification
CN108596165B (en) Road traffic marking detection method and system based on unmanned plane low latitude Aerial Images
CN105608417B (en) Traffic lights detection method and device
CN106679633B (en) A kind of vehicle-mounted distance-finding system base and method
CN114677554A (en) Statistical filtering infrared small target detection tracking method based on YOLOv5 and Deepsort
CN107729843B (en) Low-floor tramcar pedestrian identification method based on radar and visual information fusion
CN111582339B (en) Vehicle detection and recognition method based on deep learning
CN111340855A (en) Road moving target detection method based on track prediction
CN108764338B (en) Pedestrian tracking method applied to video analysis
CN111898627B (en) SVM cloud microparticle optimization classification recognition method based on PCA
CN103049909B (en) A kind of be focus with car plate exposure method
CN115546170B (en) Fan blade defect positioning method and system based on laser ranging
CN106407951A (en) Monocular vision-based nighttime front vehicle detection method
CN112488083A (en) Traffic signal lamp identification method, device and medium for extracting key points based on heatmap
CN111383286B (en) Positioning method, positioning device, electronic equipment and readable storage medium
CN114998308A (en) Defect detection method and system based on photometric stereo
CN113312987B (en) Recognition method based on unmanned aerial vehicle road surface crack image
CN114565848A (en) Liquid medicine level detection method and system in complex scene
CN116228756B (en) Method and system for detecting bad points of camera in automatic driving
CN111353481A (en) Road obstacle identification method based on laser point cloud and video image

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant