AU2019395238A1 - Method for restoring underground image on basis of ray reverse tracing technology - Google Patents

Method for restoring underground image on basis of ray reverse tracing technology Download PDF

Info

Publication number
AU2019395238A1
AU2019395238A1 AU2019395238A AU2019395238A AU2019395238A1 AU 2019395238 A1 AU2019395238 A1 AU 2019395238A1 AU 2019395238 A AU2019395238 A AU 2019395238A AU 2019395238 A AU2019395238 A AU 2019395238A AU 2019395238 A1 AU2019395238 A1 AU 2019395238A1
Authority
AU
Australia
Prior art keywords
rays
image
light intensity
underground
intersection point
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
AU2019395238A
Other versions
AU2019395238B2 (en
Inventor
Xiaoyu Li
Bowen Liu
Peng Liu
Xuliang LU
Lei SI
Chao TAN
Zhongbin WANG
Honglin Wu
Yue Wu
Hongya Zhou
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
China University of Mining and Technology CUMT
Xuzhou Goldfluid Hydraulic Technology Development Co Ltd
Original Assignee
China University of Mining and Technology CUMT
Xuzhou Goldfluid Hydraulic Technology Development Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by China University of Mining and Technology CUMT, Xuzhou Goldfluid Hydraulic Technology Development Co Ltd filed Critical China University of Mining and Technology CUMT
Publication of AU2019395238A1 publication Critical patent/AU2019395238A1/en
Application granted granted Critical
Publication of AU2019395238B2 publication Critical patent/AU2019395238B2/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/80Camera processing pipelines; Components thereof
    • H04N23/81Camera processing pipelines; Components thereof for suppressing or minimising disturbance in the image signal generation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N25/00Circuitry of solid-state image sensors [SSIS]; Control thereof
    • H04N25/60Noise processing, e.g. detecting, correcting, reducing or removing noise
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/18Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Image Generation (AREA)
  • Geophysics And Detection Of Objects (AREA)
  • Image Processing (AREA)
  • Length Measuring Devices By Optical Means (AREA)

Abstract

METHOD FOR RESTORING UNDERGROUND IMAGE ON BASIS OF RAY REVERSE TRACING TECHNOLOGY Abstract The present invention discloses a method for restoring an underground image on the basis of a ray reverse tracing technology. The method includes the following steps: setting an underground camera as a light source emitting point, and emitting rays into an underground scene; recording all intersection points of all rays and underground objects, and calculating one intersection point closest to a view point in the intersection points; calculating a direction of rays newly generated after the rays are reflected and refracted by the objects in a position of the intersection point; respectively tracing the rays newly generated; recording rays irradiated onto a view plane after a strong light source emitted from the position of the camera is reflected or refracted for three times, and calculating light intensity of the rays; converting the light intensity into a pixel value by a camera CCD photosensitive element; and eliminating the pixel value of strong light emitted from the camera in an image finally shown on the view plane to obtain an image after strong light source influence elimination. The method provided by the present invention can effectively eliminate the interference of a strong light source, restore the underground image, and ensure the smooth proceeding of underground work and the life safety of operators.

Description

METHOD FOR RESTORING UNDERGROUND IMAGE ON BASIS OF RAY REVERSE TRACING TECHNOLOGY FIELD OF THE INVENTION
[0001] The present invention belongs to the field of underground image restoration, and particularly relates to a method for restoring an underground image on the basis of a ray reverse tracing technology.
DESCRIPTION OF RELATED ART
[0002] A ray tracing technology is a method for showing a three-dimensional (3D) image on a two-dimensional (2D) screen, is widely applied to games and computer graphics at present, and brings a more vivid effect to people. A light source is supposed as a point light source capable of randomly emitting tens of thousands of rays to surroundings, and those rays are reflected, refracted or absorbed (attenuated) or generate fluorescence after touching different objects. Ray tracing is a general technology from geometrical optics, and a ray passing path model is obtained by tracing rays generating interaction effects with an optical surface. However, tens of thousands of rays exist, and the rays after reflection, refraction, absorption and fluorescence generation are countless, so that the calculation amount of ray positive tracing is great. Therefore, a ray reverse tracing method gradually comes into people's sight. The calculation amount is greatly reduced if a camera lens is used as a light source emitting point and only the part of rays entering a view plane are calculated.
[0003] Due to a fact that most explosion-proof cameras used underground at present are black and white cameras, special underground environment of a coal mine, all-weather artificial illumination, and influence by factors such as dust and dampness, an underground video has the characteristics of low image illuminance and nonuniform illumination distribution, and this special conditions cause low quality of the collected video and poor resolution of the video. When a strong light source such as a safety mine lamp occurs in a view filed of a mine camera, a collected image will have a dazzle light phenomenon, so that the quality of the video image is greatly reduced, and occurrence of safety accidents may be caused. Therefore, application of the ray reverse tracing technology to underground image restoration for image readability improvement is of great importance.
SUMMARY OF THE INVENTION
Technical Problem
[0004] Aiming at the above problems, the present invention provides a method for restoring an underground image on the basis of a ray reverse tracing technology. By aiming at a phenomenon that under the conditions of low illuminance and much dust in a coal mine, a suddenly occurring strong light source may interfere an original video image, so that black and white level contrast of the monitoring image is too great, and information in the video image cannot be recognized, a ray reverse tracing method is used, and a pixel value of the strong light source in a view plane is eliminated, so that interference of the strong light source to the original video image is eliminated.
Technical Solution
[0005] In order to achieve a goal of the present invention, the present invention adopts a technical solution that a method for restoring an underground image on the basis of a ray reverse tracing technology includes the following steps:
[0006] step 1: supposing an underground camera as a light source emitting point, i.e., a view point, and emitting rays into an underground scene;
[0007] step 2: recording all intersection points of all rays and underground objects, and calculating one intersection point closest to the view point in the intersection points;
[0008] step 3: according to illumination, object materials and a normal direction, calculating light intensity of reflection rays or refraction rays in the closest intersection point determined in the step 2;
[0009] step 4: calculating a direction of rays newly generated after the rays are reflected and refracted by the objects in a position of the intersection point;
[0010] step 5: tracing the rays newly generated in the step 4, and judging whether the third time reflection or refraction rays are emitted onto a view plane right in front of a safety mine lamp or not; if so, calculating the third time reflection light intensity and/or refraction light intensity; and otherwise, returning to the step 2 to redetermine the closest intersection point, and repeating the step 3 to the step 5;
[0011] step 6: converting the light intensity in the step 5 into a pixel value through a camera CCD photosensitive element, emitting rays obtained after the third time reflection and/or refraction of the rays emitted from the camera onto the view plane, and performing imaging on the view plane; and
[0012] step 7: eliminating the pixel value of strong light emitted from the camera in an image finally shown on the view plane to obtain an image after strong light source influence elimination.
[0013] In the step 3, the light intensity of the reflection rays or refraction rays in the closest intersection point determined in the step 2 is calculated according to the following method:
[0014] calculating the light intensity of the reflection rays in the position of the intersection point through a formula (1):
I,= 1aKa + li(N - L)di(K1Ra +KR) ()
[0015] wherein Irrepresents the light intensity of the reflection rays; IaKa represents an influence value of environment light in the position of the intersection point; Ii represents the light intensity of incident light; Krepresents a specular reflectivity coefficient; Ks represents a diffuse reflectivity coefficient; Rd represents specular reflectivity; Rs represents diffuse reflectivity; and N, L and di respectively represent an object surface normal vector, a ray direction unit vector and a solid angle;
[0016] or calculating the light intensity of the refraction rays in the position of the intersection point through a formula (2):
It = (cos 0 2 /cos 01)(It - Ir) (2)
[0017] wherein It represents the light intensity of the refraction rays, and 01 and 02 are an incidence angle and a refraction angle.
[0018] In the step 5, the rays newly generated in the step 4 are traced according to the following methods:
[0019] (1) if the rays do not intersect with any object, giving up the tracing; if the intersection point is on a nontransparent object, only calculating the light intensity of the reflection rays; if the intersection point is on a transparent object, calculating the light intensity of the reflection rays and the light intensity of the refraction rays, and tracing the rays obtained by reflecting or refracting the initial rays for three times; if the rays obtained by reflecting or refracting the initial rays for three times are emitted onto the view plane right in front of the safety mine lamp, calculating the light intensity of the rays; and if not, giving up the tracing, and entering the step (2); and
[0020] (2) if all reflection and refraction rays generated by the initial rays are not emitted onto the view plane right in front of the safety mine lamp, determining an intersection point second closest to the view point in the intersection points of the initial rays and the objects; repeating the step (1); if the second closest intersection point does not meet conditions, sequentially calculating the next closest intersection point until the intersection point found meets the conditions.
[0021] In the step 7, the pixel value of the strong light emitted from the camera is eliminated in the image finally shown on the view plane to obtain the image after the strong light source influence elimination according to the following methods:
[0022] besides light of the safety mine lamp simulated by light emitted from the camera underground, i.e., a light source A, other artificial lamp light, i.e., a light source B also exists, and meanwhile, the environment light, i.e., an artificial light source C also exists.
[0023] When the third time reflection rays and/or refraction rays are irradiated onto the view plane, the image on the view plane is shown as the following formula:
P(x, y) = R (x, y) -S(x, y) - L (x, y) (3)
[0024] wherein P(xy) represents the image finally shown on the view plane; R(xy) represents an image shown on the view plane when the camera does not emit light, i.e., the image shown on the view plane when the light source B and the light source C are overlapped; S(xy) represents an image on the view plane when only the camera emits light; and L(xy) represents an image of the environment light, i.e., the light source C, on the view plane.
[0025] 1I(x, y) = R (x, y) S(x, y) • (4) is set,
[0026] the logarithm is taken at both sides to obtain in P(x, y) = In I(x, y) + In L (x, y) (5),
[0027] and the environment light L(xy) is shown as follows through P(xy) and Gaussian kernel convolution of a Gaussian function G(xy):
L(x, y)= P(x, y) * G (x, y) (6)
[0028] wherein G(x, y) =e c2 ,
[0029] C represents a Gaussian surrounding scale, and A is one scale, and enables ff G(x, y) dx dy = 1 to be always true. Through the formulas (4), (5) and (6), it can be
obtained:
In R(x,y) =In P(x,y) - ln(P(x,y)* G(x,y)) - InS(x,y)
[0030] wherein S'(x,y) = enR(Xy is set,
[0031] and S'(xy) is the image after the strong light source influence elimination.
Advantageous Effect
[0032] Compared with the prior art, the technical solution of the present invention has the following beneficial technical effects:
[0033] the present invention changes a conventional thought on image processing by utilizing the ray reverse tracing. Conventional methods mostly use methods of linear conversion, gamma correction, histogram equalization, anti-sharpening mask, homomorphic filtering, tone mapping, dark channel algorithm and the like for the condition of sudden occurrence of a strong light source, and a processing effect is not obvious. The ray reverse tracing technology can effectively eliminate the interference of the strong light source, restore the original underground image, and ensure smooth proceeding of underground work and life safety of operators.
BRIEF DESCRIPTION OF THE DRAWINGS
[0034] Fig. 1 is a schematic diagram of an opened solid included angle diI of a unit area towards a light source;
[0035] Fig. 2 is a schematic diagram of reflection and refraction receiving of ray reverse tracing of the present invention; and
[0036] Fig. 3 is a process of eliminating strong light source interference by ray reverse tracing of the present invention.
DETAILED DESCRIPTION OF THE INVENTION
The technical solutions of the present invention are further described below with reference to the accompanying drawings and embodiments.
[0037] According to a method for restoring an underground image on the basis of a ray reverse tracing technology of the present invention, by aiming at a phenomenon that under the conditions of low illuminance, much dust and high dampness in a coal mine, a suddenly occurring strong light source may interfere an original video image, so that black and white level contrast of the monitoring image is too great, and information in the video image cannot be recognized, a ray reverse tracing method is used, and a pixel value of the strong light source in a view plane is eliminated, so that interference of the strong light source to the original video image is eliminated. As shown in Fig. 3, a process of eliminating strong light source interference by ray reverse tracing of the present invention concretely includes steps as follows.
[0038] Step 1: an underground camera is supposed as a light source emitting point, i.e., a view point, and rays are emitted into an underground scene. Intensity of the rays is equal to light intensity of rays emitted from a safety mine lamp.
[0039] Step 2: all intersection points of all rays and underground objects are recorded, and an intersection point closest to the view point in the intersection points is calculated.
[0040] Step 3: according to illumination, object materials and a normal direction, light intensity of reflection rays or refraction rays in the closest intersection point determined in the step 2 is calculated.
[0041] The light intensity of the reflection rays in the position of the intersection point is calculated through a formula (1):
= laK, ++,Ii (N • L)d IC(K R1 + KSR)
.
[0042] Irrepresents the light intensity of the reflection rays. aKa represents an influence value of environment light in the position of the intersection point. Iirepresents the light intensity of incident light. Kd represents a specular reflectivity coefficient. Ks represents a diffuse reflectivity coefficient. Rd represents specular reflectivity. Rs represents diffuse reflectivity. N, L and respectively represent an object surface normal vector, a ray direction unit vector and a solid angle. As shown in Fig. 1, a horizontal axis direction represents an object surface; a longitudinal axis direction represents a normal vector direction of the object surface; and the solid angle is defined as an angle of a projection area of an underground object on a spherical surface to an observation point after the three-dimensional spherical surface is formed by using the camera as the observation point.
[0043] Or, the light intensity of the refraction rays in the position of the intersection point is calculated through a formula (2):
It = (cos 0 2 /cos 01)(I - 1r) (2).
[0044] Itrepresents the light intensity of the refraction rays, and 0O and 02 are an incidence angle and a refraction angle.
[0045] A light and shade effect is only determined jointly by a first intersected object surface normal direction, a material, a view point and an illumination direction, and illumination intensity, and the second layer and deeper layer rays are not considered for ray projection, so that shade, reflection, refraction and fluorescence effects do not exist.
[0046] Step 4: a direction of rays newly generated after the rays are reflected and refracted by the objects in a position of the intersection point is calculated. The direction of the rays newly generated is jointly determined by an incidence light direction, an object surface normal direction and media.
[0047] Step 5: the rays newly generated in the step 4 is traced, and whether the third time reflection or refraction rays are emitted onto the view plane right in front of a safety mine lamp or not is judged; if so, the third time reflection light intensity and/or refraction light intensity is calculated; and otherwise, flow returns to the step 2 to redetermine the closest intersection point, and the step 3 to the step 5 are repeated.
[0048] After the rays are emitted from the camera, ray tracing is performed as follows: the rays may intersect with transparent objects and nontransparent objects or may not intersect with any object in the scene after being emitted from the camera.
[0049] (1) If the rays do not intersect with any object, tracing is given up. If the intersection point is on the nontransparent object, only the light intensity of the reflection rays is calculated. If the intersection point is on the transparent object, the light intensity of the reflection rays and the light intensity of refraction rays are calculated, and the rays obtained by reflecting or refracting the initial rays for three times are traced. If the rays obtained by reflecting or refracting the initial rays for three times are emitted onto the view plane right in front of the safety mine lamp, the light intensity of the rays is calculated. If not, the tracing is given up, and flow enters the step (2).
[0050] (2) If all reflection and refraction rays generated by the initial rays are not emitted onto the view plane right in front of the safety mine lamp, an intersection point second closest to the view point in the intersection points of the initial rays and the objects. The step (1) is repeated. If the second closest intersection point does not meet conditions, the next closest intersection point is sequentially calculated until the intersection point found meets the conditions.
[0051] As shown in Fig. 2, an example for calculating the light intensity of the reflection rays and the light intensity of refraction are given concretely as follows.
[0052] It is supposed that in the underground scene, the camera is positioned in the position of the view point; light is emitted from the camera; and a transparent object O1 and a nontransparent object 02 exist. Firstly, an initial ray E is emitted from the view point and intersects with the 0 at P1 , and a reflection ray R, and a refraction ray T are generated. Light
intensity of the Ri conforms to a formula "'1Ka, + I(N1 . Ll)diH(Kd, R, 1 + Kl, R, 1
) and since the Ri no longer intersect with other objects, tracing is ended. Light intensity of the
Ti conforms to a formula Ir1 = (cos 02/COS 11)(1 U- Iri). The Ti intersects inside the 0 at
P2 , and a reflection ray R 2 and a refraction ray T2 are generated. Light intensity of the R 2
conforms to a formula 1r2 - Ia2 Ka2 + Ir,(N2 - L2 )d7i (K Ra, 2 + K1 2Rs2 ), and light intensity
of the T2 conforms to a formula 42= (Cos 0 4/COS3 3 )(Itr -1rz). Recursion may be
continuously performed to trace the R 2 and the T2. For example, the T2 and 03 intersect at P3
, and since the 03 is nontransparent, only a reflection ray R3 is generated. Light intensity of the
R 3 conforms to a formula r3 = asKa, + It (N 3 - L 3 )d7iN(KdRd 3 + K 3 RS 3 ). The R 3 finally
enters the view plane.
[0053] 01 and 02 are an incidence angle and a reflection angle at the position P. 03 and 04 are an incidence angle and a reflection angle at the position P 2 . 'a 1 Kai represents an influence
value of the environment light at the position Pi. Ia 2 Ka2 represents an influence value of the
environment light at the position P2 . 'a 3 Ka3 represents an influence value of the environment
light at the position P 3 . i represents light intensity of the ray E, i.e., the light intensity of incidence light of the initial ray. Kl, Kd 2 , andKa respectively represent specular reflectivity
coefficients at the positions Pi, P 2 and P3 . KS1 , Ks ,2 and Ks3 respectively represent diffuse
reflectivity coefficients at the positions Pi, P2 and P3 . Rd, R 2 , and R d3 respectively represent
specular reflectivity at the positions Pi, P2 and P3 . RS1 , RS 2 , and Rs3 respectively represent diffuse reflectivity at the positions P, P2 and P3 . N, N2 , and N3 respectively represent normal vectors of the object surface at the positions Pi, P2 and P3 . Li, L 2 and L3 respectively represent unit vectors of ray directions of the initial ray E, the refraction ray Ti and the refraction ray T2. day-, d-, and d7,- respectively represent solid angles generated at the positions Pi, P2 and
P3 .
[0054] Step 6: the light intensity in the step 5 is converted into a pixel value through a camera CCD photosensitive element. The rays obtained after the third time reflection and/or refraction of the rays emitted from the camera are emitted onto the view plane. Imaging is performed on the view plane.
[0055] Step 7: the pixel value of strong light emitted from the camera is eliminated in an image finally shown on the view plane to obtain an image after strong light source influence elimination according to the methods as follows.
[0056] Besides light of the safety mine lamp simulated by light emitted from the camera underground, i.e., a light source A, other artificial lamp light, i.e., a light source B also exists, and meanwhile, environment light, i.e., an artificial light source C also exists.
[0057] When the third time reflection rays and/or refraction rays are irradiated onto the view plane, the image on the view plane may be shown as the following formula:
P(x,y) = R(x,y) -S(x,y) - L(x,y) (3).
[0058] P(xy) represents the image finally shown on the view plane. R(xy) represents an image shown on the view plane when the camera does not emit light, i.e., the image shown on the view plane when the light source B and the light source C are overlapped. S(xy) represents an image on the view plane when only the camera emits light. L(xy) represents an image of the environment light, i.e., the light source C, on the view plane.
[00591 I(x, y) = R(x,y) - S(x, y) (4) is set,
[0060] the logarithm is taken at both sides to obtain In P(x, y) = In I(x, y) + In L (x, y) (5),
[0061] and the environment light L(xy) may be shown as follows through P(xy) and Gaussian kernel convolution of a Gaussian function G(xy):
L (x, y) P(x, y) * G(x, y) (6) -(x2+2)
[0062] wherein G(x, y) = )e C2
[0063] C represents a Gaussian surrounding scale, and is one scale, and enables ff G(x, y) dx dy = 1 to be always true. Through the formulas (4), (5) and (6), it can be
obtained:
inR(x,y) = InP(x,y) - in(P(x,y) * G(x,y)) - inS(x,y)
[0064] wherein S'(x,y) = elnR(Xy) is set,
[0065] and S'(xy) is the image after the strong light source influence elimination.
[0066] The present invention utilizes the ray reverse tracing technology. Under the condition of greatly reducing the calculation amount of the ray tracing, the dazzle light phenomenon of the strong light source on the low-illuminance underground video image is effectively reduced, so that the effect of restoring the video image is achieved.

Claims (4)

What is claimed is:
1. A method for restoring an underground image on the basis of a ray reverse tracing technology, comprising the following steps:
step 1: supposing an underground camera as a light source emitting point, i.e., a view point, and emitting rays into an underground scene;
step 2: recording all intersection points of all rays and underground objects, and calculating one intersection point closest to the view point in the intersection points;
step 3: according to illumination, object materials and a normal direction, calculating light intensity of reflection rays or refraction rays in the closest intersection point determined in the step 2;
step 4: calculating a direction of rays newly generated after the rays are reflected and refracted by the objects in a position of the intersection point;
step 5: tracing the rays newly generated in the step 4, and judging whether the third time reflection or refraction rays are emitted onto a view plane right in front of a safety mine lamp or not; if so, calculating the third time reflection light intensity and/or refraction light intensity; and otherwise, returning to the step 2 to redetermine the closest intersection point, and repeating the step 3 to the step 5;
step 6: converting the light intensity in the step 5 into a pixel value through a camera CCD photosensitive element, emitting rays obtained after the third time reflection and/or refraction of the rays emitted from the camera onto the view plane, and performing imaging on the view plane; and
step 7: eliminating the pixel value of strong light emitted from the camera in an image finally shown on the view plane to obtain an image after strong light source influence elimination.
2. The method for restoring the underground image on the basis of the ray reverse tracing technology according to claim 1, wherein in the step 3, the light intensity of the reflection rays or refraction rays in the closest intersection point determined in the step 2 is calculated according to the following method:
calculating the light intensity of the reflection rays in the position of the intersection point through a formula (1):
I,= laKa+I(N L)di(KdRd+KsR,) wherein rrepresents the light intensity of the reflection rays; IaKa represents an influence value of environment light in the position of the intersection point; Iirepresents the light intensity of incident light; Kd represents a specular reflectivity coefficient; Ks represents a diffuse reflectivity coefficient; Rdrepresents specular reflectivity; Rs represents diffuse reflectivity; and N, L and d-respectively represent an object surface normal vector, a ray direction unit vector and a solid angle; or calculating the light intensity the refraction rays in the position of the intersection point through a formula (2):
It = (cos0 2/cos 01)(I - Ir) (2)
wherein It represents the light intensity of the refraction rays, and 0 and 02 are an incidence angle and a refraction angle.
3. The method for restoring the underground image on the basis of the ray reverse tracing technology according to claim 1 or 2, wherein in the step 5, the rays newly generated in the step 4 are traced according to the following methods:
(1) if the rays do not intersect with any object, giving up the tracing; if the intersection point is on a nontransparent object, only calculating the light intensity of the reflection rays; if the intersection point is on a transparent object, calculating the light intensity of the reflection rays and the light intensity of the refraction rays, and tracing the rays obtained by reflecting or refracting the initial rays for three times; if the rays obtained by reflecting or refracting the initial rays for three times are emitted onto the view plane right in front of the safety mine lamp, calculating the light intensity of the rays; and if not, giving up the tracing, and entering the step (2); and
(2) if all reflection and refraction rays generated by the initial rays are not emitted onto the view plane right in front of the safety mine lamp, determining an intersection point second closest to the view point in the intersection points of the initial rays and the objects; repeating the step (1); if the second closest intersection point does not meet conditions, sequentially calculating the next closest intersection point until the intersection point found meets the conditions.
4. The method for restoring the underground image on the basis of the ray reverse tracing technology according to claim 1 or 2, wherein in the step 7, the pixel value of the strong light emitted from the camera is eliminated in the image finally shown on the view plane to obtain the image after the strong light source influence elimination according to the following method: when the third time reflection rays and/or refraction rays are irradiated onto the view plane, the image on the view plane is shown as the following formula:
P(x, y) = R(x, y) - S(x, y) - L (x, y) (3)
wherein P(xy) represents the image finally shown on the view plane; R(xy) represents an image shown on the view plane when the camera does not emit light; S(xy) represents an image on the view plane when only the camera emits light; and L(xy) represents an image of the environment light on the view plane:
I(x, y) = R (x, y) -S(x, y) (4) is set,
the logarithm is taken at both sides to obtain in P(x, y) = In I(x, y) + In L (x, y) (5),
and the environment light L(xy) is shown as follows through P(xy) and Gaussian kernel convolution of a Gaussian function G(xy):
L(x,y) = P(x,y) * G(x,y) (6)
whereinG(x, y) = [e c2
C represents a Gaussian surrounding scale, and, Iis one scale; through the formulas (4), (5) and (6), it can be obtained:
In R (x, y) = ln P(x,y) - In(P(x, y) * G(x, y)) - In S(x,y)
wherein S'(x, y) = en'(xy) is set,
and S'(xy) is the image after the strong light source influence elimination.
AU2019395238A 2019-01-04 2019-06-18 Method for restoring underground image on basis of ray reverse tracing technology Active AU2019395238B2 (en)

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
CN201910006766.3A CN109862209B (en) 2019-01-04 2019-01-04 Method for restoring underground image based on light ray inverse tracking technology
CN2019100067663 2019-01-04
PCT/CN2019/091631 WO2020140397A1 (en) 2019-01-04 2019-06-18 Method for restoring downhole image based on reverse ray tracing technology

Publications (2)

Publication Number Publication Date
AU2019395238A1 true AU2019395238A1 (en) 2020-07-23
AU2019395238B2 AU2019395238B2 (en) 2021-11-25

Family

ID=66893940

Family Applications (1)

Application Number Title Priority Date Filing Date
AU2019395238A Active AU2019395238B2 (en) 2019-01-04 2019-06-18 Method for restoring underground image on basis of ray reverse tracing technology

Country Status (5)

Country Link
CN (1) CN109862209B (en)
AU (1) AU2019395238B2 (en)
CA (1) CA3079552C (en)
RU (1) RU2742814C9 (en)
WO (1) WO2020140397A1 (en)

Families Citing this family (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109862209B (en) * 2019-01-04 2021-02-26 中国矿业大学 Method for restoring underground image based on light ray inverse tracking technology
CN114286375B (en) * 2021-12-16 2023-08-18 北京邮电大学 Mobile communication network interference positioning method
CN116051450B (en) * 2022-08-15 2023-11-24 荣耀终端有限公司 Glare information acquisition method, device, chip, electronic equipment and medium
CN116681814B (en) * 2022-09-19 2024-05-24 荣耀终端有限公司 Image rendering method and electronic equipment

Family Cites Families (19)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6005916A (en) * 1992-10-14 1999-12-21 Techniscan, Inc. Apparatus and method for imaging with wavefields using inverse scattering techniques
US7389041B2 (en) * 2005-02-01 2008-06-17 Eastman Kodak Company Determining scene distance in digital camera images
US7764230B2 (en) * 2007-03-13 2010-07-27 Alcatel-Lucent Usa Inc. Methods for locating transmitters using backward ray tracing
WO2009086836A1 (en) * 2008-01-11 2009-07-16 Danmarks Tekniske Universitet A touch-sensitive device
WO2011066275A2 (en) * 2009-11-25 2011-06-03 Massachusetts Institute Of Technology Actively addressable aperture light field camera
KR101395255B1 (en) * 2010-09-09 2014-05-15 한국전자통신연구원 Apparatus and method for analysing propagation of radio wave in radio wave system
RU125335U1 (en) * 2012-11-07 2013-02-27 Общество с ограниченной ответственностью "Артек Венчурз" DEVICE FOR MONITORING LINEAR SIZES OF THREE-DIMENSIONAL OBJECTS
US9041914B2 (en) * 2013-03-15 2015-05-26 Faro Technologies, Inc. Three-dimensional coordinate scanner and method of operation
KR101716928B1 (en) * 2013-08-22 2017-03-15 주식회사 만도 Image processing method for vehicle camera and image processing apparatus usnig the same
JP2015132953A (en) * 2014-01-10 2015-07-23 キヤノン株式会社 Image processor and method thereof
US9311565B2 (en) * 2014-06-16 2016-04-12 Sony Corporation 3D scanning with depth cameras using mesh sculpting
WO2016002578A1 (en) * 2014-07-04 2016-01-07 ソニー株式会社 Image processing device and method
US9977644B2 (en) * 2014-07-29 2018-05-22 The University Of North Carolina At Chapel Hill Methods, systems, and computer readable media for conducting interactive sound propagation and rendering for a plurality of sound sources in a virtual environment scene
KR101592793B1 (en) * 2014-12-10 2016-02-12 현대자동차주식회사 Apparatus and Method for Correcting Image Distortion
KR20160071774A (en) * 2014-12-12 2016-06-22 삼성전자주식회사 Apparatus, Method and recording medium for processing image
EP3443735A4 (en) * 2016-04-12 2019-12-11 Quidient, LLC Quotidian scene reconstruction engine
CN106231286B (en) * 2016-07-11 2018-03-20 北京邮电大学 A kind of three-dimensional image generating method and device
CN109118531A (en) * 2018-07-26 2019-01-01 深圳大学 Three-dimensional rebuilding method, device, computer equipment and the storage medium of transparent substance
CN109862209B (en) * 2019-01-04 2021-02-26 中国矿业大学 Method for restoring underground image based on light ray inverse tracking technology

Also Published As

Publication number Publication date
RU2742814C9 (en) 2021-04-20
CA3079552A1 (en) 2020-07-04
CN109862209B (en) 2021-02-26
WO2020140397A1 (en) 2020-07-09
AU2019395238B2 (en) 2021-11-25
CA3079552C (en) 2021-03-16
CN109862209A (en) 2019-06-07
RU2742814C1 (en) 2021-02-11

Similar Documents

Publication Publication Date Title
AU2019395238B2 (en) Method for restoring underground image on basis of ray reverse tracing technology
CN107607040B (en) Three-dimensional scanning measurement device and method suitable for strong reflection surface
CN107797560B (en) Visual recognition system and method for robot tracking
Bazeille et al. Color-based underwater object recognition using water light attenuation
CN104101611A (en) Mirror-like object surface optical imaging device and imaging method thereof
CA2281551A1 (en) Outdoor range finder
WO2013109755A1 (en) Segmentation for wafer inspection
TWI537668B (en) Image monitoring apparatus
US11620760B2 (en) Ranging method based on laser-line scanning imaging
CN111833258B (en) Image color correction method based on double-transmissivity underwater imaging model
CN110300976B (en) Eye gaze tracking
CN116386520A (en) Spherical inner display screen correction method and device, electronic equipment and storage medium
Huang et al. Enhancing object detection in the dark using U-Net based restoration module
KR102291162B1 (en) Apparatus and method for generating virtual data for artificial intelligence learning
CN109612408B (en) Method and device for testing emission angle of semiconductor laser and readable storage medium
Bajestani et al. Event-based RGB sensing with structured light
CN101685185A (en) Lens, camera lens component having same and camera module
CN103982857B (en) Optical lens, image pickup device and optical touch system
CN117255964A (en) External illumination with reduced detectability
CN211857172U (en) Long-focus ambient light resistant cloth curtain
CN109671075A (en) Defect inspection method, device, equipment and storage medium
Jin et al. Reliable Image Dehazing by NeRF
US20180341106A1 (en) Method and system for automatically controlling lighting of a person via incident light radiation
TWI753344B (en) Hybrid depth estimation system
Li et al. Combining encoded structured light and photometric stereo for underwater 3D reconstruction

Legal Events

Date Code Title Description
FGA Letters patent sealed or granted (standard patent)