CN106200882B - Sensing element and optical ranging system - Google Patents

Sensing element and optical ranging system Download PDF

Info

Publication number
CN106200882B
CN106200882B CN201510216100.2A CN201510216100A CN106200882B CN 106200882 B CN106200882 B CN 106200882B CN 201510216100 A CN201510216100 A CN 201510216100A CN 106200882 B CN106200882 B CN 106200882B
Authority
CN
China
Prior art keywords
pixel
aperture
shielding layer
subframe
ranging system
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201510216100.2A
Other languages
Chinese (zh)
Other versions
CN106200882A (en
Inventor
许恩峯
刘家佑
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Pixart Imaging Inc
Original Assignee
Pixart Imaging Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Pixart Imaging Inc filed Critical Pixart Imaging Inc
Priority to CN201910293007.XA priority Critical patent/CN110275606B/en
Priority to CN201510216100.2A priority patent/CN106200882B/en
Publication of CN106200882A publication Critical patent/CN106200882A/en
Application granted granted Critical
Publication of CN106200882B publication Critical patent/CN106200882B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Abstract

A kind of sensing element and optical ranging system.The present invention provides a kind of sensing element, comprising the multiple sensor pixel regions arranged with array, wherein each of the multiple sensor pixel region includes the first pixel, the second pixel, the first shielding layer, the second shielding layer and at least one lenticule.Second pixel adjacent first pixel on preset direction.First shielding layer is arranged on first pixel and has the first aperture, wherein the aperture of first aperture increases from the center of first pixel along the preset direction.Second shielding layer is arranged on second pixel and has the second aperture, wherein the shape of second aperture and first aperture is mirrored into symmetrically in the preset direction.At least one described lenticule is arranged on first shielding layer and second shielding layer.

Description

Sensing element and optical ranging system
Technical field
The present invention relates to a kind of sensing elements, and in particular, to it is a kind of can be applied to range measurement or gesture identification Sensing element and optical ranging system.
Background technique
In general, Range Measurement System (Distance measurement system, DMS) would generally use light source, And energy back is reflected by the object to calculate the distance of object using the light beam of light source.Traditional, triangulation location can be used Mode or the technical solution of flight time (Time of flight, TOF) calculate distance, however use aforesaid way institute The cost of cost and the size of system comparatively all can be higher.
In addition, the usually usable 3D rendering in the exploitation basis of gesture (gesture) identification eliminates background image first to extract Foreground object image, wherein such technology will use to two imaging sensors, in this way, the size of gesture identification module Reduction can not be effectively obtained in the same manner as cost.
Based on above-mentioned, the present invention be mainly utilize the technology of detecting phase (phase detection) to obtain 3D rendering, and And it is not necessarily to additional polishing (above-mentioned time-of-flight method needs polishing), and the technical solution that the present invention illustrates can only use single image Sensor can reach the application of detecting distance and gesture identification.
Summary of the invention
In view of this, the present invention proposes a kind of sensing element and optical ranging system, have low cost, size small and The advantages of height detecting accuracy.
The present invention illustrates to provide a kind of optical ranging system, includes lens, sensing element and processing unit.The sensing Part is used to obtain the light across the lens and exports picture frame and multiple sensor pixel regions comprising being arranged with array, Described in each of multiple sensor pixel regions comprising the first pixel, the second pixel, the first shielding layer, the second shielding layer and At least one lenticule.Second pixel adjacent first pixel on preset direction.The first shielding layer setting exists On first pixel and there is the first aperture, wherein the aperture of first aperture is from the center of first pixel along institute State preset direction increase.Second shielding layer is arranged on second pixel and has the second aperture, wherein described the The shape of two apertures and first aperture is mirrored into symmetrically in the preset direction.At least one described lenticule is arranged in institute It states between the first shielding layer and second shielding layer and the lens.The processing unit is used to be generated according to described image frame The first subframe corresponding with first pixel and the second subframe corresponding with second pixel, and according to described first Subframe and second subframe calculate at least one object distance.
Present invention explanation also provides a kind of sensing element, multiple sensor pixel areas comprising array arrangement, wherein described more Each of a sensor pixel region include the first pixel, the second pixel, the first shielding layer, the second shielding layer and at least one Lenticule.Second pixel adjacent first pixel on preset direction.First shielding layer is arranged described first On pixel and there is the first aperture, wherein the aperture of first aperture is from the center of first pixel along the default side To increase.Second shielding layer is arranged on second pixel and has the second aperture, wherein second aperture with The shape of first aperture is mirrored into symmetrically in the preset direction.At least one described lenticule is arranged described first On shielding layer and second shielding layer.
The present invention also provides a kind of optical ranging systems, include lens, sensing element and processing unit.The sensing element For obtaining the light across the lens and exporting picture frame and multiple sensor pixel regions comprising arranging with array, wherein Each of the multiple sensor pixel region includes the first pixel, the second pixel, third pixel, the 4th pixel, the first screening Cover layer, the second shielding layer, third shielding layer, the 4th shielding layer and four lenticules.First shielding layer setting is described the On one pixel and there is the first aperture, wherein the aperture of first aperture is from the center of first pixel along first direction Increase.Second shielding layer is arranged on second pixel and has the second aperture, wherein second aperture and institute The shape for stating the first aperture is mirrored into symmetrically in the first direction.The third shielding layer is arranged on the third pixel And there is third aperture, wherein the aperture of the third aperture increases in a second direction from the center of the third pixel.It is described 4th shielding layer is arranged on the 4th pixel and has the 4th aperture, wherein the 4th aperture and the third aperture Shape be mirrored into symmetrically in the second direction.Four lenticules are separately positioned on first shielding layer, described Between two shielding layers, the third shielding layer and the 4th shielding layer and the lens.The processing unit is used for according to institute State picture frame generate corresponding with first pixel the first subframe, the second subframe corresponding with second pixel and The corresponding third subframe of third pixel and the 4th subframe corresponding with the 4th pixel, and according to first son Frame, second subframe, the third subframe and the 4th subframe estimate at least two object distances.
In order to which above and other objects, features and advantages of the invention can be become apparent from, will hereafter be illustrated appended by cooperation, in detail Carefully it is described as follows.In addition, in the present invention, identical component is indicated with identical symbol, first stated clearly in this.
Figure of description
Fig. 1 is the schematic diagram of the optical ranging system of first embodiment of the invention.
Fig. 2A is the top view of the sensing element of the optical ranging system of Fig. 1.
Fig. 2 B is the partial enlarged view of Fig. 2A.
Fig. 3 is the schematic diagram of the sensing element of the optical ranging system of one embodiment of the invention.
Fig. 4 is the processing unit of first embodiment of the invention according to the schematic diagram of picture frame estimated object distance.
Fig. 5 A is the top view of the sensing element of the optical ranging system of second embodiment of the invention.
Fig. 5 B is the processing unit of second embodiment of the invention according to the schematic diagram of picture frame estimated object distance.
Fig. 6 A~8B is the schematic diagram in the sensor pixel area with aperture of different shapes.
Fig. 9 is the top view of the sensing element of the optical ranging system of third embodiment of the invention.
Specific embodiment
Refering to what is shown in Fig. 1, the schematic diagram of its optical ranging system 1 for showing first embodiment of the invention.The optics is surveyed It include lens 10, sensing element 12 and processing unit 14 away from system 1.In the present embodiment, the optical ranging system 1 is for estimating At least one object distance, for example, when an object 9 enters the effective range of the optical ranging system 1, the optics Range-measurement system 1 can estimate at least one distance of the relatively described optical ranging system 1 of the object 9 (after being specified in).
It should be noted that Fig. 1 only shows an object (that is, described object 9) for illustrating the present embodiment, but this hair It is bright without being limited thereto.In some embodiments, multiple objects of the effective range of the optical ranging system 1 are appeared in all Distance can be estimated.In some embodiments, the object 9 has out-of-flatness surface, and the optical ranging system 1 can be estimated at this time Multiple distances of the relatively described optical ranging system 1 of different location, use acquirement three on the out-of-flatness surface of the object 9 Tie up image information.
The lens 10 are used for optically focused, such as the light that the object 9 is reflected.The lens 10 for example pass through gasket (spacer) fixed range is maintained with the sensing element 12, but not limited to this.In other embodiments, the lens 10 can pass through Shell or support portion and the sensing element 12 maintain the fixed range.In addition, although Fig. 1 only shows a lens 10, The invention is not limited thereto.In other embodiments, the lens 10 can be the lens group comprising multiple lens.
The sensing element 12 is for obtaining the light across the lens 10 and exporting picture frame IF.The sensing element 12 may be, for example, charge coupled cell (CCD) imaging sensor, complementary metal oxide semiconductor (CMOS) imaging sensor or Other can be used for sensing the sensor of light energy.The sensing element 12 includes the multiple sensor pixel regions arranged with array AS, wherein each of the multiple sensor pixel region ASInclude the first pixel P1, the second pixel P2, the first shielding layer S1、 Second shielding layer S2And at least one lenticule LM
With reference to Fig. 1,2A and Fig. 2 B, then illustrate the multiple sensor pixel region ASEach of each member Part.Fig. 2A is the top view of the sensing element 12 of Fig. 1, wherein not showing that at least one lenticule LM.Such as Fig. 2A It is shown, the second pixel of the embodiment of the present invention P2The adjacent first pixel P in preset direction (such as X-direction)1.In addition, this In embodiment, the first pixel P1With the second pixel P2Shape be square, but not limited to this.In other embodiments, The first pixel P1With the second pixel P2Shape can be round or rectangle.Although Fig. 2A shows the sensing element 12 have 6 × 5 sensor pixel region AS, but not limited to this, the multiple sensor pixel region ASThe visual practical application of quantity Depending on.
Fig. 2 B is the partial enlarged view of Fig. 2A, is shown as a sensor pixel region A in thisS.The first shielding layer S1If It sets in the first pixel P1On and have the first aperture O1, wherein the first aperture O1Aperture from first pixel P1Center increase or be increased monotonically along the preset direction.The second shielding layer S2It is arranged in the second pixel P2On And there is the second aperture O2, wherein the second aperture O2With the first aperture O1Shape in the preset direction (such as X Direction) it is mirrored into symmetrically.That is, the second aperture O2Aperture from the second pixel P2Center along described default The opposite direction in direction increases.It should be noted that the first shielding layer S1And the second shielding layer S2It is to be worn for covering Cross the lenticule LMA part of light, and not shielded another part light then may pass through the first shielding layer S1 The first aperture O1And the second shielding layer S2The second aperture O2And arrive at the first pixel P1And described Two pixel P2
It is understood that the first shielding layer S of Fig. 2 B1With the first aperture O1The first area summation be equal to The first pixel P1Area;The second shielding layer S2With the second aperture O2Second area summation be equal to described the Two pixel P2Area, but not limited to this.In other embodiments, the first area summation and second area summation difference Slightly larger than the first pixel P1And the second pixel P2Area to avoid light leakage.
The first shielding layer S1And the second shielding layer S2First of metal layer in CMOS technology be can be to the tenth Any two layers between road metal layer, or with other can the material of shading formed.
In one embodiment, the first shielding layer S is formed with metal material1And the second shielding layer S2Later, usually It can be in the first shielding layer S1And the second shielding layer S2On cover insulating layer or protective layer (passivation layer).At this point, the insulating layer or the protective layer are preferably that translucent material is formed, so that the first aperture O1And institute State the second aperture O2With high transmittance.It is understood that the insulating layer or the protective layer can prevent dust from entering institute State the first pixel P1And the second pixel P2To avoid influence light receiving efficiency.
In the present embodiment, although Fig. 1 shows the first shielding layer S1And the second shielding layer S2With first pixel P1And the second pixel P2A distance away (such as by protective layer or dielectric layer to form the distance), described first Shielding layer S1And the second shielding layer S2It is preferably close to the first pixel P1And the second pixel P2.Other embodiments In, the first shielding layer S1And the second shielding layer S2Coating or the first pixel P can be directly overlayed respectively1And institute State the second pixel P2On.
At least one lenticule LMIt is arranged in the first shielding layer S1And the second shielding layer S2With the lens Between 10, such as shown in Fig. 1, the multiple sensor pixel region ASEach of include two lenticule LM, and described two A lenticule LMRespectively with the first pixel P1And the second pixel P2It is corresponding.At this point, passing through described two lenticule LM And the first aperture O of mirror symmetry1And the second aperture O2Configuration, the first pixel P1And second pixel P2The incident beam of out of phase can be respectively received to carry out detecting phase.
It should be noted that by above-mentioned lenticule and the configuration of aperture, in the first pixel P1And second picture Plain P2When receiving incident beam, close to the first pixel P1And the second pixel P2The received incident light of center institute difference Between phase difference be not obvious;Opposite, close to the first pixel P1And the second pixel P2Edge receive respectively Incident light between phase difference it is then larger.Therefore, for the first aperture O1, with the first pixel P1Edge is opposite The aperture is preferably more than and the first pixel P1The opposite aperture in center.That is, the first aperture O1's The aperture is from the first pixel P1Center increases along the preset direction.Accordingly, the accuracy of detecting phase can be promoted.
It is understood that due to the first aperture O1The aperture from the first pixel P1Center is along described pre- Set direction increases, the first aperture O1Area can be less than the first pixel P1The half of area, such as shown in Fig. 2 B.Together When, the first aperture O1And the second aperture O2Area should be greater than preset area so that acquired in the sensing element 12 Described image frame IF remain with acceptable signal-to-interference ratio (SNR).Preferably, the first aperture O1Area be described first Pixel P1The 5%~45% of area.
Due to the lenticule LMWith symmetrical structure, in other embodiments, the multiple sensor pixel region ASIn it is every One can only include a lenticule LM.At this point, the lenticule LMSimultaneously with the first aperture O1And the second aperture O2 It is corresponding, as shown in Figure 3.In addition, the lenticule LMProtective layer can be formed between the shielding layer.
It please also refer to Fig. 1,2A and Fig. 4, wherein Fig. 4 is the processing unit 14 according to picture frame IF estimated object distance Schematic diagram.The processing unit 14 is, for example, digital signal processor (DSP) or processing circuit, and is electrically connected the sensing Element 12.Described image frame IF (such as corresponding with the 6 × 10 of Fig. 2A pixel array) is exported extremely in the sensing element 12 After the processing unit 14, the processing unit 14 is used to be generated and the first pixel P according to described image frame IF1Relatively The the first subframe F answered1And with the second pixel P2Corresponding second subframe F2.For example, when described image frame IF and 2A schemes Sensing element 6 × 5 sensor pixel region ASWhen (that is, 6 × 10 pixel array) is corresponding, 6 × 5 described first Pixel P1And 6 × 5 the second pixel P2The gray value information sensed can be respectively used to generate the first subframe F1And institute State the second subframe F2
In general, subject to the relatively described optical ranging system 1 of the object 9 when burnt (in focus), the sensing It will appear a clearly object images in described image frame IF acquired in element 12.Moreover, at the processing unit 14 According to first subframe F caused by described image frame IF1And the second subframe F2In, at image position corresponding to the object 9 It sets roughly the same.That is, when the object 9 is respectively in the first subframe F1And the second subframe F2Formed in When imaging position overlaps mutually (that is, distance between the two is 0).At this point, between the object 9 and the optical ranging system 1 Between linear distance may be defined as reference distance L0.
However, when the relatively described optical ranging system 1 of the object 9 is (out of focus) out of focus, the sensing It then will appear two object images in described image frame IF acquired in element 12 and respectively in the first subframe F1And it is described Second subframe F2The first imaging position I of middle formation1And the second imaging position I2, as shown in Figure 4.At this point, the first subframe F1And The second subframe F2In perpendicular to the center line of the preset direction can be respectively defined as the first reference line R1And second reference line R2.Then, the processing unit 14 calculates separately the first imaging position I1To the first reference line R1First projection away from From D1And the second imaging position I2To the second reference line R1The second projector distance D2
It should be noted that, it is assumed that the first imaging position I1And the second imaging position I2When overlapping, the reference Distance L0 is it is known that the first projector distance D at this time1And the second projector distance D2It is all 0.By the object 9 and institute The distance between optical ranging system 1 is stated with the object 9 in the first subframe F1The generated first imaging position I1 The first projector distance D1(or with the object 9 in the second subframe F2The generated second imaging position I2's The second projector distance D2) it is in preset relation, such as linear relationship or non-linear relation, the optical ranging system 1 can thing First the reference distance L0 and the preset relation are stored to storage element.Accordingly, the processing unit 14 can be according to described First imaging position I1And the second imaging position I2Estimate at least one object distance (that is, relatively described light of the object 9 Learn the distance of range-measurement system 1).
In one embodiment, the processing unit 14 can be according to the first projector distance D1With second projector distance D2Difference (such as D1-D2) estimate at least one described object distance.For example, can be opposite according to the difference and the object 9 The relativeness of the distance of the optical ranging system 1 establishes look-up table (look-up table) in advance, such as shown in the following table 1 (wherein object distance L2 > L0 > L1), and be stored among the storage element.
First projector distance D1 Second projector distance D2 Difference (D1-D2) Object distance
0 0 0 L0
-1 +1 -2 L1
+1 -1 2 L2
Table 1
In another embodiment, relativeness of the difference at a distance from the relatively described optical ranging system 1 of the object 9 Straight line equation can be formed, and is stored among the storage element, but not limited to this.In short, the optics of the invention The processing unit 14 of range-measurement system 1 can be according to the first subframe F1The first imaging position I1And the second subframe F2 The second imaging position I2At least one object distance is calculated, must use and beat compared to known distance sensing system (DMS) The first pixel P of negligible amounts can be used only without polishing in the mode of light, the embodiment of the present invention optical ranging system 11And Second pixel P2Object distance is detected, to have the advantages that low cost and small size.
Compared to the multiple sensor pixel region A of first embodiment of the inventionSPixel comprising two mirror symmetries (that is, the first pixel P1And the second pixel P2), the sensor pixel region A of second embodiment of the inventionSIt may include surpassing Two pixels are crossed, such as include four pixels.With reference to Fig. 1,5A and Fig. 5 B, wherein Fig. 5 A is second embodiment of the invention The top view of sensing element, Fig. 5 B are the processing unit of second embodiment of the invention according to the signal of picture frame estimated object distance Figure.The optical ranging system 1 of second embodiment of the invention includes lens 10, sensing element 12 and processing unit 14, wherein described The effect of lens 10, illustrates in first embodiment, therefore repeats no more in this.
The sensing element 12 be used to obtain across the lens 10 light and export picture frame (such as shown in Fig. 5 B Picture frame IF, corresponding with the 8 × 12 of Fig. 5 A pixel array), and include the multiple sensor pixel regions arranged with array AS, wherein the multiple sensor pixel region ASEach include the first pixel P1, the second pixel P2, third pixel P3And the 4th Pixel P4, as shown in Figure 5A.
The multiple sensor pixel region ASEach of also include the first shielding layer S1, the second shielding layer S2, third hide Cover a layer S3And the 4th shielding layer S4.The first shielding layer S1It is arranged in the first pixel P1On and have the first aperture O1, Wherein the first aperture O1Aperture from the first pixel P1Center increase along first direction (such as X-direction) or dull Increase.The second shielding layer S2It is arranged in the second pixel P2On and have the second aperture O2, wherein second aperture O2With the first aperture O1Shape be mirrored into symmetrically in the first direction.The third shielding layer S3Setting is described the Three pixel P3On and have third aperture O3, wherein the third aperture O3Aperture from the third pixel P3Center edge Second direction (such as Y-direction) increases or is increased monotonically.The 4th shielding layer S4It is arranged in the 4th pixel P4On and have There is the 4th aperture O4, wherein the 4th aperture O4With the third aperture O3Shape be mirrored into symmetrically in the second direction.
In the present embodiment, the first direction (such as X-direction) perpendicular to the second direction (such as Y-direction), but this It invents without being limited thereto.
Then, four lenticule (not shown) are separately positioned on the first shielding layer S1, the second shielding layer S2, institute State third shielding layer S3And the 4th shielding layer S4Between the lens 10, such as it is separately positioned on the shielding layer S1~ S4On, and with the first pixel P1, the second pixel P2, the third pixel P3And the 4th pixel P4It is corresponding, The wherein lenticule L of the lenticule and first embodimentMHave effects that identical, therefore is repeated no more in this.
It should be noted that the first aperture of the embodiment of the present invention O1And the second aperture O2Respectively with first embodiment The first aperture O1And the second aperture O2Have effects that same shape and.However, different from the first embodiment, The multiple sensor pixel region A of the embodiment of the present inventionSIt also include the third aperture O3, the 4th aperture O4And its it is right respectively The pixel and shielding layer answered.It is understood that simultaneously by the third shielding layer S3And the 4th shielding layer S4With the sense Survey pixel region ASAfter being rotated by 90 ° counterclockwise for axle center, the postrotational third shielding layer S3And the 4th shielding layer S4 Respectively with the first shielding layer S1And the second shielding layer S2With same shape.Meanwhile the third after rotating is opened Hole O3And the 4th aperture O4Respectively with the first aperture O1And the second aperture O2With same shape).Therefore, institute State third aperture O3And the 4th aperture O4The first aperture O can be reached in the second direction1And the second aperture O2 Identical effect.
Then, the processing unit 14 according to described image frame IF in addition to that can generate and the first pixel P1It is corresponding First subframe F1And with the second pixel P2Corresponding second subframe F2, also can produce and the third pixel P3It is corresponding Third subframe F3And with the 4th pixel P4Corresponding 4th subframe F4, and according to the first subframe F1The first one-tenth Image position I1, the second subframe F2The second imaging position I2, the third subframe F3Third imaging position I3And described Four subframe F4The 4th imaging position I4Estimate at least two object distances.
For example, the first subframe F1And the second subframe F2It is respectively provided with the vertical first direction (such as X-direction) Center line and be defined as the first reference line R1And the second reference line R2, the third subframe F3And the 4th subframe F4Respectively Center line with the vertical second direction (such as Y-direction) is simultaneously defined as third reference line R3And the 4th reference line R4.It is described Processing unit 14 then calculates the first imaging position I1To the first reference line R1The first projector distance D1, described second Imaging position I2To the second reference line R2The second projector distance D2, the third imaging position I3It is referred to the third Line R3Third projector distance D3And the 4th imaging position I4To the 4th reference line R4The 4th projector distance D4, and According to the first projector distance D1With the second projector distance D2The first difference and the third projector distance D3With it is described 4th projector distance D4The second difference estimate at least two object distance, wherein 14 estimated object of the processing unit away from From mode illustrated in first embodiment of the invention and Fig. 4, therefore repeated no more in this.
Further, since the third shielding layer S of the sensing element 12 of the present embodiment3With the first shielding layer S1Have Same shape, the third shielding layer S3Third aperture O3Inevitable and the first shielding layer S1The first aperture O1With phase Similar shape and area, as Fig. 5 A is shown as triangle, however, the present invention is not limited thereto.In one embodiment, the first aperture O1 With the third aperture O3With same area but do not have same shape, such as the first aperture O1To be trapezoidal, the third Aperture O3For semicircle, and it is described it is trapezoidal with the semicircular area be identical.
In the embodiment of the present invention, there is no specific limits for the shape for the aperture that the shielding layer for sensor pixel area is included System, as long as the aperture of aperture is along corresponding pixel center along preset direction increase.For example, described first opens in Fig. 6 A Hole O1Aperture from the first pixel P1Center along first direction (such as X-direction) exponential increasing.It is understood that by In the second aperture O2With the first aperture O1It is mirrored into symmetrically in the first direction, the second aperture O2Aperture From the second pixel P2Center along the first direction opposite direction (such as -X direction) exponential increasing.Furthermore work as sense When surveying pixel region comprising four sensor pixels, as shown in Figure 6B, the third aperture O3Aperture then from the third pixel P3 Center (such as Y-direction) exponential increasing in a second direction, and the 4th aperture O4Aperture from the 4th pixel P4In Opposite direction (such as -Y direction) exponential increasing of the heart along the second direction.
In one embodiment, the first aperture O1, the second aperture O2, the third aperture O3And the described 4th open Hole O4Shape be semicircle, as shown in figs. 7 a and 7b.
In one embodiment, the first aperture O1, the second aperture O2, the third aperture O3And the described 4th open Hole O4Shape be it is trapezoidal, as shown in Fig. 8 A and Fig. 8 B.
In the present invention, the multiple sensor pixel region ASEach (such as include the first pixel P1, described Two pixel P2, the third pixel P3And the 4th pixel P4) it can be the light sensing pixel independently made, or can be same Adjacent or non-conterminous light sensing pixel, has no specific limitation in pixel array.In some embodiments, it can be selected in pixel array Partial pixel is using as the multiple sensor pixel region AS, and other pixels can be used for executing other function.
For example, with reference to Fig. 9, for red (R), green (G), blue (B) three kinds of sensor pixels sensing element.A part Shielding layer and lenticule are respectively arranged in G pixel, wherein the shielding layer includes aperture (such as in first embodiment of the invention Shape is the first aperture O of triangle1And the second aperture O2).And it is then not provided in the G pixel of other parts, R pixel and B pixel Shielding layer and lenticule.At this point, the G pixel of the part can be used for obtaining the picture frame comprising subject depth information, other pictures It is plain then for obtaining the picture frame comprising two-dimensional image information.
Further, since the optical ranging system 1 of the present embodiment can only use a small amount of sensor pixel (such as first picture Plain P1And the second pixel P2) position that counter can push away the object 9, in other words, if the optical ranging system 1 of the present embodiment Sensor pixel array comprising greater number of sensor pixel, such as 300 × 300,600 × 600,900 × 900, can obtain More location informations of the object 9, and then build up the 3D rendering of the object 9.
It should be noted that numerical value in above-described embodiment, such as projector distance, difference etc., are merely to illustrate rather than use It is of the invention in limiting.
In conclusion well known Range Measurement System and gesture identification system need higher cost and size, and usually It needs in addition to provide light source.Therefore, the present invention proposes a kind of sensing element and optical ranging system (Fig. 1), utilizes mirror image pair The sensor pixel of title carries out detecting phase to image is obtained, and uses two dimension, three-dimensional position and the change in location of judgment object, and Due to having the advantages that low cost and size are small without using light source.
Although the present invention is disclosed by examples detailed above, so it is not intended to limit the present invention, skill belonging to any present invention The technical staff in art field, without departing from the spirit and scope of the present invention, when various change and modification can be made.Therefore this hair Bright protection scope is when being subject to the limited range of the appended claims.

Claims (18)

1. a kind of optical ranging system, includes:
Lens;
Sensing element, for obtaining the light across the lens and exporting picture frame, the sensing element includes to be arranged with array Multiple sensor pixel regions of column, wherein each of the multiple sensor pixel region includes:
First pixel;
Second pixel, adjacent first pixel on preset direction;
First shielding layer, be arranged on first pixel and have the first aperture, wherein the aperture of first aperture from The center of first pixel increases along the preset direction, and the area of first aperture is the area of first pixel 5%~45%;
Second shielding layer is arranged on second pixel and has the second aperture, wherein second aperture and described the The shape of one aperture is mirrored into symmetrically in the preset direction;And
At least one lenticule is arranged between first shielding layer and second shielding layer and the lens;And
Processing unit, for generating corresponding with first pixel the first subframe according to described image frame and with described second Corresponding second subframe of pixel, and at least one object distance is calculated according to first subframe and second subframe.
2. optical ranging system according to claim 1, wherein each of the multiple sensor pixel region includes One lenticule, the lenticule are corresponding with first aperture and second aperture simultaneously.
3. optical ranging system according to claim 1, wherein each of the multiple sensor pixel region includes Two lenticules, described two lenticules are corresponding with first pixel and second pixel respectively.
4. optical ranging system according to claim 1, wherein first shielding layer and second shielding layer difference It is coated with or is covered on first pixel and second pixel.
5. optical ranging system described in any one of -4 claims according to claim 1, wherein first aperture and The shape of second aperture is triangle, trapezoidal or semicircle.
6. optical ranging system described in any one of -4 claims according to claim 1, wherein first aperture The aperture is from the center of first pixel along the preset direction exponential increasing.
7. optical ranging system according to claim 1, wherein the processing unit also calculates the of first subframe Second imaging position of the first projector distance of one imaging position to the first reference line and second subframe is to the second reference line The second projector distance, and according to the estimation of the difference of first projector distance and second projector distance it is described at least one Object distance.
8. optical ranging system according to claim 7, the optical ranging system also includes:
Storage element, for storing the preset relation of the difference Yu the object distance.
9. a kind of sensing element, comprising the multiple sensor pixel regions arranged with array, wherein the multiple sensor pixel region Each of include:
First pixel;
Second pixel, adjacent first pixel on preset direction;
First shielding layer, be arranged on first pixel and have the first aperture, wherein the aperture of first aperture from The center of first pixel increases along the preset direction, and the area of first aperture is the area of first pixel 5%~45%;
Second shielding layer is arranged on second pixel and has the second aperture, wherein second aperture and described the The shape of one aperture is mirrored into symmetrically in the preset direction;And
At least one lenticule is arranged on first shielding layer and second shielding layer.
10. sensing element according to claim 9, wherein first shielding layer and second shielding layer are respectively coated Or it is covered on first pixel and second pixel.
11. according to sensing element described in any one of claim 9 to 10 claim, wherein first aperture and institute The shape for stating the second aperture is triangle, trapezoidal or semicircle.
12. according to sensing element described in any one of claim 9 to 10 claim, wherein the institute of first aperture Aperture is stated from the center of first pixel along the preset direction exponential increasing.
13. a kind of optical ranging system, includes:
Lens;
Sensing element, for obtaining the light across the lens and exporting picture frame, the sensing element includes to be arranged with array Multiple sensor pixel regions of column, wherein each of the multiple sensor pixel region includes:
First pixel, the second pixel, third pixel and the 4th pixel;
First shielding layer, be arranged on first pixel and have the first aperture, wherein the aperture of first aperture from The center of first pixel increases along first direction;
Second shielding layer is arranged on second pixel and has the second aperture, wherein second aperture and described the The shape of one aperture is mirrored into symmetrically in the first direction;
Third shielding layer, be arranged on the third pixel and have third aperture, wherein the aperture of the third aperture from The center of the third pixel increases in a second direction;
4th shielding layer is arranged on the 4th pixel and has the 4th aperture, wherein the 4th aperture and described the The shape of three apertures is mirrored into symmetrically in the second direction;And
Four lenticules are separately positioned on first shielding layer, second shielding layer, the third shielding layer and described Between four shielding layers and the lens;And
Processing unit, for generating the first subframe corresponding with first pixel and described second according to described image frame Corresponding second subframe of pixel, third subframe corresponding with the third pixel and corresponding with the 4th pixel Four subframes, and according to first subframe, second subframe, the third subframe and the 4th subframe estimation at least two Object distance.
14. optical ranging system according to claim 13, wherein first aperture has phase with the third aperture Similar shape or same area.
15. optical ranging system described in 3 or 14 according to claim 1, wherein first aperture and the third aperture Shape is triangle, trapezoidal or semicircle.
16. optical ranging system described in 3 or 14 according to claim 1, wherein first aperture and the third aperture The aperture is respectively from the center of first pixel and the third pixel along the first direction and the second party To exponential increasing.
17. optical ranging system according to claim 13, wherein the processing unit also calculates first subframe The second imaging position to the second reference line of first projector distance of the first imaging position to the first reference line, second subframe The second projector distance, the third subframe third imaging position to the third projector distance of third reference line and the described 4th 4th projector distance of the 4th imaging position to the 4th reference line of subframe, and according to first projector distance and described second First difference of projector distance and the estimation of the second difference of the third projector distance and the 4th projector distance are described at least Two object distances.
18. optical ranging system according to claim 13, wherein the first direction is perpendicular to the second direction.
CN201510216100.2A 2015-04-30 2015-04-30 Sensing element and optical ranging system Active CN106200882B (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
CN201910293007.XA CN110275606B (en) 2015-04-30 2015-04-30 Sensing element
CN201510216100.2A CN106200882B (en) 2015-04-30 2015-04-30 Sensing element and optical ranging system

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201510216100.2A CN106200882B (en) 2015-04-30 2015-04-30 Sensing element and optical ranging system

Related Child Applications (1)

Application Number Title Priority Date Filing Date
CN201910293007.XA Division CN110275606B (en) 2015-04-30 2015-04-30 Sensing element

Publications (2)

Publication Number Publication Date
CN106200882A CN106200882A (en) 2016-12-07
CN106200882B true CN106200882B (en) 2019-05-10

Family

ID=57458499

Family Applications (2)

Application Number Title Priority Date Filing Date
CN201510216100.2A Active CN106200882B (en) 2015-04-30 2015-04-30 Sensing element and optical ranging system
CN201910293007.XA Active CN110275606B (en) 2015-04-30 2015-04-30 Sensing element

Family Applications After (1)

Application Number Title Priority Date Filing Date
CN201910293007.XA Active CN110275606B (en) 2015-04-30 2015-04-30 Sensing element

Country Status (1)

Country Link
CN (2) CN106200882B (en)

Families Citing this family (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10627518B2 (en) * 2017-06-02 2020-04-21 Pixart Imaging Inc Tracking device with improved work surface adaptability
CN109040566B (en) * 2018-09-14 2024-04-02 深圳阜时科技有限公司 Image sensor, image acquisition device, identity recognition device and electronic equipment
WO2020118640A1 (en) * 2018-12-13 2020-06-18 深圳市汇顶科技股份有限公司 Optical collection apparatus and electronic device
CN111133445B (en) * 2019-08-23 2021-09-24 深圳市汇顶科技股份有限公司 Fingerprint identification device and electronic equipment
KR102494086B1 (en) 2019-08-23 2023-01-30 선전 구딕스 테크놀로지 컴퍼니, 리미티드 Fingerprint detection device, method and electronic device
CN111328398B (en) * 2019-08-23 2021-09-17 深圳市汇顶科技股份有限公司 Fingerprint identification device and electronic equipment
CN111095285B (en) * 2019-08-23 2021-09-17 深圳市汇顶科技股份有限公司 Fingerprint identification device and electronic equipment
KR102610583B1 (en) 2019-10-18 2023-12-05 선전 구딕스 테크놀로지 컴퍼니, 리미티드 Fingerprint detection devices and electronic devices
CN113218342B (en) * 2021-06-08 2023-02-28 镇江市博驰汽车配件有限公司 Axial leveling detection device and method for hydraulic motor

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102300054A (en) * 2010-06-22 2011-12-28 株式会社理光 Range-finding device and imaging apparatus
CN103453881A (en) * 2012-02-24 2013-12-18 株式会社理光 Distance measuring device and distance measuring method

Family Cites Families (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP3998549B2 (en) * 2001-11-30 2007-10-31 シャープ株式会社 Liquid crystal display
JP3642489B2 (en) * 2003-06-11 2005-04-27 シャープ株式会社 Liquid crystal display
US20100149474A1 (en) * 2005-08-01 2010-06-17 Masumi Kubo Liquid crystal display device
JP5302644B2 (en) * 2008-12-03 2013-10-02 キヤノン株式会社 Imaging apparatus and imaging system
JP2012003080A (en) * 2010-06-17 2012-01-05 Olympus Corp Imaging apparatus
JP2012043939A (en) * 2010-08-18 2012-03-01 Sony Corp Imaging element and imaging apparatus
EP2566151A4 (en) * 2011-03-07 2014-01-29 Panasonic Corp Image pickup device and rangefinder device
WO2013046820A1 (en) * 2011-09-28 2013-04-04 富士フイルム株式会社 Image sensor and imaging device
TWI475193B (en) * 2011-11-18 2015-03-01 Pixart Imaging Inc Optical distance measurement system and operation method thereof
WO2013145821A1 (en) * 2012-03-28 2013-10-03 富士フイルム株式会社 Imaging element and imaging device
CN104517541B (en) * 2015-01-08 2018-04-27 京东方科技集团股份有限公司 More visual field display units, patterned shielding and more visual field display devices

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102300054A (en) * 2010-06-22 2011-12-28 株式会社理光 Range-finding device and imaging apparatus
CN103453881A (en) * 2012-02-24 2013-12-18 株式会社理光 Distance measuring device and distance measuring method

Also Published As

Publication number Publication date
CN106200882A (en) 2016-12-07
CN110275606B (en) 2022-11-29
CN110275606A (en) 2019-09-24

Similar Documents

Publication Publication Date Title
CN106200882B (en) Sensing element and optical ranging system
KR102354260B1 (en) Method and device for processing lightfield data
JP6509914B2 (en) Image sensor for depth estimation
CN101512601B (en) Method for determining a depth map from images, device for determining a depth map
CN104618704B (en) Method and apparatus for image procossing
KR102370062B1 (en) Method of determining calibration parameter for 3d display device and 3d display device thereof
US11747140B2 (en) Sensing element having pixels exposed by aperture increasing along predetermined direction
US20120162410A1 (en) 3d image sensor
US10708486B2 (en) Generation of a depth-artificial image by determining an interpolated supplementary depth through interpolation based on the original depths and a detected edge
US10567636B2 (en) Resolution enhancement using sensor with plural photodiodes per microlens
EP3513550B1 (en) Flat digital image sensor
CN108353117A (en) 3D multiple aperture imaging devices
CN103299343A (en) Range image pixel matching method
US10760953B2 (en) Image sensor having beam splitter
CN103033166B (en) Target ranging method based on synthetic aperture focused images
CN103621078A (en) Image processing apparatus and image processing program
CN106352847B (en) Distance-measuring device and distance measurement method based on phase difference
Garcia et al. Real-time distance-dependent mapping for a hybrid ToF multi-camera rig
GB2540922B (en) Full resolution plenoptic imaging
Li et al. Depth-assisted demosaicing for light field data in layered object space
JP2012146050A (en) Premises change reading support device, premises change reading support method, and premises change reading support program
CN109348149A (en) A kind of display methods of display and image with imaging function
Garcia Becerro Sensor fusion combining 3-D and 2-D for depth data enhancement
CN101778303A (en) Global property difference-based CCD array video positioning method and system
JP6808467B2 (en) Imaging system

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant