CN110275606A - Sensing element - Google Patents
Sensing element Download PDFInfo
- Publication number
- CN110275606A CN110275606A CN201910293007.XA CN201910293007A CN110275606A CN 110275606 A CN110275606 A CN 110275606A CN 201910293007 A CN201910293007 A CN 201910293007A CN 110275606 A CN110275606 A CN 110275606A
- Authority
- CN
- China
- Prior art keywords
- pixel
- aperture
- shielding layer
- sensing element
- area
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Abstract
The present invention provides a kind of sensing element, comprising the multiple sensor pixel regions arranged with array, wherein each of the multiple sensor pixel region includes the first pixel, the second pixel, the first shielding layer, the second shielding layer and at least one lenticule.Second pixel adjacent first pixel on preset direction.First shielding layer is arranged on first pixel and has the first aperture, wherein the aperture of first aperture increases from the center of first pixel along the preset direction.Second shielding layer is arranged on second pixel and has the second aperture, wherein the shape of second aperture and first aperture is mirrored into symmetrically in the preset direction.At least one described lenticule is arranged on first shielding layer and second shielding layer.
Description
It is 30 days 04 month, entitled " sensing element in 2015 that the application, which is application No. is the 201510216100.2, applying date,
And optical ranging system " Chinese invention patent application divisional application.
Technical field
The present invention relates to a kind of sensing elements, and in particular, to it is a kind of can be applied to range measurement or gesture identification
Sensing element and optical ranging system.
Background technique
In general, Range Measurement System (Distance measurement system, DMS) would generally use light source,
And energy back is reflected by the object to calculate the distance of object using the light beam of light source.Traditional, triangulation location can be used
Mode or the technical solution of flight time (Time of flight, TOF) calculate distance, however use aforesaid way institute
The cost of cost and the size of system comparatively all can be higher.
In addition, the usually usable 3D rendering in the exploitation basis of gesture (gesture) identification eliminates background image first to extract
Foreground object image, wherein such technology will use to two imaging sensors, in this way, the size of gesture identification module
Reduction can not be effectively obtained in the same manner as cost.
Based on above-mentioned, the present invention is mainly that the technology of detecting phase (phase detection) is utilized to obtain 3D rendering,
And it is not necessarily to additional polishing (above-mentioned time-of-flight method needs polishing), and the technical solution that the present invention illustrates can only use single figure
As sensor, the application of detecting distance and gesture identification can be reached.
Summary of the invention
In view of this, the present invention proposes a kind of sensing element and optical ranging system, have low cost, size small and
The advantages of height detecting accuracy.
The present invention illustrate to provide it is a kind of comprising the sensing element in the multiple sensor pixel regions arranged with array, it is the multiple
Each of sensor pixel region includes the first pixel, the second pixel, third pixel, the 4th pixel, the first shielding layer, second
Shielding layer, third shielding layer, the 4th shielding layer and four lenticules.First shielding layer setting first pixel it
The first aperture is gone up and has, wherein the aperture of first aperture increases from the center of first pixel along first direction.Institute
It states the second shielding layer to be arranged on second pixel and there is the second aperture, wherein second aperture is opened with described first
The shape in hole is mirrored into symmetrically in the first direction.The third shielding layer is arranged on the third pixel and has the
Three apertures, wherein the aperture of the third aperture increases in a second direction from the center of the third pixel.4th masking
Layer is arranged on the 4th pixel and has the 4th aperture, wherein the shape of the 4th aperture and the third aperture exists
The second direction is mirrored into symmetrically.Four lenticules are respectively to being located at first pixel, second pixel, described
Third pixel and the 4th pixel.
In order to which above and other objects, features and advantages of the invention can be become apparent from, will hereafter be illustrated appended by cooperation, in detail
Carefully it is described as follows.In addition, in the present invention, identical component is indicated with identical symbol, first stated clearly in this.
Figure of description explanation
Fig. 1 is the schematic diagram of the optical ranging system of first embodiment of the invention.
Fig. 2A is the top view of the sensing element of the optical ranging system of Fig. 1.
Fig. 2 B is the partial enlarged view of Fig. 2A.
Fig. 3 is the schematic diagram of the sensing element of the optical ranging system of one embodiment of the invention.
Fig. 4 is the processing unit of first embodiment of the invention according to the schematic diagram of picture frame estimated object distance.
Fig. 5 A is the top view of the sensing element of the optical ranging system of second embodiment of the invention.
Fig. 5 B is the processing unit of second embodiment of the invention according to the schematic diagram of picture frame estimated object distance.
Fig. 6 A~8B is the schematic diagram in the sensor pixel area with aperture of different shapes.
Fig. 9 is the top view of the sensing element of the optical ranging system of third embodiment of the invention.
Specific embodiment
Refering to what is shown in Fig. 1, the schematic diagram of its optical ranging system 1 for showing first embodiment of the invention.The optics is surveyed
It include lens 10, sensing element 12 and processing unit 14 away from system 1.In the present embodiment, the optical ranging system 1 is for estimating
At least one object distance, for example, when an object 9 enters the effective range of the optical ranging system 1, the optics
Range-measurement system 1 can estimate at least one distance of the relatively described optical ranging system 1 of the object 9 (after being specified in).
It should be noted that Fig. 1 only shows an object (that is, described object 9) for illustrating the present embodiment, but this hair
It is bright without being limited thereto.In some embodiments, multiple objects of the effective range of the optical ranging system 1 are appeared in all
Distance can be estimated.In some embodiments, the object 9 has out-of-flatness surface, and the optical ranging system 1 can be estimated at this time
Multiple distances of the relatively described optical ranging system 1 of different location, use acquirement three on the out-of-flatness surface of the object 9
Tie up image information.
The lens 10 are used for optically focused, such as the light that the object 9 is reflected.The lens 10 for example pass through gasket
(spacer) fixed range is maintained with the sensing element 12, but not limited to this.In other embodiments, the lens 10 can pass through
Shell or support portion and the sensing element 12 maintain the fixed range.In addition, although Fig. 1 only shows a lens 10,
The invention is not limited thereto.In other embodiments, the lens 10 can be the lens group comprising multiple lens.
The sensing element 12 is for obtaining the light across the lens 10 and exporting picture frame IF.The sensing element
12 may be, for example, charge coupled cell (CCD) imaging sensor, complementary metal oxide semiconductor (CMOS) imaging sensor or
Other can be used for sensing the sensor of light energy.The sensing element 12 includes the multiple sensor pixel regions arranged with array
AS, wherein each of the multiple sensor pixel region ASInclude the first pixel P1, the second pixel P2, the first shielding layer S1、
Second shielding layer S2And at least one lenticule LM。
With reference to Fig. 1,2A and Fig. 2 B, then illustrate the multiple sensor pixel region ASEach of each member
Part.Fig. 2A is the top view of the sensing element 12 of Fig. 1, wherein not showing that at least one lenticule LM.Such as Fig. 2A
It is shown, the second pixel of the embodiment of the present invention P2The adjacent first pixel P in preset direction (such as X-direction)1.In addition, this
In embodiment, the first pixel P1With the second pixel P2Shape be square, but not limited to this.In other embodiments,
The first pixel P1With the second pixel P2Shape can be round or rectangle.Although Fig. 2A shows the sensing element
12 have 6 × 5 sensor pixel region AS, but not limited to this, the multiple sensor pixel region ASThe visual practical application of quantity
Depending on.
Fig. 2 B is the partial enlarged view of Fig. 2A, is shown as a sensor pixel region A in thisS.The first shielding layer S1If
It sets in the first pixel P1On and have the first aperture O1, wherein the first aperture O1Aperture from first pixel
P1Center increase or be increased monotonically along the preset direction.The second shielding layer S2It is arranged in the second pixel P2On
And there is the second aperture O2, wherein the second aperture O2With the first aperture O1Shape in the preset direction (such as X
Direction) it is mirrored into symmetrically.That is, the second aperture O2Aperture from the second pixel P2Center along described default
The opposite direction in direction increases.It should be noted that the first shielding layer S1And the second shielding layer S2It is to be worn for covering
Cross the lenticule LMA part of light, and not shielded another part light then may pass through the first shielding layer S1
The first aperture O1And the second shielding layer S2The second aperture O2And arrive at the first pixel P1And described
Two pixel P2。
It is understood that the first shielding layer S of Fig. 2 B1With the first aperture O1The first area summation be equal to
The first pixel P1Area;The second shielding layer S2With the second aperture O2Second area summation be equal to described the
Two pixel P2Area, but not limited to this.In other embodiments, the first area summation and second area summation difference
Slightly larger than the first pixel P1And the second pixel P2Area to avoid light leakage.
The first shielding layer S1And the second shielding layer S2First of metal layer in CMOS technology be can be to the tenth
Any two layers between road metal layer, or with other can the material of shading formed.
In one embodiment, the first shielding layer S is formed with metal material1And the second shielding layer S2Later, lead to
It often can be in the first shielding layer S1And the second shielding layer S2On cover insulating layer or protective layer (passivation
layer).At this point, the insulating layer or the protective layer are preferably that translucent material is formed, so that the first aperture O1And institute
State the second aperture O2With high transmittance.It is understood that the insulating layer or the protective layer can prevent dust from entering institute
State the first pixel P1And the second pixel P2To avoid influence light receiving efficiency.
In the present embodiment, although Fig. 1 shows the first shielding layer S1And the second shielding layer S2With first pixel
P1And the second pixel P2A distance away (such as by protective layer or dielectric layer to form the distance), described first
Shielding layer S1And the second shielding layer S2It is preferably close to the first pixel P1And the second pixel P2.Other embodiments
In, the first shielding layer S1And the second shielding layer S2Coating or the first pixel P can be directly overlayed respectively1And institute
State the second pixel P2On.
At least one lenticule LMIt is arranged in the first shielding layer S1And the second shielding layer S2With the lens
Between 10, such as shown in Fig. 1, the multiple sensor pixel region ASEach of include two lenticule LM, and described two
A lenticule LMRespectively with the first pixel P1And the second pixel P2It is corresponding.At this point, passing through described two lenticule LM
And the first aperture O of mirror symmetry1And the second aperture O2Configuration, the first pixel P1And second pixel
P2The incident beam of out of phase can be respectively received to carry out detecting phase.
It should be noted that by above-mentioned lenticule and the configuration of aperture, in the first pixel P1And second picture
Plain P2When receiving incident beam, close to the first pixel P1And the second pixel P2The received incident light of center institute difference
Between phase difference be not obvious;Opposite, close to the first pixel P1And the second pixel P2Edge receive respectively
Incident light between phase difference it is then larger.Therefore, for the first aperture O1, with the first pixel P1Edge is opposite
The aperture is preferably more than and the first pixel P1The opposite aperture in center.That is, the first aperture O1's
The aperture is from the first pixel P1Center increases along the preset direction.Accordingly, the accuracy of detecting phase can be promoted.
It is understood that due to the first aperture O1The aperture from the first pixel P1Center is along described pre-
Set direction increases, the first aperture O1Area can be less than the first pixel P1The half of area, such as shown in Fig. 2 B.Together
When, the first aperture O1And the second aperture O2Area should be greater than preset area so that acquired in the sensing element 12
Described image frame IF remain with acceptable signal-to-interference ratio (SNR).Preferably, the first aperture O1Area be described first
Pixel P1The 5%~45% of area.
Due to the lenticule LMWith symmetrical structure, in other embodiments, the multiple sensor pixel region ASIn it is every
One can only include a lenticule LM.At this point, the lenticule LMSimultaneously with the first aperture O1And the second aperture O2
It is corresponding, as shown in Figure 3.In addition, the lenticule LMProtective layer can be formed between the shielding layer.
Please also refer to Fig. 1,2A and Fig. 4, wherein Fig. 4 be the processing unit 14 according to picture frame IF estimated object away from
From schematic diagram.The processing unit 14 is, for example, digital signal processor (DSP) or processing circuit, and described in electric connection
Sensing element 12.It is (such as opposite with the 6 × 10 of Fig. 2A pixel array that described image frame IF is exported in the sensing element 12
Answer) to after the processing unit 14, the processing unit 14 is used to be generated and first pixel according to described image frame IF
P1Corresponding first subframe F1And with the second pixel P2Corresponding second subframe F2.For example, working as described image frame IF
With 6 × 5 sensor pixel region A of the sensing element of 2A figureSWhen (that is, 6 × 10 pixel array) is corresponding, 6 × 5
The first pixel P1And 6 × 5 the second pixel P2The gray value information sensed can be respectively used to generate first son
Frame F1And the second subframe F2。
In general, subject to the relatively described optical ranging system 1 of the object 9 when burnt (in focus), the sensing
It will appear a clearly object images in described image frame IF acquired in element 12.Moreover, at the processing unit 14
According to first subframe F caused by described image frame IF1And the second subframe F2In, at image position corresponding to the object 9
It sets roughly the same.That is, when the object 9 is respectively in the first subframe F1And the second subframe F2Formed in
When imaging position overlaps mutually (that is, distance between the two is 0).At this point, between the object 9 and the optical ranging system 1
Between linear distance may be defined as reference distance L0.
However, when the relatively described optical ranging system 1 of the object 9 is (out of focus) out of focus, the sensing
It then will appear two object images in described image frame IF acquired in element 12 and respectively in the first subframe F1And it is described
Second subframe F2The first imaging position I of middle formation1And the second imaging position I2, as shown in Figure 4.At this point, the first subframe F1
And the second subframe F2In perpendicular to the center line of the preset direction can be respectively defined as the first reference line R1And second reference
Line R2.Then, the processing unit 14 calculates separately the first imaging position I1To the first reference line R1First projection
Distance D1And the second imaging position I2To the second reference line R1The second projector distance D2。
It should be noted that, it is assumed that the first imaging position I1And the second imaging position I2When overlapping, the reference
Distance L0 is it is known that the first projector distance D at this time1And the second projector distance D2It is all 0.Due to the object 9 with
The distance between described optical ranging system 1 is with the object 9 in the first subframe F1Generated the first one-tenth image position
Set I1The first projector distance D1(or with the object 9 in the second subframe F2Generated second imaging position
I2The second projector distance D2) it is in preset relation, such as linear relationship or non-linear relation, the optical ranging system 1
The reference distance L0 and the preset relation can be stored to storage element in advance.Accordingly, the processing unit 14 can basis
The first imaging position I1And the second imaging position I2Estimating at least one object distance, (that is, the object 9 is with respect to institute
State the distance of optical ranging system 1).
In one embodiment, the processing unit 14 can be according to the first projector distance D1With second projector distance
D2Difference (such as D1-D2) estimate at least one described object distance.For example, can be opposite according to the difference and the object 9
The relativeness of the distance of the optical ranging system 1 establishes look-up table (look-up table) in advance, such as shown in the following table 1
(wherein object distance L2 > L0 > L1), and be stored among the storage element.
First projector distance D1 | Second projector distance D2 | Difference (D1-D2) | Object distance |
0 | 0 | 0 | L0 |
-1 | +1 | -2 | L1 |
+1 | -1 | 2 | L2 |
Table 1
In another embodiment, relativeness of the difference at a distance from the relatively described optical ranging system 1 of the object 9
Straight line equation can be formed, and is stored among the storage element, but not limited to this.In short, the optics of the invention
The processing unit 14 of range-measurement system 1 can be according to the first subframe F1The first imaging position I1And the second subframe F2
The second imaging position I2At least one object distance is calculated, must use and beat compared to known distance sensing system (DMS)
The first pixel P of negligible amounts can be used only without polishing in the mode of light, the embodiment of the present invention optical ranging system 11
And the second pixel P2Object distance is detected, to have the advantages that low cost and small size.
Compared to the multiple sensor pixel region A of first embodiment of the inventionSPixel comprising two mirror symmetries
(that is, the first pixel P1And the second pixel P2), the sensor pixel region A of second embodiment of the inventionSIt may include surpassing
Two pixels are crossed, such as include four pixels.With reference to Fig. 1,5A and Fig. 5 B, wherein Fig. 5 A is second embodiment of the invention
Sensing element top view, Fig. 5 B be second embodiment of the invention processing unit according to picture frame estimated object distance
Schematic diagram.The optical ranging system 1 of second embodiment of the invention includes lens 10, sensing element 12 and processing unit 14, wherein
The effect of lens 10, illustrates in first embodiment, therefore repeats no more in this.
The sensing element 12 be used to obtain across the lens 10 light and export picture frame (such as shown in Fig. 5 B
Picture frame IF, corresponding with the 8 × 12 of Fig. 5 A pixel array), and include the multiple sensor pixel regions arranged with array
AS, wherein the multiple sensor pixel region ASEach include the first pixel P1, the second pixel P2, third pixel P3And the 4th
Pixel P4, as shown in Figure 5A.
The multiple sensor pixel region ASEach of also include the first shielding layer S1, the second shielding layer S2, third hide
Cover a layer S3And the 4th shielding layer S4.The first shielding layer S1It is arranged in the first pixel P1On and have the first aperture O1,
Wherein the first aperture O1Aperture from the first pixel P1Center increase along first direction (such as X-direction) or dull
Increase.The second shielding layer S2It is arranged in the second pixel P2On and have the second aperture O2, wherein described second opens
Hole O2With the first aperture O1Shape be mirrored into symmetrically in the first direction.The third shielding layer S3It is arranged described
Third pixel P3On and have third aperture O3, wherein the third aperture O3Aperture from the third pixel P3Center
(such as Y-direction) increases or is increased monotonically in a second direction.The 4th shielding layer S4It is arranged in the 4th pixel P4On simultaneously
With the 4th aperture O4, wherein the 4th aperture O4With the third aperture O3Shape be mirrored into pair in the second direction
Claim.
In the present embodiment, the first direction (such as X-direction) perpendicular to the second direction (such as Y-direction), but this
It invents without being limited thereto.
Then, four lenticule (not shown) are separately positioned on the first shielding layer S1, the second shielding layer S2, institute
State third shielding layer S3And the 4th shielding layer S4Between the lens 10, such as it is separately positioned on the shielding layer S1~
S4On, and with the first pixel P1, the second pixel P2, the third pixel P3And the 4th pixel P4It is corresponding,
The wherein lenticule L of the lenticule and first embodimentMHave effects that identical, therefore is repeated no more in this.
It should be noted that the first aperture of the embodiment of the present invention O1And the second aperture O2Respectively with first embodiment
The first aperture O1And the second aperture O2Have effects that same shape and.However, different from the first embodiment,
The multiple sensor pixel region A of the embodiment of the present inventionSIt also include the third aperture O3, the 4th aperture O4And its it is right respectively
The pixel and shielding layer answered.It is understood that simultaneously by the third shielding layer S3And the 4th shielding layer S4With the sense
Survey pixel region ASAfter being rotated by 90 ° counterclockwise for axle center, the postrotational third shielding layer S3And the 4th shielding layer S4
Respectively with the first shielding layer S1And the second shielding layer S2With same shape.Meanwhile the third after rotating is opened
Hole O3And the 4th aperture O4Respectively with the first aperture O1And the second aperture O2With same shape).Therefore, institute
State third aperture O3And the 4th aperture O4The first aperture O can be reached in the second direction1And second aperture
O2Identical effect.
Then, the processing unit 14 according to described image frame IF in addition to that can generate and the first pixel P1It is corresponding
First subframe F1And with the second pixel P2Corresponding second subframe F2, also can produce and the third pixel P3It is corresponding
Third subframe F3And with the 4th pixel P4Corresponding 4th subframe F4, and according to the first subframe F1The first one-tenth
Image position I1, the second subframe F2The second imaging position I2, the third subframe F3Third imaging position I3And described
Four subframe F4The 4th imaging position I4Estimate at least two object distances.
For example, the first subframe F1And the second subframe F2It is respectively provided with the vertical first direction (such as the side X
To) center line and be defined as the first reference line R1And the second reference line R2, the third subframe F3And the 4th subframe F4Point
Not Ju You the vertical second direction (such as Y-direction) center line and be defined as third reference line R3And the 4th reference line R4.Institute
It states processing unit 14 and then calculates the first imaging position I1To the first reference line R1The first projector distance D1, described
Two imaging position I2To the second reference line R2The second projector distance D2, the third imaging position I3Join to the third
Examine line R3Third projector distance D3And the 4th imaging position I4To the 4th reference line R4The 4th projector distance D4,
And according to the first projector distance D1With the second projector distance D2The first difference and the third projector distance D3With institute
State the 4th projector distance D4The second difference estimate at least two object distance, wherein 14 estimated object of the processing unit
The mode of distance illustrates in first embodiment of the invention and Fig. 4, therefore repeats no more in this.
Further, since the third shielding layer S of the sensing element 12 of the present embodiment3With the first shielding layer S1Have
Same shape, the third shielding layer S3Third aperture O3Inevitable and the first shielding layer S1The first aperture O1With phase
Similar shape and area, as Fig. 5 A is shown as triangle, however, the present invention is not limited thereto.In one embodiment, the first aperture O1
With the third aperture O3With same area but do not have same shape, such as the first aperture O1To be trapezoidal, the third
Aperture O3For semicircle, and it is described it is trapezoidal with the semicircular area be identical.
In the embodiment of the present invention, there is no specific limits for the shape for the aperture that the shielding layer for sensor pixel area is included
System, as long as the aperture of aperture is along corresponding pixel center along preset direction increase.For example, described first opens in Fig. 6 A
Hole O1Aperture from the first pixel P1Center along first direction (such as X-direction) exponential increasing.It is understood that by
In the second aperture O2With the first aperture O1It is mirrored into symmetrically in the first direction, the second aperture O2Aperture
From the second pixel P2Center along the first direction opposite direction (such as -X direction) exponential increasing.Furthermore work as sense
When surveying pixel region comprising four sensor pixels, as shown in Figure 6B, the third aperture O3Aperture then from the third pixel P3
Center (such as Y-direction) exponential increasing in a second direction, and the 4th aperture O4Aperture from the 4th pixel P4In
Opposite direction (such as -Y direction) exponential increasing of the heart along the second direction.
In one embodiment, the first aperture O1, the second aperture O2, the third aperture O3And the described 4th open
Hole O4Shape be semicircle, as shown in figs. 7 a and 7b.
In one embodiment, the first aperture O1, the second aperture O2, the third aperture O3And the described 4th open
Hole O4Shape be it is trapezoidal, as shown in Fig. 8 A and Fig. 8 B.
In the present invention, the multiple sensor pixel region ASEach (such as include the first pixel P1, described
Two pixel P2, the third pixel P3And the 4th pixel P4) it can be the light sensing pixel independently made, or can be same
Adjacent or non-conterminous light sensing pixel, has no specific limitation in pixel array.In some embodiments, it can be selected in pixel array
Partial pixel is using as the multiple sensor pixel region AS, and other pixels can be used for executing other function.
For example, with reference to Fig. 9, for red (R), green (G), blue (B) three kinds of sensor pixels sensing element.A part
Shielding layer and lenticule are respectively arranged in G pixel, wherein the shielding layer includes aperture (such as in first embodiment of the invention
Shape is the first aperture O of triangle1And the second aperture O2).And it is then not provided in the G pixel of other parts, R pixel and B pixel
Shielding layer and lenticule.At this point, the G pixel of the part can be used for obtaining the picture frame comprising subject depth information, other pictures
It is plain then for obtaining the picture frame comprising two-dimensional image information.
Further, since the optical ranging system 1 of the present embodiment can only use a small amount of sensor pixel (such as first picture
Plain P1And the second pixel P2) position that counter can push away the object 9, in other words, if the optical ranging system 1 of the present embodiment
Sensor pixel array comprising greater number of sensor pixel, such as 300 × 300,600 × 600,900 × 900, can obtain
More location informations of the object 9, and then build up the 3D rendering of the object 9.
It should be noted that numerical value in above-described embodiment, such as projector distance, difference etc., are merely to illustrate rather than use
It is of the invention in limiting.
In conclusion well known Range Measurement System and gesture identification system need higher cost and size, and usually
It needs in addition to provide light source.Therefore, the present invention proposes a kind of sensing element and optical ranging system (Fig. 1), utilizes mirror image pair
The sensor pixel of title carries out detecting phase to image is obtained, and uses two dimension, three-dimensional position and the change in location of judgment object, and
Due to having the advantages that low cost and size are small without using light source.
Although the present invention is disclosed by examples detailed above, so it is not intended to limit the present invention, skill belonging to any present invention
The technical staff in art field, without departing from the spirit and scope of the present invention, when various change and modification can be made.Therefore this hair
Bright protection scope is when being subject to the limited range of the appended claims.
Claims (10)
1. a kind of sensing element, which includes the multiple sensor pixel regions arranged with array, multiple sensor pixel
Each of region includes:
First pixel, the second pixel, third pixel and the 4th pixel;
First shielding layer, be arranged on first pixel and have the first aperture, wherein the aperture of first aperture from
The center of first pixel increases along first direction;
Second shielding layer is arranged on second pixel and has the second aperture, wherein second aperture and described the
The shape of one aperture is mirrored into symmetrically in the first direction;
Third shielding layer, be arranged on the third pixel and have third aperture, wherein the aperture of the third aperture from
The center of the third pixel increases in a second direction;
4th shielding layer is arranged on the 4th pixel and has the 4th aperture, wherein the 4th aperture and described the
The shape of three apertures is mirrored into symmetrically in the second direction;And
Four lenticules, respectively to positioned at first pixel, second pixel, the third pixel and the 4th picture
Element.
2. sensing element according to claim 1, wherein first aperture and the third aperture have same shape
Or same area.
3. sensing element according to claim 1 or 2, wherein the shape of first aperture and the third aperture is three
It is angular, trapezoidal or semicircle.
4. sensing element according to claim 1 or 2, wherein the aperture of first aperture and the third aperture
It is passed respectively from the center of first pixel and the third pixel along the first direction and the second direction index
Increase.
5. sensing element according to claim 1 or 2, wherein the aperture of first aperture and the third aperture
Respectively from the center of first pixel and the third pixel along the first direction and the second direction monotone increasing
Add.
6. sensing element according to claim 1, wherein the first direction is perpendicular to the second direction.
7. sensing element according to claim 1, wherein first aperture, second aperture, the third aperture
And the area of the 4th aperture is respectively first pixel, second pixel, the third pixel and the 4th picture
The 5%~45% of the area of element.
8. sensing element according to claim 1, wherein first shielding layer, second shielding layer, the third
Shielding layer and the 4th shielding layer are respectively coated or are covered in first pixel, second pixel, the third pixel
And on the 4th pixel.
9. sensing element according to claim 1, wherein first pixel and second pixel configuration are in an institute
On the diagonal line for stating sensor pixel region, the third pixel and the 4th pixel configuration are in one sensor pixel region
Another diagonal line on.
10. sensing element according to claim 1, wherein
First shielding layer and the area summation of first aperture are equal to the area of first pixel;
Second shielding layer and the area summation of second aperture are equal to the area of second pixel;
The third shielding layer and the area summation of the third aperture are equal to the area of the third pixel;And
4th shielding layer and the area summation of the 4th aperture are equal to the area of the 4th pixel.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910293007.XA CN110275606B (en) | 2015-04-30 | 2015-04-30 | Sensing element |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910293007.XA CN110275606B (en) | 2015-04-30 | 2015-04-30 | Sensing element |
CN201510216100.2A CN106200882B (en) | 2015-04-30 | 2015-04-30 | Sensing element and optical ranging system |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201510216100.2A Division CN106200882B (en) | 2015-04-30 | 2015-04-30 | Sensing element and optical ranging system |
Publications (2)
Publication Number | Publication Date |
---|---|
CN110275606A true CN110275606A (en) | 2019-09-24 |
CN110275606B CN110275606B (en) | 2022-11-29 |
Family
ID=57458499
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201510216100.2A Active CN106200882B (en) | 2015-04-30 | 2015-04-30 | Sensing element and optical ranging system |
CN201910293007.XA Active CN110275606B (en) | 2015-04-30 | 2015-04-30 | Sensing element |
Family Applications Before (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201510216100.2A Active CN106200882B (en) | 2015-04-30 | 2015-04-30 | Sensing element and optical ranging system |
Country Status (1)
Country | Link |
---|---|
CN (2) | CN106200882B (en) |
Families Citing this family (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10627518B2 (en) * | 2017-06-02 | 2020-04-21 | Pixart Imaging Inc | Tracking device with improved work surface adaptability |
CN109040566B (en) * | 2018-09-14 | 2024-04-02 | 深圳阜时科技有限公司 | Image sensor, image acquisition device, identity recognition device and electronic equipment |
CN109716350B (en) * | 2018-12-13 | 2023-06-13 | 深圳市汇顶科技股份有限公司 | Optical acquisition device and electronic equipment |
CN111133445B (en) * | 2019-08-23 | 2021-09-24 | 深圳市汇顶科技股份有限公司 | Fingerprint identification device and electronic equipment |
CN111095285B (en) * | 2019-08-23 | 2021-09-17 | 深圳市汇顶科技股份有限公司 | Fingerprint identification device and electronic equipment |
EP3809315B1 (en) | 2019-08-23 | 2022-12-07 | Shenzhen Goodix Technology Co., Ltd. | Fingerprint detection method |
CN111328398B (en) * | 2019-08-23 | 2021-09-17 | 深圳市汇顶科技股份有限公司 | Fingerprint identification device and electronic equipment |
CN111095282B (en) | 2019-10-18 | 2023-09-05 | 深圳市汇顶科技股份有限公司 | Fingerprint detection device and electronic equipment |
CN113218342B (en) * | 2021-06-08 | 2023-02-28 | 镇江市博驰汽车配件有限公司 | Axial leveling detection device and method for hydraulic motor |
Citations (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2003228073A (en) * | 2001-11-30 | 2003-08-15 | Sharp Corp | Liquid crystal display device |
CN1806195A (en) * | 2003-06-11 | 2006-07-19 | 夏普株式会社 | Liquid crystal display |
CN101233447A (en) * | 2005-08-01 | 2008-07-30 | 夏普株式会社 | Liquid crystal display device |
JP2010135509A (en) * | 2008-12-03 | 2010-06-17 | Canon Inc | Imaging apparatus and imaging system |
CN102289131A (en) * | 2010-06-17 | 2011-12-21 | 奥林巴斯株式会社 | Image pickup apparatus |
CN102375293A (en) * | 2010-08-18 | 2012-03-14 | 索尼公司 | Imaging device and imaging apparatus |
US20130070146A1 (en) * | 2011-03-07 | 2013-03-21 | Panasonic Corporation | Image pickup device and rangefinder device |
CN103453881A (en) * | 2012-02-24 | 2013-12-18 | 株式会社理光 | Distance measuring device and distance measuring method |
CN103843320A (en) * | 2011-09-28 | 2014-06-04 | 富士胶片株式会社 | Image sensor and imaging device |
US20150009373A1 (en) * | 2012-03-28 | 2015-01-08 | Fujifilm Corporation | Imaging element and imaging apparatus |
CN104517541A (en) * | 2015-01-08 | 2015-04-15 | 京东方科技集团股份有限公司 | Multi-vision field display component, patterned shielding layer and multi-vision field display device |
Family Cites Families (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9267797B2 (en) * | 2010-06-22 | 2016-02-23 | Ricoh Company, Ltd. | Range-finding device and imaging apparatus |
TWI475193B (en) * | 2011-11-18 | 2015-03-01 | Pixart Imaging Inc | Optical distance measurement system and operation method thereof |
-
2015
- 2015-04-30 CN CN201510216100.2A patent/CN106200882B/en active Active
- 2015-04-30 CN CN201910293007.XA patent/CN110275606B/en active Active
Patent Citations (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2003228073A (en) * | 2001-11-30 | 2003-08-15 | Sharp Corp | Liquid crystal display device |
CN1806195A (en) * | 2003-06-11 | 2006-07-19 | 夏普株式会社 | Liquid crystal display |
CN101233447A (en) * | 2005-08-01 | 2008-07-30 | 夏普株式会社 | Liquid crystal display device |
JP2010135509A (en) * | 2008-12-03 | 2010-06-17 | Canon Inc | Imaging apparatus and imaging system |
CN102289131A (en) * | 2010-06-17 | 2011-12-21 | 奥林巴斯株式会社 | Image pickup apparatus |
CN102375293A (en) * | 2010-08-18 | 2012-03-14 | 索尼公司 | Imaging device and imaging apparatus |
US20130070146A1 (en) * | 2011-03-07 | 2013-03-21 | Panasonic Corporation | Image pickup device and rangefinder device |
CN103843320A (en) * | 2011-09-28 | 2014-06-04 | 富士胶片株式会社 | Image sensor and imaging device |
CN103453881A (en) * | 2012-02-24 | 2013-12-18 | 株式会社理光 | Distance measuring device and distance measuring method |
US20150009373A1 (en) * | 2012-03-28 | 2015-01-08 | Fujifilm Corporation | Imaging element and imaging apparatus |
CN104517541A (en) * | 2015-01-08 | 2015-04-15 | 京东方科技集团股份有限公司 | Multi-vision field display component, patterned shielding layer and multi-vision field display device |
Also Published As
Publication number | Publication date |
---|---|
CN110275606B (en) | 2022-11-29 |
CN106200882A (en) | 2016-12-07 |
CN106200882B (en) | 2019-05-10 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN106200882B (en) | Sensing element and optical ranging system | |
KR102354260B1 (en) | Method and device for processing lightfield data | |
EP3516626B1 (en) | Device and method for obtaining distance information from views | |
US9048153B2 (en) | Three-dimensional image sensor | |
JP6509914B2 (en) | Image sensor for depth estimation | |
CN101512601B (en) | Method for determining a depth map from images, device for determining a depth map | |
US11747140B2 (en) | Sensing element having pixels exposed by aperture increasing along predetermined direction | |
CN104618704B (en) | Method and apparatus for image procossing | |
US10708486B2 (en) | Generation of a depth-artificial image by determining an interpolated supplementary depth through interpolation based on the original depths and a detected edge | |
EP3513550B1 (en) | Flat digital image sensor | |
US10760953B2 (en) | Image sensor having beam splitter | |
US20190174115A1 (en) | Light field data representation | |
CN103621078A (en) | Image processing apparatus and image processing program | |
CN106352847B (en) | Distance-measuring device and distance measurement method based on phase difference | |
US9342889B2 (en) | Three-dimensional measurement system and method therefor | |
GB2540922B (en) | Full resolution plenoptic imaging | |
Li et al. | Depth-assisted demosaicing for light field data in layered object space | |
JP6703387B2 (en) | Mobile display with thin film photosensor, 2D array sensor and fingerprint sensor | |
CN109348149A (en) | A kind of display methods of display and image with imaging function | |
CN107071374B (en) | A kind of projection fusion method based on 3D scanning and Slam | |
CN107071375B (en) | A kind of Slam methods based on 3D scannings | |
CN101778303A (en) | Global property difference-based CCD array video positioning method and system | |
Choi et al. | CMOS image sensor with off-center circular apertures for depth extraction | |
JP6584898B2 (en) | Display with position detection function in 3D space | |
Kunnath et al. | Depth from Defocus on a Transmissive Diffraction Mask-based Sensor |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |