CN113790874A - Lens test system - Google Patents
Lens test system Download PDFInfo
- Publication number
- CN113790874A CN113790874A CN202110995586.XA CN202110995586A CN113790874A CN 113790874 A CN113790874 A CN 113790874A CN 202110995586 A CN202110995586 A CN 202110995586A CN 113790874 A CN113790874 A CN 113790874A
- Authority
- CN
- China
- Prior art keywords
- lens
- target
- image
- target area
- processing device
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000012360 testing method Methods 0.000 title claims abstract description 15
- 238000012545 processing Methods 0.000 claims abstract description 63
- 238000000034 method Methods 0.000 claims description 8
- 230000008569 process Effects 0.000 claims description 5
- 238000010586 diagram Methods 0.000 description 15
- 230000003287 optical effect Effects 0.000 description 12
- 230000006870 function Effects 0.000 description 8
- 239000000047 product Substances 0.000 description 8
- 238000003384 imaging method Methods 0.000 description 7
- 238000004590 computer program Methods 0.000 description 5
- 230000005540 biological transmission Effects 0.000 description 4
- 238000013461 design Methods 0.000 description 3
- 239000000835 fiber Substances 0.000 description 2
- 230000001902 propagating effect Effects 0.000 description 2
- RYGMFSIKBFXOCR-UHFFFAOYSA-N Copper Chemical compound [Cu] RYGMFSIKBFXOCR-UHFFFAOYSA-N 0.000 description 1
- 229910052802 copper Inorganic materials 0.000 description 1
- 239000010949 copper Substances 0.000 description 1
- 238000001514 detection method Methods 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 239000013067 intermediate product Substances 0.000 description 1
- 238000004519 manufacturing process Methods 0.000 description 1
- 239000000463 material Substances 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
- 230000003068 static effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01M—TESTING STATIC OR DYNAMIC BALANCE OF MACHINES OR STRUCTURES; TESTING OF STRUCTURES OR APPARATUS, NOT OTHERWISE PROVIDED FOR
- G01M11/00—Testing of optical apparatus; Testing structures by optical methods not otherwise provided for
- G01M11/02—Testing optical properties
Abstract
The utility model provides a test system of lens, this system includes processing apparatus, projection arrangement, camera device, display device and mobile device, and the lens is installed in the mobile device, and the lens lies in between projection arrangement and the display device, and projection arrangement is used for projecting the first image that processing apparatus provided through the lens; the moving device is used for controlling the lens to move between the projection device and the display device; the display device is used for respectively displaying a second image after the first image at each moving position is projected through the lens; the camera device is used for acquiring each second image and sending each second image to the processing device; the processing device is used for identifying each second image to judge whether the lens is qualified.
Description
Technical Field
The embodiment of the disclosure relates to the technical field of lens testing, and more particularly, to a lens testing system.
Background
The lens is one of the key parts of the optical product, and because the design parameters and the material object of the lens often have differences, the lens is required to be inspected and tested after being produced into a finished product, so that the lens is ensured to accord with the design parameters, the yield of the corresponding optical product is improved, and the problem which needs to be solved urgently is solved.
Disclosure of Invention
An object of the disclosed embodiments is to provide a new technical solution of a lens testing system.
The present disclosure provides a lens testing system, including: the device comprises a processing device, a projection device, a camera device, a display device and a mobile device, wherein the lens is arranged on the mobile device and is positioned between the projection device and the display device,
the projection device is used for projecting the first image provided by the processing device through the lens;
the moving device is used for controlling the lens to move between the projection device and the display device;
the display device is used for respectively displaying a second image after the first image at each moving position is projected through the lens;
the camera device is used for acquiring each second image and sending each second image to the processing device;
the processing device is used for identifying each second image to judge whether the lens is qualified.
Optionally, the moving means comprises a motor and a bracket,
the bracket is used for placing the lens;
the motor is used for controlling the lens to move between the projection device and the display device.
Optionally, a plurality of target regions are included in the first image,
the plurality of target areas at least comprise a first central area positioned in the center of the first image, and other target areas except the first central area are symmetrically distributed based on the first central area;
correspondingly, the second image comprises a plurality of target areas, the target areas at least comprise a second central area located in the center of the second image, and other target areas except the second central area are symmetrically distributed based on the second central area.
Optionally, the processing means is configured to:
respectively identifying each second image to obtain the definition of each target area in each second image;
determining a target focal length of the lens in each target area according to the definition of each target area in each second image;
and determining the diopter of the lens in each target area according to the target focal length of the lens in each target area, and judging whether the lens is qualified or not according to the diopter of the lens in each target area.
Optionally, the processing means is configured to:
acquiring the target definition of the lens in each target area according to the definition of each target area in each second image; and the number of the first and second groups,
and determining the target focal length of the corresponding lens in each target area according to the target definition of the lens in each target area.
Optionally, the motor is provided with graduation marks,
and the scale values of the scale marks correspond to the positions of the motor in the moving process one by one.
Optionally, the processing means is configured to:
acquiring the position of the motor corresponding to each target definition according to the target definition of the lens in each target area, and taking the position as a target position;
and determining the focal length of the lens in each target area according to each target position and the position between the projection device and the display device.
Optionally, the processing device is configured to compare the diopter of the lens in each of the target areas with a target diopter threshold, and determine that the lens is qualified if the diopter of the lens in each of the target areas is greater than or equal to the target diopter threshold.
Optionally, the processing device is configured to provide a configuration interface for configuring the diopter threshold, and acquire the diopter threshold input through the configuration interface as the target diopter threshold.
Optionally, the lens is comprised of at least one lens.
The camera device has the advantages that the lens is controlled to move between the projection device and the display device through the moving device, the display device displays second images obtained after the first images at each moving position are projected through the lens, each second image is collected through the camera device, and then the processing device identifies each second image to judge whether the lens is qualified or not, namely, the camera device can detect the lens to improve the yield of subsequent applications.
Other features of the present description and advantages thereof will become apparent from the following detailed description of exemplary embodiments thereof, which proceeds with reference to the accompanying drawings.
Drawings
The accompanying drawings, which are incorporated in and constitute a part of this specification, illustrate embodiments of the specification and together with the description, serve to explain the principles of the specification.
FIG. 1 is a schematic diagram of a testing system for lenses according to an embodiment of the present disclosure;
FIG. 2 is a schematic diagram of a first image according to an embodiment of the present disclosure;
FIG. 3 is a schematic view of a lens according to an embodiment of the disclosure;
FIG. 4 is a graph illustrating sharpness curves according to an embodiment of the present disclosure;
fig. 5 is a schematic diagram of a lens imaging principle according to an embodiment of the disclosure.
Detailed Description
Reference will now be made in detail to embodiments of the present application, examples of which are illustrated in the accompanying drawings, wherein like reference numerals refer to the same or similar elements or elements having the same or similar function throughout. The embodiments described below with reference to the drawings are exemplary only for the purpose of explaining the present application and are not to be construed as limiting the present application. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present application.
The features of the terms first and second in the description and in the claims of the present application may explicitly or implicitly include one or more of such features. In the description of the present application, "a plurality" means two or more unless otherwise specified.
In the description of the present application, it is to be understood that the terms "center," "length," "width," "thickness," "upper," "lower," "front," "rear," "left," "right," "vertical," "horizontal," "top," "bottom," "inner," "outer," "clockwise," "counterclockwise," "axial," "circumferential," and the like are used in the orientations and positional relationships indicated in the drawings for convenience in describing the present application and to simplify the description, and are not intended to indicate or imply that the device or element being referred to must have a particular orientation, be constructed and operated in a particular orientation, and are therefore not to be considered limiting of the present application.
In the description of the present application, it is to be noted that, unless otherwise explicitly specified or limited, the terms "mounted," "connected," and "connected" are to be construed broadly, e.g., as meaning either a fixed connection, a removable connection, or an integral connection; can be mechanically or electrically connected; they may be connected directly or indirectly through intervening media, or they may be interconnected between two elements. The specific meaning of the above terms in the present application can be understood in a specific case by those of ordinary skill in the art.
The embodiment of the present disclosure provides a test system for a lens, where the lens is an optical lens and is composed of at least one lens, and the at least one lens includes a concave lens and a convex lens. The lens is an essential component in an optical product, and plays a key role in the imaging quality of the optical product. Fig. 1 is a block schematic diagram of a lens testing system according to an embodiment of the present disclosure.
Referring to fig. 1, the lens testing system 10 includes a processing device 110, a projection device 120, an image capturing device 130, a display device 140, and a moving device 150, wherein the lens 20 is mounted on the moving device 150, and the lens 20 is located between the projection device 120 and the display device 140. The processing device 110 is communicatively coupled to the projection device 120, and the imaging device 130 is communicatively coupled to the processing device 110.
In this embodiment, the processing device 110 is configured to provide the first image to the projection device 120. The projection device 120 is configured to project the first image provided by the processing device 110 through the lens 20. The moving device 150 is used for controlling the lens 20 to move between the projection device 120 and the display device 140. The display device 140 is used for displaying the second image projected by the lens 20 after the first image at each moving position is displayed. The camera device 130 is configured to capture each second image and send each second image to the processing device 110. The processing device 110 is used for identifying each second image to determine whether the lens is qualified.
The processing device 110 may be an industrial personal computer having important computer attributes and features, such as a computer CPU, a hard disk, an internal memory, an external device, an interface, an operating system, a control network, a protocol, computing capabilities, and a friendly human-computer interface. The product and technology of industrial control industry are very special, belong to the intermediate product, it is reliable, embedded, intelligent industrial computer to provide for other every trade. The processing device 110 is capable of providing a first image, which may be an image of a black line with white background as shown in fig. 2.
The first image comprises a plurality of target areas, the plurality of target areas at least comprise a first central area positioned in the center of the first image, and other target areas except the first central area are symmetrically distributed based on the first central area. It will be appreciated that to distinguish the central region of the first image from the central region of the second image, the central region of the first image will be referred to herein as the first central region and the central region of the second image will be referred to herein as the second central region.
As shown in fig. 2, the first image includes four target areas, namely, a target area 0, a target area 1, a target area 2, a target area 3, and a target area 4. The target area 0 is a central area of the first image and is called a first central area, the target area 1 is located in an upper left area of the first image, the target area 2 is located in an upper right area of the first image, the target area 3 is located in a lower right area of the first image, and the target area 4 is located in a lower left area of the first image. The target area 1, the target area 2, the target area 3, and the target area 4 are distributed based on the target area 1 in a central symmetry manner. Wherein, 1, 2, 3, 4 are only for distinguishing different target areas, and the digital mark does not exist in the actual first image.
The projection device 120 may be a projection apparatus, or may be only an optical engine in the projection apparatus, where the optical engine is a standard optical engine. The projection device 120 may be used as a light source in the lens testing system 10. The projection device 120 is communicatively connected to the processing device 110 to receive the first image provided by the processing device 110 and project the first image.
The above lens 20 is composed of at least one lens including a concave lens and a convex lens, and the lens 20 is an indispensable component in an optical product, and plays a key role in the imaging quality of the optical product. It is understood that the light of the first image projected by the projection device 120 is generally divergent, and the light is converged by the lens 20 and projected to the display device 140. As shown in fig. 3, which is an image of the first image shown in fig. 2 as it passes through the lens 20.
The above moving means 150 includes a motor and a stand for placing the lens 20. The motor is used to control the lens 20 to move between the projection device 120 and the display device 140. The motor is a stepping motor, scale marks are arranged on the stepping motor, and the scale values of the scale marks correspond to the positions of the motor in the moving process one by one.
The above display device 140 may be a rear projection imaging light curtain. The display device 140 is used for displaying the second image projected by the lens 20 after the first image at each moving position is displayed.
The above camera device 130 is used for capturing each second image and sending each second image to the processing device 110. For example, the moving device 150 may control the lens 20 to move by a set step length, the display device 140 may display a second image processed by the lens 20 from the first image at the corresponding position, and the camera device 130 captures the second image at the corresponding position and sends the second image to the processing device 110. The second image comprises a plurality of target areas, the plurality of target areas at least comprise a second central area positioned in the center of the second image, and other target areas except the second central area are symmetrically distributed based on the second central area.
The processing device 110 is used for identifying each second image to determine whether the lens is qualified.
In this embodiment, the processing device 110 configured to identify each second image to determine whether the lens 20 is qualified may further include:
In this step 110, the processing device 110 identifies each second image, so that the definition of each target area in each second image can be obtained, and here, a definition graph can be constructed with the definition of each target area as a reference. Fig. 4 is a constructed sharpness graph, in which the horizontal axis represents the position of the motor, i.e., the distance between the lens and the projection apparatus, and the vertical axis represents sharpness. Wherein the curve TL represents the sharpness curve of the target region 1 in the second image, the curve TR represents the sharpness curve of the target region 2 in the second image, the curve BL represents the sharpness curve of the target region 4 in the second image, the curve CT represents the sharpness curve of the target region 0 in the second image, and the curve BR represents the sharpness curve of the target region 3 in the second image.
In step 120, the processing device 110 is configured to determine a target focal length of the lens in each target area according to the sharpness of each target area in each second image.
The target focal length of the lens in each target area is the maximum focal length of the lens in each target area.
In this embodiment, the determining, by the processing device in step 120, the target focal length of the lens in each target area according to the sharpness of each target area in each second image may further include:
in step 121, the processing device 110 is configured to obtain a target sharpness of the lens 20 in each target area according to the sharpness of each target area in each second image.
The target sharpness of the lens 20 in each target area is the maximum sharpness of the lens in each target area.
Taking the example that the second image includes the five target regions, in this step 121, the processing device 110 can obtain the maximum definitions of the five target regions according to the definitions of the five target regions in each second image.
Step 122, the processing device 110 is configured to determine a target focal length of the corresponding lens 20 in each target area according to the target sharpness of the lens in each target area.
Because the motor is provided with the scale marks, and each scale value of the scale marks corresponds to each position of the motor in the moving process one by one. Here, the step 122 of determining, by the processing device 110, the target focal length of the corresponding lens 20 in each target area according to the target sharpness of the lens 20 in each target area may further include:
the processing device 110 obtains a position of the motor corresponding to each target definition as a target position according to the target definition of the lens 20 in each target area. The processing device 110 determines a target focal length of the lens 20 in each target area according to each target position and a position between the projection device 120 and the display device 140.
The position of the motor can reflect the distance between the projection device 120 and the lens 20, i.e., the object distance u. Fig. 5 shows a schematic diagram of lens imaging, wherein BB' is a fixed length and represents the distance between the projection device 120 and the display device 140. When the motor moves the lens 20, the object distance u is changed, and the image distance v is BB' -u. According to lens imaging formulaThe focal length f is obtained.
Taking the example that the second image includes the five target areas, the processing device 110 determines the positions of the motors corresponding to the maximum definitions of the five target areas as target positions according to the maximum definitions of the five target areas respectively and according to the definition curves shown in fig. 4. Each target position can reflect the distance u between the projection device 120 and the lens 20 when the corresponding target region is at maximum sharpness. V can be obtained according to the formula v ═ BB' -u, and the formula can be obtainedAnd obtaining the target focal length f.
Taking the example that the second image includes the five target regions, in this step 120, the processing device 110 can determine the target focal lengths of the lens 20 in the five target regions according to the degrees of sharpness of the five target regions in each second image, that is, five maximum focal lengths, f0,f1,f2,f3,f4Wherein f is0Denotes the target focal length, f, of the lens 20 in the second image for the target area 01Representing the target focal length, f, of the lens 20 in the target area 1 in the second image2Representing the target focal length, f, of the target area 2 of the lens 20 in the second image3Representing the target focal length, f, of the target area 3 of the lens 20 in the second image4Which represents the target focal length of the lens 20 in the second image for the target area 4.
In step 130, the processing device 110 determines the diopter of the lens 20 in each target area according to the target focal length of the lens 20 in each target area, and determines whether the lens 20 is qualified according to the diopter of the lens 20 in each target area.
Taking the example that the five target regions are included in the second image, in the above step 130, the processing device 110 can determine the target focal lengths of the lens 20 in the five target regions according to the degrees of sharpness of the five target regions, that is, five maximum focal lengths, f0,f1,f2,f3,f4Wherein f is0Indicating that the lens 20 is in the second imageTarget focal length of target area 0, f1Representing the target focal length, f, of the lens 20 in the target area 1 in the second image2Representing the target focal length, f, of the target area 2 of the lens 20 in the second image3Representing the target focal length, f, of the target area 2 of the lens 20 in the second image4Which represents the target focal length of the lens 20 in the second image for the target area 4. By the formulaWhere D represents diopter and f represents the target focal length. The diopter D of the lens 20 in the five target areas can be obtained0,D1,D2,D3,D4Wherein D is0Represents the diopter, D, of the lens 20 at the target area 01Shows the diopter, D, of the lens 20 in the target area 12Shows the diopter, D, of the lens 20 in the target area 13Shows the diopter, D, of the lens 20 in the target area 14Representing the diopter of the lens 20 at the target area 4.
In this step 130, the determining, by the processing device 110, whether the lens 20 is qualified according to the diopter of the lens 20 in each target area may further include:
the processing device 110 determines that the lens 20 is acceptable in accordance with comparing the diopter of the lens 20 in each target area with the target diopter threshold, and in the case where the diopter of the lens 20 in each target area is greater than or equal to the target diopter threshold.
Taking the example that the second image includes the five target areas, the diopters D of the lens 10 in the five target areas can be respectively adjusted0,D1,D2,D3,D4And a target diopter threshold value DdiffMaking a comparison if D0Is less than Ddiff,D1Is less than Ddiff,D2Is less than Ddiff,D3Is less than Ddiff,D4Is less than DdiffIn the case of (2), the lens 20 is determined to be acceptable.
According to the lens testing system disclosed by the embodiment of the disclosure, the lens is controlled to move between the projection device and the display device through the moving device, the display device displays the second images obtained after the first images at each moving position are projected through the lens, each second image is collected by the camera device, and then the processing device identifies each second image to judge whether the lens is qualified or not, namely, the lens testing system can detect the lens to improve the yield of subsequent applications.
In one embodiment, the processing device 110 is further configured to provide a configuration interface for configuring the diopter threshold value, and acquire the diopter threshold value input through the configuration interface as the target diopter threshold value.
The configuration interface may be an input box, a drop-down list, a voice input, etc., for example, a tester may input a target diopter threshold value through the input box; for another example, the detector may select the target diopter threshold through a drop-down list; for another example, the operator may voice-input the target diopter threshold.
The embodiment provides a human-computer interaction interface to support detection personnel to select the required target diopter threshold value according to the current actual requirement, and the customized design is realized.
The present disclosure may be systems, methods, and/or computer program products. The computer program product may include a computer-readable storage medium having computer-readable program instructions embodied thereon for causing a processor to implement various aspects of the present disclosure.
The computer readable storage medium may be a tangible device that can hold and store the instructions for use by the instruction execution device. The computer readable storage medium may be, for example, but not limited to, an electronic memory device, a magnetic memory device, an optical memory device, an electromagnetic memory device, a semiconductor memory device, or any suitable combination of the foregoing. More specific examples (a non-exhaustive list) of the computer readable storage medium would include the following: a portable computer diskette, a hard disk, a Random Access Memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or flash memory), a Static Random Access Memory (SRAM), a portable compact disc read-only memory (CD-ROM), a Digital Versatile Disc (DVD), a memory stick, a floppy disk, a mechanical coding device, such as punch cards or in-groove projection structures having instructions stored thereon, and any suitable combination of the foregoing. Computer-readable storage media as used herein is not to be construed as transitory signals per se, such as radio waves or other freely propagating electromagnetic waves, electromagnetic waves propagating through a waveguide or other transmission medium (e.g., optical pulses through a fiber optic cable), or electrical signals transmitted through electrical wires.
The computer-readable program instructions described herein may be downloaded from a computer-readable storage medium to a respective computing/processing device, or to an external computer or external storage device via a network, such as the internet, a local area network, a wide area network, and/or a wireless network. The network may include copper transmission cables, fiber optic transmission, wireless transmission, routers, firewalls, switches, gateway computers and/or edge servers. The network adapter card or network interface in each computing/processing device receives computer-readable program instructions from the network and forwards the computer-readable program instructions for storage in a computer-readable storage medium in the respective computing/processing device.
The computer program instructions for carrying out operations of the present disclosure may be assembler instructions, Instruction Set Architecture (ISA) instructions, machine-related instructions, microcode, firmware instructions, state setting data, or source or object code written in any combination of one or more programming languages, including an object oriented programming language such as Smalltalk, C + + or the like and conventional procedural programming languages, such as the "C" programming language or similar programming languages. The computer-readable program instructions may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server. In the case of a remote computer, the remote computer may be connected to the user's computer through any type of network, including a Local Area Network (LAN) or a Wide Area Network (WAN), or the connection may be made to an external computer (for example, through the Internet using an Internet service provider). In some embodiments, the electronic circuitry that can execute the computer-readable program instructions implements aspects of the present disclosure by utilizing the state information of the computer-readable program instructions to personalize the electronic circuitry, such as a programmable logic circuit, a Field Programmable Gate Array (FPGA), or a Programmable Logic Array (PLA).
Various aspects of the present disclosure are described herein with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems) and computer program products according to embodiments of the disclosure. It will be understood that each block of the flowchart illustrations and/or block diagrams, and combinations of blocks in the flowchart illustrations and/or block diagrams, can be implemented by computer-readable program instructions.
These computer-readable program instructions may be provided to a processor of a general purpose computer, special purpose computer, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks. These computer-readable program instructions may also be stored in a computer-readable storage medium that can direct a computer, programmable data processing apparatus, and/or other devices to function in a particular manner, such that the computer-readable medium storing the instructions comprises an article of manufacture including instructions which implement the function/act specified in the flowchart and/or block diagram block or blocks.
The computer readable program instructions may also be loaded onto a computer, other programmable data processing apparatus, or other devices to cause a series of operational steps to be performed on the computer, other programmable apparatus or other devices to produce a computer implemented process such that the instructions which execute on the computer, other programmable apparatus or other devices implement the functions/acts specified in the flowchart and/or block diagram block or blocks.
The flowchart and block diagrams in the figures illustrate the architecture, functionality, and operation of possible implementations of systems, methods and computer program products according to various embodiments of the present disclosure. In this regard, each block in the flowchart or block diagrams may represent a module, segment, or portion of instructions, which comprises one or more executable instructions for implementing the specified logical function(s). In some alternative implementations, the functions noted in the block may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved. It will also be noted that each block of the block diagrams and/or flowchart illustration, and combinations of blocks in the block diagrams and/or flowchart illustration, can be implemented by special purpose hardware-based systems which perform the specified functions or acts, or combinations of special purpose hardware and computer instructions. It is well known to those skilled in the art that implementation by hardware, by software, and by a combination of software and hardware are equivalent.
Having described embodiments of the present disclosure, the foregoing description is intended to be exemplary, not exhaustive, and not limited to the disclosed embodiments. Many modifications and variations will be apparent to those of ordinary skill in the art without departing from the scope and spirit of the described embodiments. The terminology used herein is chosen in order to best explain the principles of the embodiments, the practical application, or improvements made to the technology in the marketplace, or to enable others of ordinary skill in the art to understand the embodiments disclosed herein. The scope of the present disclosure is defined by the appended claims.
Claims (10)
1. A lens test system is characterized by comprising a processing device, a projection device, a camera device, a display device and a mobile device, wherein the lens is arranged on the mobile device and is positioned between the projection device and the display device,
the projection device is used for projecting the first image provided by the processing device through the lens;
the moving device is used for controlling the lens to move between the projection device and the display device;
the display device is used for respectively displaying a second image after the first image at each moving position is projected through the lens;
the camera device is used for acquiring each second image and sending each second image to the processing device;
the processing device is used for identifying each second image to judge whether the lens is qualified.
2. The system of claim 1, wherein the moving means comprises a motor and a carriage,
the bracket is used for placing the lens;
the motor is used for controlling the lens to move between the projection device and the display device.
3. The system of claim 2, wherein the first image includes a plurality of target regions therein,
the plurality of target areas at least comprise a first central area positioned in the center of the first image, and other target areas except the first central area are symmetrically distributed based on the first central area;
correspondingly, the second image comprises a plurality of target areas, the target areas at least comprise a second central area located in the center of the second image, and other target areas except the second central area are symmetrically distributed based on the second central area.
4. The system of claim 3, wherein the processing device is configured to:
respectively identifying each second image to obtain the definition of each target area in each second image;
determining a target focal length of the lens in each target area according to the definition of each target area in each second image;
and determining the diopter of the lens in each target area according to the target focal length of the lens in each target area, and judging whether the lens is qualified or not according to the diopter of the lens in each target area.
5. The system of claim 4, wherein the processing device is configured to:
acquiring the target definition of the lens in each target area according to the definition of each target area in each second image; and the number of the first and second groups,
and determining the target focal length of the corresponding lens in each target area according to the target definition of the lens in each target area.
6. System according to claim 5, characterized in that the motor is provided with graduation marks,
and the scale values of the scale marks correspond to the positions of the motor in the moving process one by one.
7. The system of claim 6, wherein the processing device is configured to:
acquiring the position of the motor corresponding to each target definition according to the target definition of the lens in each target area, and taking the position as a target position;
and determining the target focal length of the lens in each target area according to each target position and the position between the projection device and the display device.
8. The system of claim 4,
the processing device is used for comparing the diopter of the lens in each target area with a target diopter threshold value, and determining that the lens is qualified under the condition that the diopter of the lens in each target area is greater than or equal to the target diopter threshold value.
9. The system of claim 8,
the processing device is used for providing a configuration interface for configuring the diopter threshold value and acquiring the diopter threshold value input through the configuration interface as the target diopter threshold value.
10. The system of claim 1, wherein the lens is comprised of at least one lens.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110995586.XA CN113790874A (en) | 2021-08-27 | 2021-08-27 | Lens test system |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110995586.XA CN113790874A (en) | 2021-08-27 | 2021-08-27 | Lens test system |
Publications (1)
Publication Number | Publication Date |
---|---|
CN113790874A true CN113790874A (en) | 2021-12-14 |
Family
ID=79182217
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202110995586.XA Pending CN113790874A (en) | 2021-08-27 | 2021-08-27 | Lens test system |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN113790874A (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN115661158A (en) * | 2022-12-28 | 2023-01-31 | 歌尔光学科技有限公司 | Method and device for detecting optical-mechanical lens assembly, electronic equipment and storage medium |
Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2011151165A (en) * | 2010-01-21 | 2011-08-04 | Casio Computer Co Ltd | Wafer inspection device, and wafer inspection method |
CN102252821A (en) * | 2011-07-25 | 2011-11-23 | 福州福赛特光学仪器有限公司 | Parameter detection method for optical lens and detecting instrument thereof |
CN105371885A (en) * | 2015-10-14 | 2016-03-02 | 凌云光技术集团有限责任公司 | Automatic test method and system of imaging chip MTF value |
US20180202941A1 (en) * | 2016-04-12 | 2018-07-19 | Nippon Steel & Sumitomo Metal Corporation | Inspection object imaging apparatus, inspection object imaging method, surface inspection apparatus, and surface inspection method |
CN109413334A (en) * | 2018-12-13 | 2019-03-01 | 浙江舜宇光学有限公司 | Image pickup method and filming apparatus |
CN110568704A (en) * | 2019-09-02 | 2019-12-13 | 歌尔股份有限公司 | Ultra-short-focus lens detection system and detection method |
CN111679541A (en) * | 2020-07-16 | 2020-09-18 | 歌尔光学科技有限公司 | Calibration control method and system for projection module, storage medium and control device |
CN111929038A (en) * | 2020-08-20 | 2020-11-13 | 歌尔光学科技有限公司 | Testing device and method of micro lens, testing equipment and computer storage medium |
CN113042400A (en) * | 2021-03-26 | 2021-06-29 | 裴方虚 | Optical glass lens testing system and testing method thereof |
-
2021
- 2021-08-27 CN CN202110995586.XA patent/CN113790874A/en active Pending
Patent Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2011151165A (en) * | 2010-01-21 | 2011-08-04 | Casio Computer Co Ltd | Wafer inspection device, and wafer inspection method |
CN102252821A (en) * | 2011-07-25 | 2011-11-23 | 福州福赛特光学仪器有限公司 | Parameter detection method for optical lens and detecting instrument thereof |
CN105371885A (en) * | 2015-10-14 | 2016-03-02 | 凌云光技术集团有限责任公司 | Automatic test method and system of imaging chip MTF value |
US20180202941A1 (en) * | 2016-04-12 | 2018-07-19 | Nippon Steel & Sumitomo Metal Corporation | Inspection object imaging apparatus, inspection object imaging method, surface inspection apparatus, and surface inspection method |
CN109413334A (en) * | 2018-12-13 | 2019-03-01 | 浙江舜宇光学有限公司 | Image pickup method and filming apparatus |
CN110568704A (en) * | 2019-09-02 | 2019-12-13 | 歌尔股份有限公司 | Ultra-short-focus lens detection system and detection method |
CN111679541A (en) * | 2020-07-16 | 2020-09-18 | 歌尔光学科技有限公司 | Calibration control method and system for projection module, storage medium and control device |
CN111929038A (en) * | 2020-08-20 | 2020-11-13 | 歌尔光学科技有限公司 | Testing device and method of micro lens, testing equipment and computer storage medium |
CN113042400A (en) * | 2021-03-26 | 2021-06-29 | 裴方虚 | Optical glass lens testing system and testing method thereof |
Non-Patent Citations (2)
Title |
---|
曾明智;李海峰;: "投影机镜头的MTF实时检测系统", 《光电工程》, vol. 35, no. 02, 15 February 2008 (2008-02-15), pages 135 - 139 * |
朱宇栋: "光学镜片外观瑕疵视觉检测技术研究及实现", 《中国优秀硕士学位论文全文数据库工程科技Ⅱ辑》, no. 03, 15 March 2020 (2020-03-15), pages 030 - 29 * |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN115661158A (en) * | 2022-12-28 | 2023-01-31 | 歌尔光学科技有限公司 | Method and device for detecting optical-mechanical lens assembly, electronic equipment and storage medium |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US11933604B2 (en) | Detection method and apparatus for automatic driving sensor, and electronic device | |
EP3291004A1 (en) | Ranging method, automatic focusing method and device | |
US11835417B2 (en) | Method, apparatus and device for detecting field curvature of VR optical module | |
CN107578047B (en) | Eccentricity detection method for power cable | |
KR20200018411A (en) | Method and apparatus for detecting burr of electrode piece | |
TW201510878A (en) | Measurement device | |
US20210093227A1 (en) | Image processing system and control method thereof | |
CN103837828A (en) | Travel curve test method and system of camera-shooting mode motor and mobile terminal | |
CN108055532A (en) | Automate the method and apparatus of matching test card | |
KR20210058931A (en) | Signal indicator state detection method and device, operation control method and device | |
CN112595496A (en) | Method, device and equipment for detecting defects of near-eye display equipment and storage medium | |
CN113790874A (en) | Lens test system | |
JP2006309405A (en) | Meter recognition system, meter recognition method, and meter recognition program | |
US10965858B2 (en) | Image processing apparatus, control method thereof, and non-transitory computer-readable storage medium for detecting moving object in captured image | |
JP2018182593A (en) | Image processing apparatus and image processing method | |
US11119304B2 (en) | Determining an error in a moving distance of a microscope | |
WO2015096824A1 (en) | Analysis device and analysis method | |
CN105427315B (en) | Digital instrument image position testing method and device | |
JP7467107B2 (en) | Image processing device, image processing method, and program | |
CN106846401B (en) | Detection method and equipment of double-camera module | |
JP2016114785A (en) | Imaging device, imaging method and imaging program | |
US20230206425A1 (en) | Systems and methods for alignment and comparison of a reference drawing with images of a part | |
CN115496759B (en) | Dust detection method and device and storage medium | |
JP2019139588A (en) | Image analysis apparatus and image analysis method | |
US11070716B2 (en) | Image sensor system, image sensor, data generation method of image sensor in image sensor system, and non-transitory computer-readable recording medium |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination |