US20130088586A1 - Magnification Observation Device - Google Patents
Magnification Observation Device Download PDFInfo
- Publication number
- US20130088586A1 US20130088586A1 US13/559,799 US201213559799A US2013088586A1 US 20130088586 A1 US20130088586 A1 US 20130088586A1 US 201213559799 A US201213559799 A US 201213559799A US 2013088586 A1 US2013088586 A1 US 2013088586A1
- Authority
- US
- United States
- Prior art keywords
- stage
- axis
- image
- imaging
- imaging region
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Images
Classifications
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B21/00—Microscopes
- G02B21/24—Base structure
- G02B21/26—Stages; Adjusting means therefor
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B21/00—Microscopes
- G02B21/36—Microscopes arranged for photographic purposes or projection purposes or digital imaging or video purposes including associated control and data processing arrangements
- G02B21/365—Control or image processing arrangements for digital or video microscopes
Definitions
- the present invention relates to a magnification observation device.
- a magnification observation device provided with a microscope, for observing an observation object.
- the microscope includes an imaging unit for imaging the observation object.
- An image of the observation object imaged by the imaging unit is displayed by a display part.
- a user can acquire detailed information of the observation object based on the image displayed by the display part.
- a microscope described in Japanese Unexamined Patent Publication No. 2006-337470 has a camera for measuring an observation object and a stage on which the observation object is placed.
- the camera is attached to a camera attaching part so as to be along a Z-axis direction (perpendicular direction).
- the stage is rotatably provided below the camera.
- the stage is made up of an X-Y stage movable in an X-axis direction and a Y-axis direction. The stage is moved along the X-axis direction and the Y-axis direction, to thereby align the observation object with respect to an imaging region of the camera.
- a rotational axis of the stage is fixed, and an X-axis and a Y-axis of the stage are rotated in association with rotation of the stage.
- a moving direction of the stage is difficult to be controlled and the observation object cannot be moved in a desired direction.
- a magnification observation device includes: an imaging part that images an observation object, to acquire imaging data of an imaging region; a display part that displays an image of the observation object as an observed image based on the image data acquired by the imaging part; a stage that has a placement surface on which the observation object is placed, and is provided relatively rotatably with respect to the imaging part around a rotational axis substantially vertical to the placement surface, and movably along first and second axes intersecting with each other within a plane substantially parallel to the placement surface; a rotational angle detecting part that detects a rotational angle of the stage; a stage driving part that moves the stage relatively with respect to the imaging part along the first and second axes; an instruction accepting part that accepts an instruction for a moving direction of the stage; and a control part that provides the stage driving part with moving amounts of the stage along the first and second axes, to control movement of the stage, wherein the first and second axes of the stage rotate around the rotational axis integrally with the stage,
- the stage is provided relatively rotatably with respect to the imaging part around the rotational axis substantially vertical to the placement surface, and movably along the first and second axes intersecting with each other within the plane substantially parallel to the placement surface.
- the observation object is placed on the placement surface of the stage.
- the observation object is imaged by the imaging part, to acquire image data of the imaging region. Based on the image data acquired by the imaging part, the image of the observation object is displayed by the display part as an observed image.
- the moving amounts of the stage along the first and second axes are provided from the control part to the stage driving part. Based on the provided moving amounts, the stage is moved by the stage driving part along the first and second axes, relatively with respect to the imaging part. Accordingly, the observation object moves relatively with respect to the imaging region.
- the first and second axes of the stage rotate around the rotational axis integrally with the stage.
- the stage can be moved without moving the rotational axis of the stage with respect to the imaging part.
- a desired region of the observation object can be prevented from going outside the imaging region due to rotation of the stage. It is therefore possible to change an orientation of the observation object while imaging the desired area of the imaging object.
- the moving amounts along the first and second axes to be provided to the stage driving part are controlled such that the moving direction of the observation object in the imaging region of the imaging part agrees with the direction accepted by the instruction accepting part, based on the rotational angle detected by the rotational angle detecting part. Therefore, even when the stage is rotated, the observation object can be easily moved in a desired direction in the imaging region.
- the imaging region of the imaging part may be provided with third and fourth axes intersecting with each other, and the control part may be configured to detect an angle difference between the first and second axes of the stage and the third and fourth axes of the imaging region during predetermined calibration, and the moving amounts of the stage along the first and second axes to be provided to the stage driving part may be corrected based on the detected angle difference in normal time.
- the observation object can be moved with the third and fourth axes of the imaging region taken as references. This allows the user to easily move the observation object in the imaging region while viewing the observation object displayed in the display part, without recognizing the orientation of the imaging part.
- the control part may be configured to detect the angle difference between the first and second axes of the stage and the third and fourth axes of the imaging region based on the moving direction of the observed image in a case where the stage has been moved by the stage driving part.
- the angle difference between the first and second axes of the stage and the third and fourth axes of the imaging region can be easily detected, to thereby correct the moving amounts of the stage along the first and second axes by means of the detected angle difference.
- a direction specifying part for specifying directions of the first and second axes may be provided on the stage, and the control part may be configured to detect the angle difference between the first and second axes of the stage and the third and fourth axes of the imaging region based on an image of the direction specifying part included in the observed image.
- the angle difference between the first and second axes of the stage and the third and fourth axes of the imaging region can be easily detected, to thereby correct the moving amounts of the stage along the first and second axes by means of the detected angle difference.
- the display part may display a position specifying image indicating a position of the imaging region on the stage.
- the user can easily recognize the position of the imaging region on the stage. This allows the user to efficiently align the desired region of the observation object to the imaging region.
- the position specifying image may include a first image representing a region wider than the imaging region on the stage and a second image corresponding to the imaging region, and the control part may rotate the first image relatively with respect to the second image based on the rotational angle detected by the rotational angle detecting part.
- the user can easily recognize a positional relation between the stage and the imaging region and a rotational angle of the stage with respect to the imaging region. This allows the user to efficiently align the desired region of the observation object to the imaging region.
- the imaging part may be configured to be capable of imaging the region wider than the imaging region, and the display part may display the first image based on image data obtained upon imaging a wide range by the imaging part.
- the user can easily recognize a positional relation between the observation object on the stage and the imaging region. This allows the user to efficiently align the desired region of the observation object to the imaging region.
- the first image may show a diagram for specifying the region wider than the imaging region on the stage.
- the first image can be easily displayed in the display part without increasing a process load of the control part.
- the first image may correspond to a relatively movable range of the imaging region on the stage.
- the user can easily recognize a relative movable range of the imaging region on the stage. This allows the user to efficiently align the desired region of the observation object to the imaging region.
- the instruction accepting part may be configured to be capable of accepting designation of an arbitrary position in the position specifying image, and the control part may control movement of the stage by the stage driving part such that the imaging region moves to a position on the stage which corresponds to the position designated by the instruction accepting part.
- the user designates a desired position on the position specifying image, and thus the stage moves such that the position on the stage which corresponds to the designated position becomes the imaging region. This allows the user to efficiently align the desired region of the observation object to the imaging region.
- the control part may be configured to detect a moving amount of the observed image in the case where the stage has been moved, calculate a relation between the moving amount of the observed image and the moving amount of the stage based on the detected moving amount, and correct the moving amount to be provided to the stage driving part based on the calculated relation.
- the present invention it is possible to easily move an observation object in a desired direction even when the stage is rotated, and also prevent a desired region of the observation object from going outside an imaging region due to rotation of the stage.
- FIG. 1 is a block diagram showing a configuration of a magnification observation device according to one embodiment of the present invention
- FIG. 2 is a perspective view showing a microscope of the magnification observation device according to one embodiment of the present invention
- FIG. 3 is a schematic view showing a state where an imaging unit of the microscope is fixed parallel to a Z-axis;
- FIG. 4 is a schematic view showing a state where the imaging unit of the microscope is inclined at a desired angle from the Z-axis;
- FIGS. 5A to 5C are schematic views showing an example of a method for calculating an x t moving amount and a y t moving amount
- FIG. 6 is a flowchart for a stage driving process by a CPU
- FIG. 7 is a plan view showing an example of an observation object placed on a placement surface of the stage
- FIGS. 8A and 8B are views showing a relation between an operation of a stage unit and an observed image displayed in a display part
- FIGS. 9A and 9B are views showing a relation between an operation of the stage unit and an observed image displayed in the display part;
- FIGS. 10A and 10B are views showing a relation between an operation of the stage unit and an observed image displayed in the display part;
- FIGS. 11A and 11B are views showing a relation between an operation of the stage unit and an observed image displayed in the display part;
- FIGS. 12A and 12B are views showing a relation between an operation of the stage unit and an observed image displayed in the display part;
- FIGS. 13A and 13B are views for describing correction of a moving direction of the stage
- FIGS. 14A and 14B are views for describing correction of a moving direction of the stage
- FIG. 15 is a flowchart for an inclination angle acquiring process by the
- FIG. 16 is a view showing an example of a position specifying image
- FIG. 17 is a view showing an example of display of the position specifying image by the display part
- FIG. 18 is a view showing an example of display of the position specifying image by the display part.
- FIGS. 19A and 19B are views showing an example of display of the position specifying image by the display part.
- FIG. 1 is a block diagram showing a configuration of a magnification observation device according to one embodiment of the present invention.
- a magnification observation device 300 is provided with a microscope 100 and an image processing apparatus 200 .
- the microscope 100 includes an imaging unit 10 , a stage unit 20 , and a rotational angle sensor 30 .
- the imaging unit 10 includes a color CCD (charge coupled device) 11 , a half mirror 12 , an object lens 13 , an A/D converter (analog/digital converter) 15 , an illumination light source 16 , and a lens driving part 17 .
- the stage unit 20 includes a stage 21 , a stage driving part 22 , and a stage supporting part 23 . An observation object S is placed on the stage 21 .
- an X-axis and a Y-axis intersecting with each other within a horizontal plane, and a Z-axis (perpendicular direction) vertical to the X-axis and the Y-axis are defined.
- a direction parallel to the X-axis is referred to as an X-axis direction
- a direction parallel to the Y-axis is referred to as a Y-axis direction
- a direction parallel to the Z-axis is referred to as a Z-axis direction.
- the illumination light source 16 is, for example, a halogen lamp or a white light LED (light-emitting diode) which generates white light.
- White light generated by the illumination light source 16 is reflected by the half mirror 12 , and thereafter collected by the object lens 13 onto the observation object S on the stage 21 .
- the white light reflected by the observation object S is transmitted through the object lens 13 and the half mirror 12 , and incident on the color CCD 11 .
- the color CCD 11 has a plurality of pixels for red that receive red wavelength light, a plurality of pixels for green that receive green wavelength light, and a plurality of pixels for blue that receive blue wavelength light.
- the plurality of pixels for red, the plurality of pixels for green, and the plurality of pixels for blue are two-dimensionally arrayed. From each of the pixels in the color CCD 11 , an electric signal corresponding to a light receiving amount is outputted.
- the output signal of the color CCD 11 is converted to a digital signal by the A/D converter 15 .
- the digital signal outputted from the A/D converter 15 is sequentially provided as image data to the image processing apparatus 200 .
- an imaging element such as a CMOS (complementary metal oxide semiconductor) image sensor may be used.
- the object lens 13 is, for example, a zoom lens, and is configured such that its magnification is manually or automatically changeable.
- the magnification of the object lens 13 is detected by a magnification detecting part (not shown). Further, the object lens 13 is provided movably in the Z-axis direction.
- the lens driving part 17 moves the object lens 13 in the Z-axis direction by control of the image processing apparatus 200 . Thereby, a position of a focus of the imaging unit 10 is moved in the Z-axis direction.
- the stage 21 is rotatably provided on the stage supporting part 23 around a rotational axis parallel to the Z-axis.
- the stage driving part 22 moves the stage 21 along a later-mentioned x-axis and y-axis relatively with respect to the imaging unit 10 based on a movement command signal (drive pulse) provided from the image processing apparatus 200 . Further, the stage driving part 22 rotates the stage 21 around the rotational axis parallel to the Z-axis relatively with respect to the imaging unit 10 based on the rotation command signal (drive pulse) provided from the image processing apparatus 200 .
- a position of the rotational axis of the stage 21 is constant with respect to the stage supporting part 23 .
- the stage driving part 22 uses a stepping motor.
- the rotational angle sensor 30 detects a rotational angle of the stage 21 , and provides the image processing apparatus 200 with an angle detection signal indicating the detected angle.
- the image processing apparatus 200 includes an interface 210 , a CPU (central processing unit) 220 , a ROM (read only memory) 230 , a storage unit 240 , an input unit 250 , a display part 260 , and an operation memory 270 .
- a CPU central processing unit
- ROM read only memory
- a system program is stored into the ROM 230 .
- the storage unit 240 is made up of a hard disk and the like. In the storage unit 240 , a variety of control programs and programs are stored, and a variety of data, such as image data provided from the microscope 100 through the interface 210 , are also stored.
- the input unit 250 includes a keyboard and a pointing device, and is configured to be able to accept a movement instruction and a rotation instruction for the stage 21 from the user or an external device.
- the movement instruction includes a moving direction and a moving amount
- the rotation instruction includes a rotational direction and a rotational angle.
- the pointing device a mouse, a joystick, or the like is used.
- the display part 260 is configured, for example, by a liquid crystal display panel or an organic EL (electroluminescent) panel. As described later, the display part 260 is set with an x d -axis and a y d -axis intersecting with each other.
- the x d -axis corresponds to a lateral direction of the screen of the display part 260
- the y d -axis corresponds to a vertical direction of the screen of the display part 260 .
- the operation memory 270 is made up of a RAM (random access memory), and used for processing a variety of data.
- the CPU 220 executes the control programs stored in the storage unit 240 , to perform a variety of processes by means of the operation memory 270 , and displays an image based on the image data in the display part 260 . Further, the CPU 220 controls the color CCD 11 , the illumination light source 16 , the lens driving part 17 , and the stage driving part 22 of the microscope 100 through the interface 210 .
- FIG. 2 is a perspective view showing the microscope 100 of the magnification observation device 300 according to one embodiment of the present invention.
- the X-axis, the Y-axis, and the Z-axis are indicated by arrows.
- the microscope 100 has a base 1 .
- a first supporting base 2 is attached onto the base 1
- a second supporting base 3 is also attached to the front surface of the first supporting base 2 so as to be embedded thereinto.
- a connecting part 4 is rotatably attached to the top edge of the first supporting base 2 around a rotational axis R 1 extending in the Y-axis direction.
- a rotational column 5 is attached to the connecting part 4 .
- the rotational column 5 is inclinable within a vertical plane parallel to the Z-axis with the rotational axis R 1 taken as a fulcrum point in association with rotation of the connecting part 4 .
- the user can fix the connecting part 4 to the first supporting base 2 by means of a fixing knob 9 .
- a circular supporting part 7 is attached to the front surface of a connecting part 6 .
- a substantially tubular imaging unit 10 is attached to the supporting part 7 .
- a light axis R 2 of the imaging unit 10 is parallel to the Z-axis.
- the supporting part 7 has a plurality of adjustment screws 41 for moving the imaging unit 10 within a horizontal plane. It is possible to adjust a position of the imaging unit 10 such that the light axis R 2 of the imaging unit 10 vertically intersects with a rotational axis R 1 by means of the plurality of adjustment screws 41 .
- a slider 8 is attached, slidably in the Z-axis direction, to the front surface of the second supporting base 3 on the base 1 .
- An adjustment knob 42 is provided on the side surface of the second supporting base 3 . A position of the slider 8 in the Z-axis direction (height direction) is adjustable by the adjustment knob 42 .
- the supporting part 23 of the stage unit 20 is attached onto the slider 8 .
- the stage 21 is rotationally provided around a rotational axis R 3 parallel to the Z-axis with respect to the supporting part 23 .
- the stage 21 has a placement surface 21 a on which an observation object S is placed. Further, an x t -axis and a y t -axis intersecting with each other within the horizontal plane are set on the stage 21 .
- the stage 21 is provided movably along the x t -axis and the y t -axis by the stage driving part 22 of FIG. 1 .
- the stage 21 rotates around the rotational axis R 3
- the x t -axis and the y t -axis of the stage 21 also rotate. Thereby, the x t -axis and the y t -axis of the stage 21 are inclined with respect to the X-axis and the Y-axis within a horizontal plane.
- An imaging range (visual field range) of the imaging unit 10 on the placement surface 21 a of the stage 21 varies depending on a magnification of the imaging unit 10 (magnification of the object lens 13 ).
- the imaging range of the imaging unit 10 is referred to as an imaging region, and as described later, an x r -axis and a y r -axis intersecting with each other are set in the imaging region.
- the x r -axis and the y r -axis of the imaging region respectively correspond to the x d -axis and the y d -axis of the display part 260 .
- an image of the observation object S is displayed in the display part 260 of FIG. 1 .
- the image of the observation object S displayed in the display part 260 is referred to as an observed image.
- the observed image is set with an x i -axis and a y i -axis intersecting with each other.
- the x i -axis and the y i -axis of the observed image respectively correspond to the x t -axis and the y t -axis of the stage 21 .
- FIG. 3 is a schematic view showing a state where the imaging unit 10 of the microscope 100 is fixed parallel to the Z-axis.
- FIG. 4 is a schematic view showing a state where the imaging unit 10 of the microscope 100 is inclined at a desired angle with respect to the Z-axis.
- the fixing knob 9 is fastened, to fix the connecting part 4 to the second supporting base 3 .
- the light axis R 2 of the imaging unit 10 vertically intersects with the rotational axis R 1 while being in a parallel state to the Z-axis.
- the light axis R 2 of the imaging unit 10 is vertical to the surface of the stage 21 .
- the fixing knob 9 is loosened, to make the connecting part 4 rotatable around the rotational axis R 1 , and the rotational column 5 inclinable with the rotational axis R 1 taken as a fulcrum point. Therefore, as shown in FIG. 4 , the light axis R 2 of the imaging unit 10 is inclinable at an arbitrary angle ⁇ 1 with respect to the Z-axis. In this case, the light axis R 2 of the imaging unit 10 vertically intersects with the rotational axis R 1 . Similarly, the light axis R 2 of the imaging unit 10 is inclinable at an arbitrary angle on the side opposite to the side in FIG. 4 with respect to the Z-axis.
- a height of the surface of the observation object S on the stage 21 can be made to agree with a height of the rotational axis R 1 , to thereby observe the same portion of the observation object in a vertical direction and an oblique direction.
- the CPU 220 Upon provision of a movement instruction for the stage 21 or a rotation instruction for the stage 21 from the user or the external device to the input unit 250 of the image processing apparatus 200 ( FIG. 1 ), the CPU 220 provides a movement command signal or a rotation command signal to the stage driving part 22 .
- the stage driving part 22 moves the stage 21 along the x t -axis and the y t -axis based on the movement command signal from the CPU 220 , or rotates the stage 21 around the rotational axis R 3 based on the rotation command signal from the CPU 220 .
- the movement command signal from the CPU 220 includes a moving amount of the stage 21 along the x t -axis (hereinafter, referred to as x t moving amount), and a moving amount of the stage 21 along the y t -axis (hereinafter, referred to as y t moving amount).
- the rotation command signal from the CPU 220 includes a rotational direction and a rotational angle of the stage 21 .
- the x t -axis and the y t -axis of the stage 21 are inclined with respect to the X-axis and the Y-axis. For this reason, in order to move the stage 21 in a moving direction provided as a movement instruction, it is necessary to control of the x t moving amount and the y t moving amount in accordance with the rotational angle of the stage 21 .
- the x t moving amount and the y t moving amount are calculated such that the moving direction provided as the movement instruction agrees with an actual moving direction of the stage 21 .
- FIGS. 5A to 5C are schematic views showing an example of a method for calculating an x t moving amount and a y t moving amount.
- the x t -axis and the y t -axis are inclined at an angle ⁇ 0 with respect to the X-axis and the Y-axis.
- the input unit 250 is provided with a moving direction A 0 and a moving amount B 0 as a movement instruction. In this case, as shown in FIG.
- a moving amount dx 1 along the X-axis and a moving amount dy 1 along the Y-axis are calculated based on the moving direction A 0 and the moving amount B 0 .
- an x t moving amount dx 2 and a y t moving amount dy 2 are calculated based on the calculated moving amounts dx 1 , dy 1 and the rotational angle ⁇ 0 of the stage 21 detected by the rotational angle sensor 30 .
- FIG. 6 is a flowchart for a stage driving process by the CPU 220 .
- the CPU 220 performs the stage driving process of FIG. 6 based on the control program stored in the storage unit 240 .
- the CPU 220 determines whether or not a movement instruction of the stage 21 has been provided to the input unit 250 (step S 1 ).
- the CPU 220 acquires a rotational angle of the stage 21 based on an angle detection signal from the rotational angle sensor 30 (step S 2 ).
- the CPU 220 calculates an x t moving amount and a y t moving amount as described above based on the provided movement instruction and the acquired rotational angle (step S 3 ).
- the CPU 220 provides the calculated x t moving amount and y t moving amount as a movement command signal to the stage driving part 22 (step S 4 ). Thereafter, the CPU 220 returns to the process of step S 1 .
- step S 1 when the movement instruction of the stage 21 has not been provided, the CPU 220 determines whether or not a rotation instruction of the stage 21 has been provided to the input unit 250 (step S 5 ). When the rotation instruction of the stage 21 has not been provided, the CPU 220 returns to the process of step S 1 . When the rotation instruction of the stage 21 has been provided, the CPU 220 provides a rotational direction and a rotational angle as a rotation command signal to the stage driving part 22 based on the provided rotation instruction (step S 6 ). Thereafter, the CPU 220 returns to the process of step S 1 .
- the x t moving amount and the y t moving amount are calculated based on the movement instruction provided to the input unit 250 and the rotational angle detected by the rotational angle sensor 30 , and the calculated x t moving amount and the y t moving amount are provided to the stage driving part 22 .
- the stage 21 is moved from the stage driving part 22 along the x t -axis and the y t -axis. Therefore, even when the stage 21 is rotated, the observation object S can be moved in a desired direction in the imaging region.
- the rotational angle of the stage 21 detected by the rotational angle sensor 30 , may be displayed by the display part 260 . In this case, the user can easily recognize the rotational angle of the stage 21 .
- a coordinate (hereinafter, referred to as target coordinate) showing a position to which a movement is made may be provided as a movement instruction.
- a coordinate hereinafter, referred to as target coordinate
- a target coordinate in the XY coordinate system is provided as a movement instruction.
- the provided target coordinate is converted to a coordinate in the x t y t coordinate system based on the rotational angle of the stage 21 . Based on the converted target coordinate, an x t moving amount and a y t moving amount can be calculated.
- an x t moving amount and a y t moving amount may be calculated using the XY coordinate system and the x t y t coordinate system in a similar manner to the above. For example, a target coordinate in the XY coordinate system is calculated based on the provided moving amount and the moving direction. The calculated target coordinate is converted to a coordinate in the x t y t coordinate system based on a rotational angle of the stage 21 . Based on the converted target coordinate, an x t moving amount and a y t moving amount can be calculated.
- FIG. 7 is a plan view showing an example of the observation object S placed on the placement surface 21 a of the stage 21 .
- FIGS. 8A to 10B are views each showing a relation between an operation of the stage unit 20 and an observed image displayed in the display part 260 .
- FIGS. 8A , 9 A, and 10 A each show a positional relation between the stage 21 and the imaging region
- FIGS. 8B , 9 B, and 10 B each show an observed image displayed in the display part 260 .
- FIGS. 8A , 9 A, and 10 A each show a positional relation between the stage 21 and the imaging region
- FIGS. 8B , 9 B, and 10 B each show an observed image displayed in the display part 260 .
- FIGS. 8A , 9 A, and 10 A each show the X-axis and the Y-axis, the x t -axis and the y t -axis of the stage 21 , and the x r -axis and the y r -axis of the imaging region.
- FIGS. 8B , 9 B, and 10 B each show the x d -axis and the y d -axis of the display part 260 and the xi-axis and the y i -axis of the observed image.
- a direction An (n is a natural number) is a direction of movement or rotation of the stage 21
- a direction Cn is a direction of movement or rotation of an observed image IM.
- the direction An which is relative with respect to the X-axis and the Y-axis agrees with the direction Cn which is relative with respect to the x d -axis and the y d -axis of the display part 260 .
- a moving amount Bn is a moving amount of the stage 21
- a moving amount Dn is a moving amount of the observed image.
- a ratio between the moving amount Bn and the moving amount Dn is constant.
- the x t -axis and the y t -axis of the stage 21 respectively agree with the X-axis and the Y-axis.
- the x r -axis and the y r -axis of an imaging region IR respectively agree with the X-axis and the Y-axis.
- the x i -axis and the y i -axis of the observed image IM respectively agree with the x d -axis and the y d -axis of the display part 260 .
- FIG. 9A when a rotational direction A 2 and a rotational angle ⁇ 2 are provided as a rotation instruction to the input unit 250 , the stage 21 is rotated just by the angle ⁇ 2 in the direction A 2 around the rotational axis R 3 .
- FIG. 9B the observed image IM rotates just by the angle ⁇ 2 in a rotational direction C 2 .
- the position of the imaging unit 10 is adjusted such that the rotational axis R 3 is located at the central part of the imaging region IR. Hence, even when the stage 21 is rotated around the rotational axis R 3 , a desired region of the observation object S is prevented from going outside the imaging region IR. It is thus possible to rotate the observation object S while imaging the desired region of the observation object S by the imaging unit 10 .
- the desired region of the observation object S is imaged from a variety of angles. This allows the user to stereoscopically observe the desired region of the observation object S.
- the position of the rotational axis R 3 of the stage 21 is fixed with respect to the stage supporting part 23 in the present example, this is not restrictive.
- the position of the rotational axis R 3 of the stage 21 may be changeable with respect to the stage supporting part 23 . In this case, the position of the rotational axis R 3 of the stage 21 can be adjusted such that the rotational axis R 3 is located at the central part of the imaging region IR.
- the stage 21 moves just by the moving amount B 3 in the moving direction A 3 .
- an x t moving amount and a y t moving amount are calculated based on the moving direction A 3 , the moving amount B 3 , and the angle ⁇ 2, and the stage 21 is moved based on the x t moving amount and the y t moving amount.
- the observed image IM moves just by a moving amount D 3 in a direction C 3 .
- the moving direction of the stage 21 with respect to the X-axis and the Y-axis agrees with the moving direction of the observed image IM with respect to the x d -axis and the y d -axis of the display part 260 .
- the x r -axis and the y r -axis of the imaging region IR are displaced with respect to the X-axis and the Y-axis due to an assembly error or the like.
- the imaging unit 10 when the imaging unit 10 is provided detachably from the supporting part 7 , the imaging unit 10 may be attached to the supporting part 7 with the x r -axis and the y r -axis of the imaging region IR in a displaced state with respect to the X-axis and the Y-axis.
- the imaging unit 10 may be attached to the supporting part 7 with the x r -axis and the y r -axis of the imaging region IR in a displaced state with respect to the X-axis and the Y-axis.
- FIGS. 11A , 11 B, 12 A, and 12 B are views each showing a relation between an operation of the stage unit 20 and an observed image displayed in the display part 260 in the case where the x r -axis and the y r -axis of the imaging region IR are displaced with respect to the X-axis and the Y-axis.
- the x t -axis and the y r -axis of the stage 21 respectively agree with the X-axis and the Y-axis
- the x r -axis and the y r -axis of the imaging region IR are respectively inclined just by an angle ⁇ 3 with respect to the X-axis and the Y-axis.
- the x i -axis and the y i -axis of the observed image IM are respectively inclined just by the angle ⁇ 3 with respect to the x d -axis and the y d -axis of the display part 260 .
- the stage 21 moves just by the moving amount B 4 in the moving direction A 4 .
- the observed image IM moves just by a moving amount D 4 in a direction C 4 ′ different from the direction C 4 .
- the moving direction of the stage 21 with respect to the X-axis and the Y-axis is different from the moving direction of the observed image IM with respect to the x d -axis and the y d -axis of the display part 260 .
- the observed image IM may move in a direction different from the desired direction.
- an angle of the x r -axis and the y r -axis of the imaging region IR with respect to the x t -axis and the y t -axis of the stage 21 is detected based on the moving direction of the observed image IM. Based on the detected angle, an angle of the x r -axis and the y r -axis of the imaging region with respect to the X-axis and the Y-axis (hereinafter, referred to as imaging region inclination angle) is calculated, and based on the calculated imaging region inclination angle, the moving direction of the stage 21 is corrected.
- FIGS. 13A , 13 B, 14 A, and 14 B are views for describing correction of the moving direction of the stage 21 .
- the moving direction of the stage 21 is corrected based on the moving direction of the observed image IM in the examples of FIGS. 11A , 11 B, 12 A, and 12 B.
- the observed image IM of FIG. 13A correspond to the observed image IM of FIG. 11B
- the observed image IM of FIG. 14B corresponds to the observed image IM of FIG. 12B .
- the angle of the x r -axis and the y r -axis of the imaging region IR with respect to the x t -axis and the y t -axis of the stage 21 is equivalent to the imaging region inclination angle.
- a feature point CP is detected from the observed image IM before movement. For example, a portion at which a change in contrast is relatively large, or the like, is detected as the feature point CP. Subsequently, the stage 21 is moved. In this case, the moving amount of the stage 21 is set in accordance with the magnification of the imaging unit 10 such that the feature point CP does not go outside the screen of the display part 260 . In the present example, as described above, the stage 21 is moved just by the moving amount B 4 in the moving direction A 4 . Subsequently, as shown in FIG. 13B , the feature point CP is detected from the observed image TM after the movement by pattern matching, for example.
- the moving direction C 4 ′ of the feature point CP is specified as the moving direction of the observed image IM. Based on the specified moving direction C 4 ′, the imaging region inclination angle ⁇ 3 ( FIG. 12 ) is acquired.
- the moving direction of the stage 21 is controlled such that the observed image IM moves in the desired direction.
- the x t moving amount and the y t moving amount are calculated such that the stage 21 is moved just by a moving amount B 5 in a moving direction A 5 ′ displaced just by the imaging region inclination angle ⁇ 3 from a moving direction A 5 , as shown in FIG. 14A .
- the stage 21 is moved based on the calculated x t moving amount and y t moving amount.
- the observed image IM moves in the desired direction.
- the user can easily align a desired region of the observation object S to the imaging region IR, while viewing the observed image IM.
- the CPU 220 of the image processing apparatus 200 performs an inclination angle acquiring process based on the control program stored in the storage unit 240 , thereby realizing acquirement of the imaging region inclination angle.
- FIG. 15 is a flowchart for the inclination angle acquiring process by the CPU 220 .
- the timing to perform the inclination angle acquiring process is not particularly limited.
- the inclination angle acquiring process may be performed at the startup of the image processing apparatus 200 , or the inclination angle acquiring process may be performed in accordance with an instruction by the user.
- the CPU 220 acquires a magnification of the imaging unit 10 (magnification of the object lens 13 ) from a magnification detecting part (not shown) (step S 11 ).
- the CPU 220 decides a moving amount of the stage 21 on the XY plane based on the acquired magnification of the imaging unit 10 (step 12 ).
- the CPU 220 moves the imaging unit 10 in the Z-direction by means of the lens driving part 17 such that the focus of the imaging unit 10 agrees with the surface of the observation object S (step S 13 ).
- the CPU 220 detects a feature point CP from the observed image IM before the movement (step S 14 ).
- the CPU 220 moves the stage 21 just by the moving amount decided in step S 12 (step S 15 ).
- the moving direction of the stage 21 may be previously decided, or may be set as appropriate in accordance with a position of the detected feature point.
- the CPU 220 detects the feature point CP, detected in step S 14 , from the observed image IM after the movement by pattern matching, for example (step S 16 ). Then, the CPU 220 specifies a moving direction of the feature point CP based on the position of the feature point CP, detected in steps S 14 , S 16 (step S 17 ). Next, the CPU 220 acquires an imaging region inclination angle based on the calculated moving direction (step S 18 ), and completes the process.
- the acquired imaging region inclination angle is stored into the storage unit 240 or the operation memory 270 . Thereafter, when a movement instruction of the stage 21 is provided to the input unit 250 , the CPU 220 calculates an x t moving amount and a y t moving amount by use of the stored imaging region inclination angle such that the observed image IM moves in the desired direction.
- an imaging region inclination angle is calculated based on a moving direction of the observed image IM at the time of predetermined calibration (at the time of the inclination angle acquiring process), and the moving direction of the stage 21 is corrected based on the calculated imaging region inclination angle at the normal time subsequent thereto. It is thereby possible to control movement of the stage 21 such that the observed image IM moves in the desired direction. Hence, the user can easily align a desired region of the observation object S to the imaging region IR, while viewing the observed image IM. Further, the imaging region inclination angle can be changed as appropriate in accordance with an observation purpose, user's preferences, or the like, without impairing the operability.
- a display part 160 can display a position specifying image showing a position of the imaging region IR on the stage 21 .
- FIG. 16 is a view showing an example of the position specifying image.
- a position specifying image NI of FIG. 16 includes a wide-range image BD that represents a region wider than the current imaging region IR on the stage 21 and an imaging region corresponding image CI corresponding to the imaging region IR.
- the wide-range image BD is displayed based on image data acquired by the imaging unit 10 .
- the imaging object S is previously imaged by the imaging unit 10 at a lower magnification than the current magnification, to acquire wide-field image data, and a wide-range image BD is displayed based on the image data.
- image data of the imaging region IR are sequentially acquired by the imaging unit 10 .
- a plurality of images based on the acquired plurality of pieces of image data are connected, to display the wide-range image BD.
- the wide-range image BD favorably corresponds to a relative movable range of the imaging region IR on the stage 21 .
- the user can easily recognize the relative movable range of the imaging region IR on the stage 21 . This allows the user to efficiently align the desired region of the observation object S to the imaging region IR.
- the wide-range image BD may be a diagram representing a region wider than the current imaging region IR on the stage 21 .
- the diagram is configured by a linear line or a curve, and includes a rectangle, a circle, parentheses, and an arrow, for example.
- the wide-range image BD is displayed without increasing a process load on the CPU 220 .
- the imaging region corresponding image CI has a rectangular shape.
- a position of the imaging region corresponding image CI inside the wide-range image BD corresponds to the position of the imaging region IR on the stage 21 .
- FIGS. 17 to 19 are views showing examples of display of the position specifying image NI by the display part 260 .
- the position specifying image NI is displayed so as to overlap the observed image IM.
- the user can easily recognize a positional relation between the stage 21 and the imaging region IR based on the position specifying image NI.
- the user can easily recognize a direction and a moving amount in which the stage 21 is to be moved based on the position specifying image NI.
- the position specifying image NI is rotated along with rotation of the observed image IM. This allows the user to easily recognize the rotational angle of the stage 21 with respect to the imaging region IR.
- the input unit 250 is configured so as to be able to accept designation of an arbitrary position in the position specifying image NI.
- the CPU 220 controls movement of the stage 21 by means of the stage driving part 22 such that a position on the stage 21 which corresponds to the position designated by the input unit 250 is the imaging region IR.
- the user operates the input unit 250 to designate a position PS.
- the stage 21 is moved such that a position on the stage 21 which corresponds to the position PS is the imaging region IR.
- the CPU 220 calculates an x t moving amount and a y t moving amount based on the position on the position specifying image NI designated by the input unit 250 , and provides the stage driving part 22 with the x t moving amount and the y t moving amount as a movement instruction.
- the stage driving part 22 moves the stage 21 based on the provided movement instruction. This allows the user to efficiently align the desired region of the observation object S to the imaging region IR.
- the CPU 220 may calculate a relation between a moving amount in pixel units on the position specifying image NI and an actual moving amount on the stage 21 based on the moving amount of the observed image IM in the case where the stage 21 has been moved, and may correct the x t moving amount and the y t moving amount to be provided to the stage driving part 22 based on the calculated relation.
- the common feature point CP can be detected from the observed image IM before and after movement, to thereby acquire the moving amount of the feature point CP in pixel units.
- the relation between the moving amount in pixel units in the observed image IM and the actual moving amount on the stage 21 can be calculated.
- the relation between the moving amount in pixel units in the observed image IM and the moving amount in pixel units in the position specifying image NI can be calculated based on the magnification of the imaging unit 10 and the magnification of the position specifying image NI.
- a relation between the moving amount in pixel units in the position specifying image NI and the actual moving amount on the stage 21 can be calculated. Based on the relation calculated in such a manner, the x t moving amount and the y t moving amount are corrected, to thereby improve accuracy in movement of the stage 21 .
- the stage 21 is moved by designation of the position in the position specifying image NI in the present example, this is not restrictive, and the stage 21 may be moved by designation of a desired position on the observed image IM. For example, when an arbitrary position on the observed image IM is designated, the stage 21 is moved such that a position on the stage 21 which corresponds to the designated position moves to the center of the imaging region IR.
- the CPU 220 may calculate a relation between a moving amount in pixel units in the observed image IM and an actual moving amount on the stage 21 based on the moving amount of the observed image IM in the case where the stage 21 has been moved, and may correct the x t moving amount and the y t moving amount to be provided to the stage driving part 22 based on the calculated relation.
- the rotational angle of the stage 21 is detected by the rotational angle sensor 30 in the above embodiment, this is not restrictive, and the rotational angle of the stage 21 may be detected by another method.
- a line representing at least one of the X-axis and the Y-axis may be added to the placement surface 21 a of the stage 21 , and this line may be imaged by the imaging unit 10 .
- the rotational angle of the stage 21 can be detected based on inclination of a line in the observed image IM.
- the imaging region inclination angle is calculated based on the moving direction of the observed image IM and the moving direction of the stage 21 is corrected based on the calculated imaging region inclination angle in the above embodiment (cf. FIGS. 13A to 15 ), the moving direction of the stage 21 may be corrected by another method.
- a direction specifying part for specifying a direction of the X-axis or the Y-axis may be provided on the placement surface 21 a of the stage 21 , an angle of the x t -axis and the y t -axis of the stage 21 with respect to the x r -axis and the y r -axis of the imaging region IR may be detected based on an image of the direction specifying part included in the observed image IM, and the moving direction of the stage 21 may be corrected based on the detected angle.
- the direction specifying part may, for example, be a line, a projection, a groove or the like which represents at least one of the X-axis and the Y-axis.
- the direction specifying part on the stage 21 is imaged, and the angle of the x t -axis and the y t -axis of the stage 21 with respect to the x r -axis and the y r -axis of the imaging region IR may be detected based on the direction shown by the direction specifying part in the observed image IM.
- the moving direction of the stage 21 can be corrected such that the observed image IM moves in the desired direction by means of the detected angle in a manner similar to the above example.
- the moving direction and the moving amount of the stage 21 are provided as the movement instruction in the above embodiment, this is not restrictive, and a moving direction and a moving amount of the imaging region IR being relative with respect to the stage 21 may be provided.
- the moving direction and the moving amount of the imaging region IR are converted to a moving direction and a moving amount of the stage 21 , and an x t moving amount and a y t moving amount are calculated based on the converted moving direction and moving amount in a manner similar to the above.
- stage driving part 22 rotates the stage 21 based on the rotation command signal from the CPU 220 in the above embodiment, this is not restrictive, and the stage 21 may be manually rotated.
- the imaging unit 10 is an example of the imaging part
- the display part 260 is an example of the display part
- the stage 21 is an example of the stage
- the placement surface 21 a is an example of the placement surface
- the rotational angle sensor 30 is an example of the rotational angle detecting part
- the stage driving part 22 is an example of the stage driving part
- the input unit 250 is an example of the instruction accepting part
- the CPU 220 is an example of the control part.
- the supporting part 7 is an example of the supporting part
- the wide-range image BD is an example of the first image
- the imaging region corresponding image CI is an example of the second image
- the x t -axis and the y t -axis are examples of the first and second axes
- the x r -axis and the y r -axis are examples of the third and fourth axes
- the X-axis and the Y-axis are examples of the fifth and sixth axes.
- the present invention is available for a variety of magnification observation devices.
Landscapes
- Physics & Mathematics (AREA)
- Chemical & Material Sciences (AREA)
- Analytical Chemistry (AREA)
- General Physics & Mathematics (AREA)
- Optics & Photonics (AREA)
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Microscoopes, Condenser (AREA)
- Studio Devices (AREA)
Abstract
Description
- The present application claims foreign priority based on Japanese Patent Application No. 2011-222575, filed Oct. 7, 2011, the contents of which is incorporated herein by reference.
- 1. Field of the Invention
- The present invention relates to a magnification observation device.
- 2. Description of Related Art
- Conventionally, there has been used a magnification observation device, provided with a microscope, for observing an observation object. The microscope includes an imaging unit for imaging the observation object. An image of the observation object imaged by the imaging unit is displayed by a display part. A user can acquire detailed information of the observation object based on the image displayed by the display part.
- A microscope described in Japanese Unexamined Patent Publication No. 2006-337470 has a camera for measuring an observation object and a stage on which the observation object is placed. The camera is attached to a camera attaching part so as to be along a Z-axis direction (perpendicular direction). The stage is rotatably provided below the camera. The stage is made up of an X-Y stage movable in an X-axis direction and a Y-axis direction. The stage is moved along the X-axis direction and the Y-axis direction, to thereby align the observation object with respect to an imaging region of the camera.
- In the above microscope, a rotational axis of the stage is fixed, and an X-axis and a Y-axis of the stage are rotated in association with rotation of the stage. In this case, with the X-axis and the Y-axis of the stage rotating, a moving direction of the stage is difficult to be controlled and the observation object cannot be moved in a desired direction.
- On the other hand, there can also be considered a configuration in which the X-axis and the Y-axis of the stage are fixed and the rotational axis of the stage moves in association with movement of the stage in the X-axis direction and the Y-axis direction. However, when the rotational axis of the stage goes outside the imaging region of the camera, a moving amount of the observation object with respect to the imaging region becomes large due to rotation of the stage. Thereby, a desired region of the observation object tends to go outside the imaging region.
- It is an object of the present invention to provide a magnification observation device capable of easily moving an observation object in a desired direction even when a stage is rotated, and capable of preventing a desired region of the observation object from going outside an imaging region due to rotation of the stage.
- (1) A magnification observation device according to the present invention includes: an imaging part that images an observation object, to acquire imaging data of an imaging region; a display part that displays an image of the observation object as an observed image based on the image data acquired by the imaging part; a stage that has a placement surface on which the observation object is placed, and is provided relatively rotatably with respect to the imaging part around a rotational axis substantially vertical to the placement surface, and movably along first and second axes intersecting with each other within a plane substantially parallel to the placement surface; a rotational angle detecting part that detects a rotational angle of the stage; a stage driving part that moves the stage relatively with respect to the imaging part along the first and second axes; an instruction accepting part that accepts an instruction for a moving direction of the stage; and a control part that provides the stage driving part with moving amounts of the stage along the first and second axes, to control movement of the stage, wherein the first and second axes of the stage rotate around the rotational axis integrally with the stage, and the control part is configured to control the moving amounts along the first and second axes to be provided to the stage driving part such that a moving direction of the observation object in the imaging region of the imaging part agrees with the direction accepted by the instruction accepting part, based on the rotational angle detected by the rotational angle detecting part.
- In the magnification observation device, the stage is provided relatively rotatably with respect to the imaging part around the rotational axis substantially vertical to the placement surface, and movably along the first and second axes intersecting with each other within the plane substantially parallel to the placement surface. The observation object is placed on the placement surface of the stage. The observation object is imaged by the imaging part, to acquire image data of the imaging region. Based on the image data acquired by the imaging part, the image of the observation object is displayed by the display part as an observed image.
- When the instruction for the moving direction of the stage is accepted by the instruction accepting part, the moving amounts of the stage along the first and second axes are provided from the control part to the stage driving part. Based on the provided moving amounts, the stage is moved by the stage driving part along the first and second axes, relatively with respect to the imaging part. Accordingly, the observation object moves relatively with respect to the imaging region.
- The first and second axes of the stage rotate around the rotational axis integrally with the stage. In this case, the stage can be moved without moving the rotational axis of the stage with respect to the imaging part. For this reason, by arranging the imaging region on the rotational axis of the stage, a desired region of the observation object can be prevented from going outside the imaging region due to rotation of the stage. It is therefore possible to change an orientation of the observation object while imaging the desired area of the imaging object.
- When the stage is rotated, the moving amounts along the first and second axes to be provided to the stage driving part are controlled such that the moving direction of the observation object in the imaging region of the imaging part agrees with the direction accepted by the instruction accepting part, based on the rotational angle detected by the rotational angle detecting part. Therefore, even when the stage is rotated, the observation object can be easily moved in a desired direction in the imaging region.
- (2) The imaging region of the imaging part may be provided with third and fourth axes intersecting with each other, and the control part may be configured to detect an angle difference between the first and second axes of the stage and the third and fourth axes of the imaging region during predetermined calibration, and the moving amounts of the stage along the first and second axes to be provided to the stage driving part may be corrected based on the detected angle difference in normal time.
- In this case, even when the third and fourth axes of the imaging region are displaced from a previously set state, the observation object can be moved with the third and fourth axes of the imaging region taken as references. This allows the user to easily move the observation object in the imaging region while viewing the observation object displayed in the display part, without recognizing the orientation of the imaging part.
- (3) The control part may be configured to detect the angle difference between the first and second axes of the stage and the third and fourth axes of the imaging region based on the moving direction of the observed image in a case where the stage has been moved by the stage driving part.
- In this case, the angle difference between the first and second axes of the stage and the third and fourth axes of the imaging region can be easily detected, to thereby correct the moving amounts of the stage along the first and second axes by means of the detected angle difference.
- (4) A direction specifying part for specifying directions of the first and second axes may be provided on the stage, and the control part may be configured to detect the angle difference between the first and second axes of the stage and the third and fourth axes of the imaging region based on an image of the direction specifying part included in the observed image.
- In this case, the angle difference between the first and second axes of the stage and the third and fourth axes of the imaging region can be easily detected, to thereby correct the moving amounts of the stage along the first and second axes by means of the detected angle difference.
- (5) The display part may display a position specifying image indicating a position of the imaging region on the stage.
- In this case, the user can easily recognize the position of the imaging region on the stage. This allows the user to efficiently align the desired region of the observation object to the imaging region.
- (6) The position specifying image may include a first image representing a region wider than the imaging region on the stage and a second image corresponding to the imaging region, and the control part may rotate the first image relatively with respect to the second image based on the rotational angle detected by the rotational angle detecting part.
- In this case, the user can easily recognize a positional relation between the stage and the imaging region and a rotational angle of the stage with respect to the imaging region. This allows the user to efficiently align the desired region of the observation object to the imaging region.
- (7) The imaging part may be configured to be capable of imaging the region wider than the imaging region, and the display part may display the first image based on image data obtained upon imaging a wide range by the imaging part.
- In this case, the user can easily recognize a positional relation between the observation object on the stage and the imaging region. This allows the user to efficiently align the desired region of the observation object to the imaging region.
- (8) The first image may show a diagram for specifying the region wider than the imaging region on the stage.
- In this case, the first image can be easily displayed in the display part without increasing a process load of the control part.
- (9) The first image may correspond to a relatively movable range of the imaging region on the stage.
- In this case, the user can easily recognize a relative movable range of the imaging region on the stage. This allows the user to efficiently align the desired region of the observation object to the imaging region.
- (10) The instruction accepting part may be configured to be capable of accepting designation of an arbitrary position in the position specifying image, and the control part may control movement of the stage by the stage driving part such that the imaging region moves to a position on the stage which corresponds to the position designated by the instruction accepting part.
- In this case, the user designates a desired position on the position specifying image, and thus the stage moves such that the position on the stage which corresponds to the designated position becomes the imaging region. This allows the user to efficiently align the desired region of the observation object to the imaging region.
- (11) The control part may be configured to detect a moving amount of the observed image in the case where the stage has been moved, calculate a relation between the moving amount of the observed image and the moving amount of the stage based on the detected moving amount, and correct the moving amount to be provided to the stage driving part based on the calculated relation.
- In this case, occurrence of an error in the relation between the moving amount of the observed image and the moving amount of the stage is prevented. This allows accurate control of movement of the stage based on the observed image.
- According to the present invention, it is possible to easily move an observation object in a desired direction even when the stage is rotated, and also prevent a desired region of the observation object from going outside an imaging region due to rotation of the stage.
-
FIG. 1 is a block diagram showing a configuration of a magnification observation device according to one embodiment of the present invention; -
FIG. 2 is a perspective view showing a microscope of the magnification observation device according to one embodiment of the present invention; -
FIG. 3 is a schematic view showing a state where an imaging unit of the microscope is fixed parallel to a Z-axis; -
FIG. 4 is a schematic view showing a state where the imaging unit of the microscope is inclined at a desired angle from the Z-axis; -
FIGS. 5A to 5C are schematic views showing an example of a method for calculating an xt moving amount and a yt moving amount; -
FIG. 6 is a flowchart for a stage driving process by a CPU; -
FIG. 7 is a plan view showing an example of an observation object placed on a placement surface of the stage; -
FIGS. 8A and 8B are views showing a relation between an operation of a stage unit and an observed image displayed in a display part; -
FIGS. 9A and 9B are views showing a relation between an operation of the stage unit and an observed image displayed in the display part; -
FIGS. 10A and 10B are views showing a relation between an operation of the stage unit and an observed image displayed in the display part; -
FIGS. 11A and 11B are views showing a relation between an operation of the stage unit and an observed image displayed in the display part; -
FIGS. 12A and 12B are views showing a relation between an operation of the stage unit and an observed image displayed in the display part; -
FIGS. 13A and 13B are views for describing correction of a moving direction of the stage; -
FIGS. 14A and 14B are views for describing correction of a moving direction of the stage; -
FIG. 15 is a flowchart for an inclination angle acquiring process by the - CPU;
-
FIG. 16 is a view showing an example of a position specifying image; -
FIG. 17 is a view showing an example of display of the position specifying image by the display part; -
FIG. 18 is a view showing an example of display of the position specifying image by the display part; and -
FIGS. 19A and 19B are views showing an example of display of the position specifying image by the display part. - Hereinafter, a magnification observation device according to one embodiment of the present invention will be described with reference to the drawings.
- (1) Configuration of Magnification Observation Device
-
FIG. 1 is a block diagram showing a configuration of a magnification observation device according to one embodiment of the present invention. As shown inFIG. 1 , amagnification observation device 300 is provided with amicroscope 100 and animage processing apparatus 200. - The
microscope 100 includes animaging unit 10, astage unit 20, and arotational angle sensor 30. Theimaging unit 10 includes a color CCD (charge coupled device) 11, ahalf mirror 12, anobject lens 13, an A/D converter (analog/digital converter) 15, anillumination light source 16, and alens driving part 17. Thestage unit 20 includes astage 21, astage driving part 22, and astage supporting part 23. An observation object S is placed on thestage 21. - In the
microscope 100, an X-axis and a Y-axis intersecting with each other within a horizontal plane, and a Z-axis (perpendicular direction) vertical to the X-axis and the Y-axis are defined. In the following description, a direction parallel to the X-axis is referred to as an X-axis direction, a direction parallel to the Y-axis is referred to as a Y-axis direction, and a direction parallel to the Z-axis is referred to as a Z-axis direction. - The
illumination light source 16 is, for example, a halogen lamp or a white light LED (light-emitting diode) which generates white light. White light generated by theillumination light source 16 is reflected by thehalf mirror 12, and thereafter collected by theobject lens 13 onto the observation object S on thestage 21. - The white light reflected by the observation object S is transmitted through the
object lens 13 and thehalf mirror 12, and incident on thecolor CCD 11. Thecolor CCD 11 has a plurality of pixels for red that receive red wavelength light, a plurality of pixels for green that receive green wavelength light, and a plurality of pixels for blue that receive blue wavelength light. The plurality of pixels for red, the plurality of pixels for green, and the plurality of pixels for blue are two-dimensionally arrayed. From each of the pixels in thecolor CCD 11, an electric signal corresponding to a light receiving amount is outputted. The output signal of thecolor CCD 11 is converted to a digital signal by the A/D converter 15. The digital signal outputted from the A/D converter 15 is sequentially provided as image data to theimage processing apparatus 200. Instead of thecolor CCD 11, an imaging element such as a CMOS (complementary metal oxide semiconductor) image sensor may be used. - The
object lens 13 is, for example, a zoom lens, and is configured such that its magnification is manually or automatically changeable. The magnification of theobject lens 13 is detected by a magnification detecting part (not shown). Further, theobject lens 13 is provided movably in the Z-axis direction. Thelens driving part 17 moves theobject lens 13 in the Z-axis direction by control of theimage processing apparatus 200. Thereby, a position of a focus of theimaging unit 10 is moved in the Z-axis direction. - The
stage 21 is rotatably provided on thestage supporting part 23 around a rotational axis parallel to the Z-axis. Thestage driving part 22 moves thestage 21 along a later-mentioned x-axis and y-axis relatively with respect to theimaging unit 10 based on a movement command signal (drive pulse) provided from theimage processing apparatus 200. Further, thestage driving part 22 rotates thestage 21 around the rotational axis parallel to the Z-axis relatively with respect to theimaging unit 10 based on the rotation command signal (drive pulse) provided from theimage processing apparatus 200. A position of the rotational axis of thestage 21 is constant with respect to thestage supporting part 23. Thestage driving part 22 uses a stepping motor. Therotational angle sensor 30 detects a rotational angle of thestage 21, and provides theimage processing apparatus 200 with an angle detection signal indicating the detected angle. - The
image processing apparatus 200 includes aninterface 210, a CPU (central processing unit) 220, a ROM (read only memory) 230, astorage unit 240, aninput unit 250, adisplay part 260, and anoperation memory 270. - A system program is stored into the
ROM 230. Thestorage unit 240 is made up of a hard disk and the like. In thestorage unit 240, a variety of control programs and programs are stored, and a variety of data, such as image data provided from themicroscope 100 through theinterface 210, are also stored. Theinput unit 250 includes a keyboard and a pointing device, and is configured to be able to accept a movement instruction and a rotation instruction for thestage 21 from the user or an external device. In the present embodiment, the movement instruction includes a moving direction and a moving amount, and the rotation instruction includes a rotational direction and a rotational angle. As the pointing device, a mouse, a joystick, or the like is used. - The
display part 260 is configured, for example, by a liquid crystal display panel or an organic EL (electroluminescent) panel. As described later, thedisplay part 260 is set with an xd-axis and a yd-axis intersecting with each other. The xd-axis corresponds to a lateral direction of the screen of thedisplay part 260, and the yd-axis corresponds to a vertical direction of the screen of thedisplay part 260. - The
operation memory 270 is made up of a RAM (random access memory), and used for processing a variety of data. TheCPU 220 executes the control programs stored in thestorage unit 240, to perform a variety of processes by means of theoperation memory 270, and displays an image based on the image data in thedisplay part 260. Further, theCPU 220 controls thecolor CCD 11, theillumination light source 16, thelens driving part 17, and thestage driving part 22 of themicroscope 100 through theinterface 210. -
FIG. 2 is a perspective view showing themicroscope 100 of themagnification observation device 300 according to one embodiment of the present invention. InFIG. 2 , the X-axis, the Y-axis, and the Z-axis are indicated by arrows. - As shown in
FIG. 2 , themicroscope 100 has abase 1. A first supportingbase 2 is attached onto thebase 1, and a second supportingbase 3 is also attached to the front surface of the first supportingbase 2 so as to be embedded thereinto. - A connecting
part 4 is rotatably attached to the top edge of the first supportingbase 2 around a rotational axis R1 extending in the Y-axis direction. Arotational column 5 is attached to the connectingpart 4. Thereby, therotational column 5 is inclinable within a vertical plane parallel to the Z-axis with the rotational axis R1 taken as a fulcrum point in association with rotation of the connectingpart 4. The user can fix the connectingpart 4 to the first supportingbase 2 by means of a fixingknob 9. - A circular supporting
part 7 is attached to the front surface of a connectingpart 6. A substantiallytubular imaging unit 10 is attached to the supportingpart 7. In the state ofFIG. 2 , a light axis R2 of theimaging unit 10 is parallel to the Z-axis. The supportingpart 7 has a plurality of adjustment screws 41 for moving theimaging unit 10 within a horizontal plane. It is possible to adjust a position of theimaging unit 10 such that the light axis R2 of theimaging unit 10 vertically intersects with a rotational axis R1 by means of the plurality of adjustment screws 41. - A
slider 8 is attached, slidably in the Z-axis direction, to the front surface of the second supportingbase 3 on thebase 1. Anadjustment knob 42 is provided on the side surface of the second supportingbase 3. A position of theslider 8 in the Z-axis direction (height direction) is adjustable by theadjustment knob 42. - The supporting
part 23 of thestage unit 20 is attached onto theslider 8. Thestage 21 is rotationally provided around a rotational axis R3 parallel to the Z-axis with respect to the supportingpart 23. Thestage 21 has aplacement surface 21 a on which an observation object S is placed. Further, an xt-axis and a yt-axis intersecting with each other within the horizontal plane are set on thestage 21. Thestage 21 is provided movably along the xt-axis and the yt-axis by thestage driving part 22 ofFIG. 1 . When thestage 21 rotates around the rotational axis R3, the xt-axis and the yt-axis of thestage 21 also rotate. Thereby, the xt-axis and the yt-axis of thestage 21 are inclined with respect to the X-axis and the Y-axis within a horizontal plane. - An imaging range (visual field range) of the
imaging unit 10 on theplacement surface 21 a of thestage 21 varies depending on a magnification of the imaging unit 10 (magnification of the object lens 13). Hereinafter, the imaging range of theimaging unit 10 is referred to as an imaging region, and as described later, an xr-axis and a yr-axis intersecting with each other are set in the imaging region. The xr-axis and the yr-axis of the imaging region respectively correspond to the xd-axis and the yd-axis of thedisplay part 260. - Based on image data of the imaging region, an image of the observation object S is displayed in the
display part 260 ofFIG. 1 . Hereinafter, the image of the observation object S displayed in thedisplay part 260 is referred to as an observed image. As described later, the observed image is set with an xi-axis and a yi-axis intersecting with each other. The xi-axis and the yi-axis of the observed image respectively correspond to the xt-axis and the yt-axis of thestage 21. -
FIG. 3 is a schematic view showing a state where theimaging unit 10 of themicroscope 100 is fixed parallel to the Z-axis. Further,FIG. 4 is a schematic view showing a state where theimaging unit 10 of themicroscope 100 is inclined at a desired angle with respect to the Z-axis. - As shown in
FIG. 3 , with therotational column 5 in a parallel state to the Z-axis, the fixingknob 9 is fastened, to fix the connectingpart 4 to the second supportingbase 3. Thereby, the light axis R2 of theimaging unit 10 vertically intersects with the rotational axis R1 while being in a parallel state to the Z-axis. In this case, the light axis R2 of theimaging unit 10 is vertical to the surface of thestage 21. - The fixing
knob 9 is loosened, to make the connectingpart 4 rotatable around the rotational axis R1, and therotational column 5 inclinable with the rotational axis R1 taken as a fulcrum point. Therefore, as shown inFIG. 4 , the light axis R2 of theimaging unit 10 is inclinable at an arbitrary angle θ1 with respect to the Z-axis. In this case, the light axis R2 of theimaging unit 10 vertically intersects with the rotational axis R1. Similarly, the light axis R2 of theimaging unit 10 is inclinable at an arbitrary angle on the side opposite to the side inFIG. 4 with respect to the Z-axis. - Therefore, a height of the surface of the observation object S on the
stage 21 can be made to agree with a height of the rotational axis R1, to thereby observe the same portion of the observation object in a vertical direction and an oblique direction. - (2) Operation of Stage Unit
- Upon provision of a movement instruction for the
stage 21 or a rotation instruction for thestage 21 from the user or the external device to theinput unit 250 of the image processing apparatus 200 (FIG. 1 ), theCPU 220 provides a movement command signal or a rotation command signal to thestage driving part 22. Thestage driving part 22 moves thestage 21 along the xt-axis and the yt-axis based on the movement command signal from theCPU 220, or rotates thestage 21 around the rotational axis R3 based on the rotation command signal from theCPU 220. - In the present embodiment, the movement command signal from the
CPU 220 includes a moving amount of thestage 21 along the xt-axis (hereinafter, referred to as xt moving amount), and a moving amount of thestage 21 along the yt-axis (hereinafter, referred to as yt moving amount). The rotation command signal from theCPU 220 includes a rotational direction and a rotational angle of thestage 21. - Herein, as described above, when the
stage 21 rotates around the rotational axis R3, the xt-axis and the yt-axis of thestage 21 are inclined with respect to the X-axis and the Y-axis. For this reason, in order to move thestage 21 in a moving direction provided as a movement instruction, it is necessary to control of the xt moving amount and the yt moving amount in accordance with the rotational angle of thestage 21. In the present embodiment, based on the rotational angle of thestage 21 detected by therotational angle sensor 30, the xt moving amount and the yt moving amount are calculated such that the moving direction provided as the movement instruction agrees with an actual moving direction of thestage 21. -
FIGS. 5A to 5C are schematic views showing an example of a method for calculating an xt moving amount and a yt moving amount. In the example ofFIGS. 5A to 5C , the xt-axis and the yt-axis are inclined at an angle θ0 with respect to the X-axis and the Y-axis. As shown inFIG. 5A , theinput unit 250 is provided with a moving direction A0 and a moving amount B0 as a movement instruction. In this case, as shown inFIG. 5B , a moving amount dx1 along the X-axis and a moving amount dy1 along the Y-axis are calculated based on the moving direction A0 and the moving amount B0. Further, as shown inFIG. 5C , an xt moving amount dx2 and a yt moving amount dy2 are calculated based on the calculated moving amounts dx1, dy1 and the rotational angle θ0 of thestage 21 detected by therotational angle sensor 30. -
FIG. 6 is a flowchart for a stage driving process by theCPU 220. TheCPU 220 performs the stage driving process ofFIG. 6 based on the control program stored in thestorage unit 240. - As shown in
FIG. 6 , first, theCPU 220 determines whether or not a movement instruction of thestage 21 has been provided to the input unit 250 (step S1). When the movement instruction of thestage 21 has been provided, theCPU 220 acquires a rotational angle of thestage 21 based on an angle detection signal from the rotational angle sensor 30 (step S2). Next, theCPU 220 calculates an xt moving amount and a yt moving amount as described above based on the provided movement instruction and the acquired rotational angle (step S3). Subsequently, theCPU 220 provides the calculated xt moving amount and yt moving amount as a movement command signal to the stage driving part 22 (step S4). Thereafter, theCPU 220 returns to the process of step S1. - In step S1, when the movement instruction of the
stage 21 has not been provided, theCPU 220 determines whether or not a rotation instruction of thestage 21 has been provided to the input unit 250 (step S5). When the rotation instruction of thestage 21 has not been provided, theCPU 220 returns to the process of step S1. When the rotation instruction of thestage 21 has been provided, theCPU 220 provides a rotational direction and a rotational angle as a rotation command signal to thestage driving part 22 based on the provided rotation instruction (step S6). Thereafter, theCPU 220 returns to the process of step S1. - As described above, the xt moving amount and the yt moving amount are calculated based on the movement instruction provided to the
input unit 250 and the rotational angle detected by therotational angle sensor 30, and the calculated xt moving amount and the yt moving amount are provided to thestage driving part 22. Based on the xt moving amount and the yt moving amount, thestage 21 is moved from thestage driving part 22 along the xt-axis and the yt-axis. Therefore, even when thestage 21 is rotated, the observation object S can be moved in a desired direction in the imaging region. - It is to be noted that the rotational angle of the
stage 21, detected by therotational angle sensor 30, may be displayed by thedisplay part 260. In this case, the user can easily recognize the rotational angle of thestage 21. - Although the moving amount and the moving direction are provided as the movement instruction in the present example, this is not restrictive. For example, a coordinate (hereinafter, referred to as target coordinate) showing a position to which a movement is made may be provided as a movement instruction. For example, there are set an XY coordinate system with the X-axis and the Y-axis taken as coordinate axes, and an xtyt coordinate system with the xt-axis and the yt-axis taken as coordinate axes. In this case, a target coordinate in the XY coordinate system is provided as a movement instruction. The provided target coordinate is converted to a coordinate in the xtyt coordinate system based on the rotational angle of the
stage 21. Based on the converted target coordinate, an xt moving amount and a yt moving amount can be calculated. - Further, even when a moving amount and a moving direction are provided as a movement instruction, an xt moving amount and a yt moving amount may be calculated using the XY coordinate system and the xtyt coordinate system in a similar manner to the above. For example, a target coordinate in the XY coordinate system is calculated based on the provided moving amount and the moving direction. The calculated target coordinate is converted to a coordinate in the xtyt coordinate system based on a rotational angle of the
stage 21. Based on the converted target coordinate, an xt moving amount and a yt moving amount can be calculated. - (3) Relation Between Operation of Stage Unit and Observed Image
- Hereinafter, a relation between an operation of the
stage unit 20 and an observed image displayed in thedisplay part 260 will be described.FIG. 7 is a plan view showing an example of the observation object S placed on theplacement surface 21 a of thestage 21.FIGS. 8A to 10B are views each showing a relation between an operation of thestage unit 20 and an observed image displayed in thedisplay part 260.FIGS. 8A , 9A, and 10A each show a positional relation between thestage 21 and the imaging region, andFIGS. 8B , 9B, and 10B each show an observed image displayed in thedisplay part 260. Further,FIGS. 8A , 9A, and 10A each show the X-axis and the Y-axis, the xt-axis and the yt-axis of thestage 21, and the xr-axis and the yr-axis of the imaging region.FIGS. 8B , 9B, and 10B each show the xd-axis and the yd-axis of thedisplay part 260 and the xi-axis and the yi-axis of the observed image. - In the following description, a direction An (n is a natural number) is a direction of movement or rotation of the
stage 21, and a direction Cn is a direction of movement or rotation of an observed image IM. The direction An which is relative with respect to the X-axis and the Y-axis agrees with the direction Cn which is relative with respect to the xd-axis and the yd-axis of thedisplay part 260. Further, a moving amount Bn is a moving amount of thestage 21, and a moving amount Dn is a moving amount of the observed image. In the present example, a ratio between the moving amount Bn and the moving amount Dn is constant. - In the state of
FIG. 8A , the xt-axis and the yt-axis of thestage 21 respectively agree with the X-axis and the Y-axis. Further, the xr-axis and the yr-axis of an imaging region IR respectively agree with the X-axis and the Y-axis. In this case, as shown inFIG. 8B , the xi-axis and the yi-axis of the observed image IM respectively agree with the xd-axis and the yd-axis of thedisplay part 260. - As shown in
FIG. 9A , when a rotational direction A2 and a rotational angle θ2 are provided as a rotation instruction to theinput unit 250, thestage 21 is rotated just by the angle θ2 in the direction A2 around the rotational axis R3. In this case, as shown inFIG. 9B , the observed image IM rotates just by the angle θ2 in a rotational direction C2. - It is to be noted that, in the present example, the position of the
imaging unit 10 is adjusted such that the rotational axis R3 is located at the central part of the imaging region IR. Hence, even when thestage 21 is rotated around the rotational axis R3, a desired region of the observation object S is prevented from going outside the imaging region IR. It is thus possible to rotate the observation object S while imaging the desired region of the observation object S by theimaging unit 10. - Especially, as in the example of
FIG. 4 , when thestage 21 is rotated with theimaging unit 10 in an inclined state with respect to the Z-axis, the desired region of the observation object S is imaged from a variety of angles. This allows the user to stereoscopically observe the desired region of the observation object S. - Further, although the position of the rotational axis R3 of the
stage 21 is fixed with respect to thestage supporting part 23 in the present example, this is not restrictive. The position of the rotational axis R3 of thestage 21 may be changeable with respect to thestage supporting part 23. In this case, the position of the rotational axis R3 of thestage 21 can be adjusted such that the rotational axis R3 is located at the central part of the imaging region IR. - As shown in
FIG. 10A , when a moving direction A3 and a moving amount B3 are provided as a movement instruction to theinput unit 250, thestage 21 moves just by the moving amount B3 in the moving direction A3. In this case, as described above, an xt moving amount and a yt moving amount are calculated based on the moving direction A3, the moving amount B3, and the angle θ2, and thestage 21 is moved based on the xt moving amount and the yt moving amount. As shown inFIG. 10B , the observed image IM moves just by a moving amount D3 in a direction C3. - (4) Displacement of Imaging Unit
- In the examples of
FIGS. 8 and 9 , since the xr-axis and the yr-axis of the imaging region IR respectively agree with the X-axis and the Y-axis, the moving direction of thestage 21 with respect to the X-axis and the Y-axis agrees with the moving direction of the observed image IM with respect to the xd-axis and the yd-axis of thedisplay part 260. In practice, there are cases where the xr-axis and the yr-axis of the imaging region IR are displaced with respect to the X-axis and the Y-axis due to an assembly error or the like. Further, when theimaging unit 10 is provided detachably from the supportingpart 7, theimaging unit 10 may be attached to the supportingpart 7 with the xr-axis and the yr-axis of the imaging region IR in a displaced state with respect to the X-axis and the Y-axis. When the xr-axis and the yr-axis of the imaging region IR are displaced with respect to the X-axis and the Y-axis, the following problem occurs. -
FIGS. 11A , 11B, 12A, and 12B are views each showing a relation between an operation of thestage unit 20 and an observed image displayed in thedisplay part 260 in the case where the xr-axis and the yr-axis of the imaging region IR are displaced with respect to the X-axis and the Y-axis. - In the state of
FIG. 11A , the xt-axis and the yr-axis of thestage 21 respectively agree with the X-axis and the Y-axis, and the xr-axis and the yr-axis of the imaging region IR are respectively inclined just by an angle θ3 with respect to the X-axis and the Y-axis. In this case, as shown inFIG. 11B , the xi-axis and the yi-axis of the observed image IM are respectively inclined just by the angle θ3 with respect to the xd-axis and the yd-axis of thedisplay part 260. - As shown in
FIG. 12A , when a moving direction A4 and a moving amount B4 are provided as a movement instruction to theinput unit 250, thestage 21 moves just by the moving amount B4 in the moving direction A4. In this case, as shown inFIG. 12B , the observed image IM moves just by a moving amount D4 in a direction C4′ different from the direction C4. - As described above, when the xr-axis and the yr-axis of the imaging region IR are displaced with respect to the X-axis and the Y-axis, the moving direction of the
stage 21 with respect to the X-axis and the Y-axis is different from the moving direction of the observed image IM with respect to the xd-axis and the yd-axis of thedisplay part 260. For this reason, the observed image IM may move in a direction different from the desired direction. Thereat, in the present embodiment, an angle of the xr-axis and the yr-axis of the imaging region IR with respect to the xt-axis and the yt-axis of thestage 21 is detected based on the moving direction of the observed image IM. Based on the detected angle, an angle of the xr-axis and the yr-axis of the imaging region with respect to the X-axis and the Y-axis (hereinafter, referred to as imaging region inclination angle) is calculated, and based on the calculated imaging region inclination angle, the moving direction of thestage 21 is corrected. -
FIGS. 13A , 13B, 14A, and 14B are views for describing correction of the moving direction of thestage 21. Herein, examples will be described in which the moving direction of thestage 21 is corrected based on the moving direction of the observed image IM in the examples ofFIGS. 11A , 11B, 12A, and 12B. The observed image IM ofFIG. 13A correspond to the observed image IM ofFIG. 11B , and the observed image IM ofFIG. 14B corresponds to the observed image IM ofFIG. 12B . Further, in the present example, with the xt-axis and the yt-axis of thestage 21 respectively agreeing with the X-axis and the Y-axis, the angle of the xr-axis and the yr-axis of the imaging region IR with respect to the xt-axis and the yt-axis of thestage 21 is equivalent to the imaging region inclination angle. - First, as shown in
FIG. 13A , a feature point CP is detected from the observed image IM before movement. For example, a portion at which a change in contrast is relatively large, or the like, is detected as the feature point CP. Subsequently, thestage 21 is moved. In this case, the moving amount of thestage 21 is set in accordance with the magnification of theimaging unit 10 such that the feature point CP does not go outside the screen of thedisplay part 260. In the present example, as described above, thestage 21 is moved just by the moving amount B4 in the moving direction A4. Subsequently, as shown inFIG. 13B , the feature point CP is detected from the observed image TM after the movement by pattern matching, for example. - Based on the position of the feature point CP in the observed image IM before the movement and the position of the feature point CP in the observed image IM after the movement, the moving direction C4′ of the feature point CP is specified as the moving direction of the observed image IM. Based on the specified moving direction C4′, the imaging region inclination angle θ3 (
FIG. 12 ) is acquired. - Based on the acquired imaging region inclination angle θ3, the moving direction of the
stage 21 is controlled such that the observed image IM moves in the desired direction. Specifically, in the case of moving the observed image IM just by a moving amount D5 in a direction C5 as shown inFIG. 14B , the xt moving amount and the yt moving amount are calculated such that thestage 21 is moved just by a moving amount B5 in a moving direction A5′ displaced just by the imaging region inclination angle θ3 from a moving direction A5, as shown inFIG. 14A . Thestage 21 is moved based on the calculated xt moving amount and yt moving amount. Thereby, the observed image IM moves in the desired direction. Hence, the user can easily align a desired region of the observation object S to the imaging region IR, while viewing the observed image IM. - The
CPU 220 of theimage processing apparatus 200 performs an inclination angle acquiring process based on the control program stored in thestorage unit 240, thereby realizing acquirement of the imaging region inclination angle.FIG. 15 is a flowchart for the inclination angle acquiring process by theCPU 220. The timing to perform the inclination angle acquiring process is not particularly limited. For example, the inclination angle acquiring process may be performed at the startup of theimage processing apparatus 200, or the inclination angle acquiring process may be performed in accordance with an instruction by the user. - As shown in
FIG. 15 , first, theCPU 220 acquires a magnification of the imaging unit 10 (magnification of the object lens 13) from a magnification detecting part (not shown) (step S11). Next, theCPU 220 decides a moving amount of thestage 21 on the XY plane based on the acquired magnification of the imaging unit 10 (step 12). Subsequently, theCPU 220 moves theimaging unit 10 in the Z-direction by means of thelens driving part 17 such that the focus of theimaging unit 10 agrees with the surface of the observation object S (step S13). - Then, as shown in
FIG. 13A , theCPU 220 detects a feature point CP from the observed image IM before the movement (step S14). Next, theCPU 220 moves thestage 21 just by the moving amount decided in step S12 (step S15). The moving direction of thestage 21 may be previously decided, or may be set as appropriate in accordance with a position of the detected feature point. - Subsequently, the
CPU 220 detects the feature point CP, detected in step S14, from the observed image IM after the movement by pattern matching, for example (step S16). Then, theCPU 220 specifies a moving direction of the feature point CP based on the position of the feature point CP, detected in steps S14, S16 (step S17). Next, theCPU 220 acquires an imaging region inclination angle based on the calculated moving direction (step S18), and completes the process. - The acquired imaging region inclination angle is stored into the
storage unit 240 or theoperation memory 270. Thereafter, when a movement instruction of thestage 21 is provided to theinput unit 250, theCPU 220 calculates an xt moving amount and a yt moving amount by use of the stored imaging region inclination angle such that the observed image IM moves in the desired direction. - As described above, in the present embodiment, an imaging region inclination angle is calculated based on a moving direction of the observed image IM at the time of predetermined calibration (at the time of the inclination angle acquiring process), and the moving direction of the
stage 21 is corrected based on the calculated imaging region inclination angle at the normal time subsequent thereto. It is thereby possible to control movement of thestage 21 such that the observed image IM moves in the desired direction. Hence, the user can easily align a desired region of the observation object S to the imaging region IR, while viewing the observed image IM. Further, the imaging region inclination angle can be changed as appropriate in accordance with an observation purpose, user's preferences, or the like, without impairing the operability. - (5) Position Specifying Image
- In the present embodiment, a display part 160 can display a position specifying image showing a position of the imaging region IR on the
stage 21.FIG. 16 is a view showing an example of the position specifying image. A position specifying image NI ofFIG. 16 includes a wide-range image BD that represents a region wider than the current imaging region IR on thestage 21 and an imaging region corresponding image CI corresponding to the imaging region IR. - In the present example, the wide-range image BD is displayed based on image data acquired by the
imaging unit 10. For example, the imaging object S is previously imaged by theimaging unit 10 at a lower magnification than the current magnification, to acquire wide-field image data, and a wide-range image BD is displayed based on the image data. Further, as thestage 21 is sequentially moved along the xt-axis and the yt-axis, image data of the imaging region IR are sequentially acquired by theimaging unit 10. A plurality of images based on the acquired plurality of pieces of image data are connected, to display the wide-range image BD. - Further, the wide-range image BD favorably corresponds to a relative movable range of the imaging region IR on the
stage 21. In this case, the user can easily recognize the relative movable range of the imaging region IR on thestage 21. This allows the user to efficiently align the desired region of the observation object S to the imaging region IR. - Further, the wide-range image BD may be a diagram representing a region wider than the current imaging region IR on the
stage 21. The diagram is configured by a linear line or a curve, and includes a rectangle, a circle, parentheses, and an arrow, for example. In this case, the wide-range image BD is displayed without increasing a process load on theCPU 220. - In the present example, the imaging region corresponding image CI has a rectangular shape. A position of the imaging region corresponding image CI inside the wide-range image BD corresponds to the position of the imaging region IR on the
stage 21. -
FIGS. 17 to 19 are views showing examples of display of the position specifying image NI by thedisplay part 260. In the example ofFIG. 17 , the position specifying image NI is displayed so as to overlap the observed image IM. In this case, the user can easily recognize a positional relation between thestage 21 and the imaging region IR based on the position specifying image NI. Further, when a portion to be observed of the observation object S is not displayed as the observed image IM, the user can easily recognize a direction and a moving amount in which thestage 21 is to be moved based on the position specifying image NI. - When the
stage 21 is rotated in the state ofFIG. 17 , as shown inFIG. 18 , the position specifying image NI is rotated along with rotation of the observed image IM. This allows the user to easily recognize the rotational angle of thestage 21 with respect to the imaging region IR. - In the present embodiment, the
input unit 250 is configured so as to be able to accept designation of an arbitrary position in the position specifying image NI. TheCPU 220 controls movement of thestage 21 by means of thestage driving part 22 such that a position on thestage 21 which corresponds to the position designated by theinput unit 250 is the imaging region IR. - For example, in the position specifying image NI of
FIG. 19A , the user operates theinput unit 250 to designate a position PS. In this case, as shown inFIG. 19B , thestage 21 is moved such that a position on thestage 21 which corresponds to the position PS is the imaging region IR. - Specifically, the
CPU 220 calculates an xt moving amount and a yt moving amount based on the position on the position specifying image NI designated by theinput unit 250, and provides thestage driving part 22 with the xt moving amount and the yt moving amount as a movement instruction. Thestage driving part 22 moves thestage 21 based on the provided movement instruction. This allows the user to efficiently align the desired region of the observation object S to the imaging region IR. - The
CPU 220 may calculate a relation between a moving amount in pixel units on the position specifying image NI and an actual moving amount on thestage 21 based on the moving amount of the observed image IM in the case where thestage 21 has been moved, and may correct the xt moving amount and the yt moving amount to be provided to thestage driving part 22 based on the calculated relation. - For example, as shown in
FIGS. 13A and 13B , the common feature point CP can be detected from the observed image IM before and after movement, to thereby acquire the moving amount of the feature point CP in pixel units. Hence, the relation between the moving amount in pixel units in the observed image IM and the actual moving amount on thestage 21 can be calculated. Further, the relation between the moving amount in pixel units in the observed image IM and the moving amount in pixel units in the position specifying image NI can be calculated based on the magnification of theimaging unit 10 and the magnification of the position specifying image NI. As a result, a relation between the moving amount in pixel units in the position specifying image NI and the actual moving amount on thestage 21 can be calculated. Based on the relation calculated in such a manner, the xt moving amount and the yt moving amount are corrected, to thereby improve accuracy in movement of thestage 21. - Although the
stage 21 is moved by designation of the position in the position specifying image NI in the present example, this is not restrictive, and thestage 21 may be moved by designation of a desired position on the observed image IM. For example, when an arbitrary position on the observed image IM is designated, thestage 21 is moved such that a position on thestage 21 which corresponds to the designated position moves to the center of the imaging region IR. - Also in this case, similarly to the above, the
CPU 220 may calculate a relation between a moving amount in pixel units in the observed image IM and an actual moving amount on thestage 21 based on the moving amount of the observed image IM in the case where thestage 21 has been moved, and may correct the xt moving amount and the yt moving amount to be provided to thestage driving part 22 based on the calculated relation. - (6) Other Embodiments
- Although the rotational angle of the
stage 21 is detected by therotational angle sensor 30 in the above embodiment, this is not restrictive, and the rotational angle of thestage 21 may be detected by another method. For example, a line representing at least one of the X-axis and the Y-axis may be added to theplacement surface 21 a of thestage 21, and this line may be imaged by theimaging unit 10. In this case, the rotational angle of thestage 21 can be detected based on inclination of a line in the observed image IM. - Further, although the imaging region inclination angle is calculated based on the moving direction of the observed image IM and the moving direction of the
stage 21 is corrected based on the calculated imaging region inclination angle in the above embodiment (cf.FIGS. 13A to 15 ), the moving direction of thestage 21 may be corrected by another method. For example, a direction specifying part for specifying a direction of the X-axis or the Y-axis may be provided on theplacement surface 21 a of thestage 21, an angle of the xt-axis and the yt-axis of thestage 21 with respect to the xr-axis and the yr-axis of the imaging region IR may be detected based on an image of the direction specifying part included in the observed image IM, and the moving direction of thestage 21 may be corrected based on the detected angle. The direction specifying part may, for example, be a line, a projection, a groove or the like which represents at least one of the X-axis and the Y-axis. - Specifically, at the time of predetermined calibration (e.g. at the startup of the image processing apparatus 200), the direction specifying part on the
stage 21 is imaged, and the angle of the xt-axis and the yt-axis of thestage 21 with respect to the xr-axis and the yr-axis of the imaging region IR may be detected based on the direction shown by the direction specifying part in the observed image IM. This leads to detection of a relative angle between the imaging region IR and thestage 21 in an initial state. At the normal time subsequent thereto, the moving direction of thestage 21 can be corrected such that the observed image IM moves in the desired direction by means of the detected angle in a manner similar to the above example. - Further, although the moving direction and the moving amount of the
stage 21 are provided as the movement instruction in the above embodiment, this is not restrictive, and a moving direction and a moving amount of the imaging region IR being relative with respect to thestage 21 may be provided. In this case, for example, the moving direction and the moving amount of the imaging region IR are converted to a moving direction and a moving amount of thestage 21, and an xt moving amount and a yt moving amount are calculated based on the converted moving direction and moving amount in a manner similar to the above. - Further, although the
stage driving part 22 rotates thestage 21 based on the rotation command signal from theCPU 220 in the above embodiment, this is not restrictive, and thestage 21 may be manually rotated. - (7) Corresponding Relation Between Each Constitutional Element of Claims and Each Part of Embodiment
- Although an example of correspondence between each constitutional element of the claims and each part of the embodiment will be hereinafter described, the present invention is not limited to the following example.
- In the above embodiment, the
imaging unit 10 is an example of the imaging part, thedisplay part 260 is an example of the display part, thestage 21 is an example of the stage, theplacement surface 21 a is an example of the placement surface, therotational angle sensor 30 is an example of the rotational angle detecting part, thestage driving part 22 is an example of the stage driving part, theinput unit 250 is an example of the instruction accepting part, and theCPU 220 is an example of the control part. - Further, the supporting
part 7 is an example of the supporting part, the wide-range image BD is an example of the first image, the imaging region corresponding image CI is an example of the second image, the xt-axis and the yt-axis are examples of the first and second axes, the xr-axis and the yr-axis are examples of the third and fourth axes, and the X-axis and the Y-axis are examples of the fifth and sixth axes. - As each constitutional element of the claims, a variety of other elements having the configuration or the function described in the claims can be employed.
- The present invention is available for a variety of magnification observation devices.
Claims (11)
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2011-222575 | 2011-10-07 | ||
JP2011222575A JP5841398B2 (en) | 2011-10-07 | 2011-10-07 | Magnifying observation device |
Publications (2)
Publication Number | Publication Date |
---|---|
US20130088586A1 true US20130088586A1 (en) | 2013-04-11 |
US8994810B2 US8994810B2 (en) | 2015-03-31 |
Family
ID=47909055
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/559,799 Active 2033-07-10 US8994810B2 (en) | 2011-10-07 | 2012-07-27 | Magnification observation device |
Country Status (4)
Country | Link |
---|---|
US (1) | US8994810B2 (en) |
JP (1) | JP5841398B2 (en) |
CN (1) | CN103105671B (en) |
DE (1) | DE102012217281B4 (en) |
Cited By (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2015127778A (en) * | 2013-12-27 | 2015-07-09 | 株式会社キーエンス | Magnifying observation device and observation method using the same |
US20150287177A1 (en) * | 2014-04-08 | 2015-10-08 | Mitutoyo Corporation | Image measuring device |
EP3006980A1 (en) * | 2014-10-06 | 2016-04-13 | Leica Microsystems (Schweiz) AG | Digital microscope with a radial piston braking system |
US9690089B2 (en) | 2013-12-27 | 2017-06-27 | Keyence Corporation | Magnifying observation apparatus, magnified image observing method and computer-readable recording medium |
US9690088B2 (en) | 2013-12-27 | 2017-06-27 | Keyence Corporation | Magnifying observation apparatus, magnified image observing method and computer-readable recording medium |
US9726873B2 (en) | 2013-12-27 | 2017-08-08 | Keyence Corporation | Microscope and magnifying observation method using the same |
US10048477B1 (en) | 2017-02-10 | 2018-08-14 | Nanotronics Imaging, Inc. | Camera and specimen alignment to facilitate large area imaging in microscopy |
US20200073106A1 (en) * | 2018-08-30 | 2020-03-05 | Keyence Corporation | Magnifying observation apparatus |
US10890749B2 (en) * | 2018-08-30 | 2021-01-12 | Keyence Corporation | Magnifying observation apparatus |
EP3526637B1 (en) * | 2016-10-13 | 2023-04-05 | Life Technologies Holdings Pte Limited | Systems, methods, and apparatuses for optimizing field of view |
EP4343404A1 (en) * | 2022-09-20 | 2024-03-27 | Leica Microsystems CMS GmbH | Controller for an imaging system, system and corresponding method |
Families Citing this family (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
DE102013222295A1 (en) * | 2013-11-04 | 2015-05-07 | Carl Zeiss Microscopy Gmbh | Digital microscope, method for calibration and method for automatic focus and image center tracking for such a digital microscope |
CN204241755U (en) * | 2014-05-13 | 2015-04-01 | 徕卡显微系统(上海)有限公司 | For the regulating device of microscopical illuminace component, lighting device and microscope |
CN104570315B (en) | 2014-12-30 | 2017-06-27 | 中国科学院西安光学精密机械研究所 | Color three-dimensional tomography microscopic imaging system and method based on structural illumination |
JP6534318B2 (en) * | 2015-09-02 | 2019-06-26 | アズビル株式会社 | Measuring method of fluorescent particle |
JP6800698B2 (en) * | 2016-10-21 | 2020-12-16 | 株式会社キーエンス | How to control the magnifying observation device and the magnifying observation device |
JP7061518B2 (en) * | 2018-06-22 | 2022-04-28 | オリンパス株式会社 | Magnifying observation device |
JP6704156B1 (en) * | 2019-09-03 | 2020-06-03 | 日本分光株式会社 | Polarimetry device with automatic alignment function |
Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5550635A (en) * | 1993-10-29 | 1996-08-27 | Canon Kabushiki Kaisha | Rotational deviation detecting method and system using a periodic pattern |
US5734930A (en) * | 1992-09-04 | 1998-03-31 | Canon Kabushiki Kaisha | Optical apparatus having a visual axis detector |
US20060124850A1 (en) * | 2004-12-10 | 2006-06-15 | Hitachi High Technologies Corporation | Scanning interference electron microscope |
US20110169936A1 (en) * | 2010-01-14 | 2011-07-14 | Olympus Corporation | Microscope |
US20110216183A1 (en) * | 2010-03-03 | 2011-09-08 | Olympus Corporation | Microscope apparatus and observation position reproduction method |
US20120001070A1 (en) * | 2010-07-02 | 2012-01-05 | Keyence Corporation | Magnifying Observation Apparatus |
US20120001069A1 (en) * | 2010-07-02 | 2012-01-05 | Keyence Corporation | Magnifying Observation Apparatus |
US20130050464A1 (en) * | 2011-08-31 | 2013-02-28 | Keyence Corporation | Magnification Observation Device, Magnification Observation Method, And Magnification Observation Program |
US20130100170A1 (en) * | 2011-10-19 | 2013-04-25 | Keyence Corporation | Image Processing Apparatus And Image Processing Program |
Family Cites Families (17)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
DE3718066A1 (en) * | 1987-05-29 | 1988-12-08 | Zeiss Carl Fa | METHOD FOR MICROINJECTION IN CELLS OR. FOR SUCTION FROM SINGLE CELLS OR WHOLE CELLS FROM CELL CULTURES |
JPH08194162A (en) | 1995-01-17 | 1996-07-30 | Nikon Corp | Driving device for stage |
JP4473992B2 (en) | 1999-06-15 | 2010-06-02 | 株式会社キーエンス | Magnification observation stand device and magnification observation method |
JP2004153462A (en) | 2002-10-29 | 2004-05-27 | Keyence Corp | Magnifying observation apparatus, operating method of the magnifying observation apparatus, magnifying observation apparatus operating program, and computer-readable recording medium |
JP4831998B2 (en) | 2005-05-31 | 2011-12-07 | 株式会社キーエンス | Magnifying observation stand device and microscope equipped with the same |
JP4436278B2 (en) | 2005-05-31 | 2010-03-24 | 株式会社キーエンス | Magnifying observation stand device and microscope equipped with the same |
JP2007024927A (en) * | 2005-07-12 | 2007-02-01 | Olympus Corp | Microscopic image photographing apparatus |
JP2008175893A (en) | 2007-01-16 | 2008-07-31 | Keyence Corp | Microscope camera support stand |
JP4860551B2 (en) | 2007-06-01 | 2012-01-25 | 株式会社キーエンス | Magnification observation apparatus, high gradation image file creation method, high gradation image file creation method, high gradation image file creation program, and computer-readable recording medium |
JP5188101B2 (en) | 2007-06-01 | 2013-04-24 | 株式会社キーエンス | Magnification observation apparatus, magnified image photographing method, magnified image photographing program, and computer-readable recording medium |
US8554344B2 (en) * | 2008-03-13 | 2013-10-08 | Yu Sun | Method and apparatus for microscopy |
JP5154392B2 (en) | 2008-12-12 | 2013-02-27 | 株式会社キーエンス | Imaging device |
JP5096302B2 (en) | 2008-12-12 | 2012-12-12 | 株式会社キーエンス | Imaging device |
JP5096303B2 (en) | 2008-12-12 | 2012-12-12 | 株式会社キーエンス | Imaging device |
JP2010266750A (en) * | 2009-05-15 | 2010-11-25 | Olympus Corp | Observation device and observation system |
JP2011048013A (en) * | 2009-08-25 | 2011-03-10 | Nikon Corp | Control device and microscope system using the control device |
JP2011145470A (en) * | 2010-01-14 | 2011-07-28 | Olympus Corp | Microscope |
-
2011
- 2011-10-07 JP JP2011222575A patent/JP5841398B2/en active Active
-
2012
- 2012-07-27 US US13/559,799 patent/US8994810B2/en active Active
- 2012-09-25 DE DE102012217281.5A patent/DE102012217281B4/en active Active
- 2012-10-08 CN CN201210377716.4A patent/CN103105671B/en active Active
Patent Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5734930A (en) * | 1992-09-04 | 1998-03-31 | Canon Kabushiki Kaisha | Optical apparatus having a visual axis detector |
US5550635A (en) * | 1993-10-29 | 1996-08-27 | Canon Kabushiki Kaisha | Rotational deviation detecting method and system using a periodic pattern |
US20060124850A1 (en) * | 2004-12-10 | 2006-06-15 | Hitachi High Technologies Corporation | Scanning interference electron microscope |
US20110169936A1 (en) * | 2010-01-14 | 2011-07-14 | Olympus Corporation | Microscope |
US20110216183A1 (en) * | 2010-03-03 | 2011-09-08 | Olympus Corporation | Microscope apparatus and observation position reproduction method |
US20120001070A1 (en) * | 2010-07-02 | 2012-01-05 | Keyence Corporation | Magnifying Observation Apparatus |
US20120001069A1 (en) * | 2010-07-02 | 2012-01-05 | Keyence Corporation | Magnifying Observation Apparatus |
US20130050464A1 (en) * | 2011-08-31 | 2013-02-28 | Keyence Corporation | Magnification Observation Device, Magnification Observation Method, And Magnification Observation Program |
US20130100170A1 (en) * | 2011-10-19 | 2013-04-25 | Keyence Corporation | Image Processing Apparatus And Image Processing Program |
Cited By (22)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2015127778A (en) * | 2013-12-27 | 2015-07-09 | 株式会社キーエンス | Magnifying observation device and observation method using the same |
US9690089B2 (en) | 2013-12-27 | 2017-06-27 | Keyence Corporation | Magnifying observation apparatus, magnified image observing method and computer-readable recording medium |
US9690088B2 (en) | 2013-12-27 | 2017-06-27 | Keyence Corporation | Magnifying observation apparatus, magnified image observing method and computer-readable recording medium |
US9726873B2 (en) | 2013-12-27 | 2017-08-08 | Keyence Corporation | Microscope and magnifying observation method using the same |
US20150287177A1 (en) * | 2014-04-08 | 2015-10-08 | Mitutoyo Corporation | Image measuring device |
EP3006980A1 (en) * | 2014-10-06 | 2016-04-13 | Leica Microsystems (Schweiz) AG | Digital microscope with a radial piston braking system |
WO2016055339A1 (en) * | 2014-10-06 | 2016-04-14 | Leica Microsystems (Schweiz) Ag | Eucentric digital microscope having a pivotally mounted pivot unit |
US10185136B2 (en) | 2014-10-06 | 2019-01-22 | Leica Microsystems (Schweiz) Ag | Eucentric digital microscope having a pivotally mounted pivot unit |
EP3526637B1 (en) * | 2016-10-13 | 2023-04-05 | Life Technologies Holdings Pte Limited | Systems, methods, and apparatuses for optimizing field of view |
US20180335614A1 (en) * | 2017-02-10 | 2018-11-22 | Nanotronics Imaging, Inc. | Camera and specimen alignment to facilitate large area imaging in microscopy |
US20180307016A1 (en) * | 2017-02-10 | 2018-10-25 | Nanotronics Imaging, Inc. | Camera and specimen alignment to facilitate large area imaging in microscopy |
WO2018147888A1 (en) * | 2017-02-10 | 2018-08-16 | Nanotronics Imaging, Inc. | Camera and specimen alignment to facilitate large area imaging in microscopy |
US10416426B2 (en) * | 2017-02-10 | 2019-09-17 | Nanotronics Imaging, Inc. | Camera and specimen alignment to facilitate large area imaging in microscopy |
US10698191B2 (en) * | 2017-02-10 | 2020-06-30 | Nanotronics Imaging, Inc. | Camera and specimen alignment to facilitate large area imaging in microscopy |
TWI734891B (en) * | 2017-02-10 | 2021-08-01 | 美商奈米創尼克影像公司 | Method for imaging a specimen along a desired x'-direction of the specimen and microscope system |
US11099368B2 (en) * | 2017-02-10 | 2021-08-24 | Nanotronics Imaging, Inc. | Camera and specimen alignment to facilitate large area imaging in microscopy |
TWI764786B (en) * | 2017-02-10 | 2022-05-11 | 美商奈米創尼克影像公司 | Method for imaging a specimen along a desired x'-direction of the specimen and microscope system |
US10048477B1 (en) | 2017-02-10 | 2018-08-14 | Nanotronics Imaging, Inc. | Camera and specimen alignment to facilitate large area imaging in microscopy |
US20200073106A1 (en) * | 2018-08-30 | 2020-03-05 | Keyence Corporation | Magnifying observation apparatus |
US10859806B2 (en) * | 2018-08-30 | 2020-12-08 | Keyence Corporation | Magnifying observation apparatus |
US10890749B2 (en) * | 2018-08-30 | 2021-01-12 | Keyence Corporation | Magnifying observation apparatus |
EP4343404A1 (en) * | 2022-09-20 | 2024-03-27 | Leica Microsystems CMS GmbH | Controller for an imaging system, system and corresponding method |
Also Published As
Publication number | Publication date |
---|---|
CN103105671B (en) | 2016-08-03 |
US8994810B2 (en) | 2015-03-31 |
JP2013083743A (en) | 2013-05-09 |
JP5841398B2 (en) | 2016-01-13 |
DE102012217281B4 (en) | 2022-08-25 |
CN103105671A (en) | 2013-05-15 |
DE102012217281A1 (en) | 2013-04-11 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US8994810B2 (en) | Magnification observation device | |
US9959451B2 (en) | Image inspection device, image inspection method and image inspection program | |
US9007452B2 (en) | Magnification observation device, magnification observation method, and magnification observation program | |
US9970745B2 (en) | Measurement target measuring program, measurement target measuring method, and magnifying observation device | |
US9383569B2 (en) | Magnification observation device | |
US10508902B2 (en) | Three-dimensional measurement device | |
US9122048B2 (en) | Image processing apparatus and image processing program | |
US10088301B2 (en) | Image measurement device | |
JP5841427B2 (en) | Image processing apparatus and image processing method | |
US10088302B2 (en) | Image measurement device | |
JP6800698B2 (en) | How to control the magnifying observation device and the magnifying observation device | |
JP2005318652A (en) | Projector with distortion correcting function | |
US11776110B2 (en) | Image measuring apparatus | |
US10379331B2 (en) | Magnifying observation apparatus | |
US10846881B2 (en) | Image measuring apparatus | |
JP6312410B2 (en) | Alignment apparatus, microscope system, alignment method, and alignment program | |
JP6140255B2 (en) | Image processing apparatus and image processing method | |
JP2017032360A (en) | Image inspection device, image inspection method and image inspection program | |
JP6882651B2 (en) | Measurement preparation of surface shape measuring device Alignment method and surface shape measuring device | |
JP2010266750A (en) | Observation device and observation system | |
JP2014115179A (en) | Measuring device, document camera and measuring method | |
US12013230B2 (en) | Three-dimensional shape measuring apparatus | |
JP2012154862A (en) | Three-dimensional dimension measuring device | |
JP2008091690A (en) | Appearance inspection equipment | |
JP2018066910A (en) | Magnifying observation device |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: KEYENCE CORPORATION, JAPAN Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:SEKIYA, SUGURA;KANG, WOOBUM;INOMATA, MASAHIRO;REEL/FRAME:028655/0003 Effective date: 20120713 |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 4 |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1552); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 8 |