1. Field of the Invention
This disclosure relates to an image acquiring apparatus.
2. Description of the Related Art
An image acquiring apparatus configured to acquire digital images by imaging an object (mount) attracts attention in a field of pathology and the like. The image acquiring apparatus enables a doctor to diagnose a pathological condition by using acquired image data. Since the diagnosis by the doctor is required to be accurate and speedy, the image data is required to be acquired at a high speed, and the acquired image data is required to be an image which contributes to easy diagnosis. In order to do so, it is effective to take an image of the mount at once over the largest possible area at a high resolution.
If an angle of view of an objective lens is increased to enlarge an image size which can be acquired at once, the image data can be acquired at a high speed. However, acquisition of an image focused over an entire angle of view becomes difficult. This is because a surface to be imaged of the object is not flat and has “waviness”, and hence part of image taking surface may not be included within a depth of focus of the objective lens.
In view of such a problem, US2013/0169788 discloses an image acquiring apparatus having a plurality of image taking systems and capable of changing at least one (posture) of a position and an inclination of each of the plurality of the image taking systems. By setting the plurality of the image taking systems to take their own positions, the posture of the image taking surface with respect to the objective lens can be changed. The postures of the respective image taking systems are controlled so that the entire image taking surface is included within the depth of focus of the objective lens by measuring the waviness of the surface to be imaged of the object.
Japanese Patent Laid-Open No. 2012-078330 discloses a technology of a lens inspection instrument configured to measure by moving a camera unit for correcting a movement of a camera unit automatically so that measurement of a lens being tested can be performed accurately and simply irrespective of a positioning accuracy of a three-axis stage to be moved. Specifically, before mounting the lens to be tested on a lens mount, a check plate is mounted on as a jig for focus checking, and a center portion and a peripheral portion of a pattern printed on the plate are imaged by the camera unit, whereby a best focus position is obtained. On the basis of a difference between a position of the camera unit at which the best focus is obtained at the center portion and a position at which the best focus is obtained in the peripheral portion, a correction coefficient in a direction of an optical axis when moving the camera unit in an in-plane direction perpendicular to the optical axis by a three-axis stage.
As disclosed in US2013/0169788, when controlling the posture of the image taking system, even when a driving device is controlled so as to achieve a target posture, a resulting posture may become unintended posture since a movement component of another axis (different axis movement component) different from an intended operating axis is superimposed to an intended movement component. For example, in the case where the image taking system is controlled so that the inclination with respect to the optical axis is changed, the posture may be misaligned with the target posture by the movement of the position also in a direction perpendicular to the optical axis. The movement in the direction perpendicular to the optical axis is a different axis movement component.
If there is the different axis movement component, part of the image taking surface which has controlled so as to follow waviness of the surface to be imaged moves out of the depth of focus, so that a blurring image may be acquired. Therefore, a portion excluding a peripheral edge portion of an effective pixel area of the image taking system is treated as a usable area for formation of the image data (image formation), so that an effective usage of pixels is interfered.
However, in US2013/0169788 and Japanese Patent Laid-Open No. 2012-078330, a specific method of correcting the different axis movement component as described above is not disclosed. In addition, in the different axis movement component is caused by a mechanism error, deformation, a measurement error, a control computation error, and the like, and a state of variation is not monotonous, and regularity and reproducibility such as a non-linear shape, hysteresis, and a change with time are low. Therefore, even when a compensation coefficient for an assumed target value is acquired and a correction is performed on the basis of this coefficient, sufficient effects may not be obtained.
An aspect of this disclosure is an image acquiring apparatus configured to acquire an image of an object by joining a plurality of divided images obtained by taking images of a plurality of divided areas in the object, including: an imaging optical system configured to image light from the object; an image taking element configured to take an image of the object; a changing mechanism configured to change a posture of the object or the image taking element; a control unit configured to calculate a control target value for causing the changing mechanism to reach a target posture; and a correcting mechanism configured to correct the posture such that a reached posture after the changing mechanism has changed the posture in accordance with the control target value so as to approach the target posture, wherein the control unit compares reached image data obtained as a result that the image taking element actually takes an image of a correction chart whereof drawing information is known in a state that the posture is the reached posture, and target image data which is expected to be obtained when the image taking element takes an image of the correction chart in a state that the posture is the target posture to calculate a correction value of the posture, and the correcting mechanism corrects the posture on the basis of the correction value.
Further features of the present invention will become apparent from the following description of exemplary embodiments with reference to the attached drawings.
In an image acquiring apparatus described in embodiments given below, a transmission-type digital microscope is described as the image acquiring apparatus and a mount is described as an object as an object of acquisition of the image as preferable examples. However, this disclosure is not limited thereto. Numerical values exemplified for specifically describing the disclosure are not limited unless otherwise specifically noted. In the respective drawings, the same members are denoted by the same reference numerals and overlapped description is omitted.
Referring now to
The apparatus 100 includes the objective lens 102, an image taking unit 103, a stage 104, a preliminary measuring unit 105, a control unit 106, a display unit 107, a correcting chart 108 installed on the stage 104 (hereinafter, referred to as a “chart 108”).
A mount 101 is an image-acquired object (object) which is an object of the image acquisition. The mount 101 includes a cover glass, a sample 11 such as a sample of a living body, that is, a section of tissue, and a slid glass, and the sample 11 arranged on the slide glass is sealed with a cover glass and an adhesive agent. The mount 101 is arranged on the stage 104, is measured preliminary by the preliminary measuring unit 105, and then moved by the stage 104 on the basis of a preliminary measurement result, and the image of the mount 101 is taken by the image taking unit 103 via the objective lens 102.
The objective lens 102 has an imaging optical system configured to image the mount 101 and, specifically, is an imaging optical system for forming an image on reflecting surfaces of reflecting members 31 in the image taking unit 103 described later while enlarging an image of the mount 101 at a predetermined magnification. The objective lens 102 is retained by a body frame and a lens barrel, which are not illustrated, and is configured by a combination of a lens and a mirror. The objective lens 102 is arranged so that the reflecting surfaces of the reflecting members 31 of the image taking unit 103 and the mount 101 are optically conjugated, an object side corresponds to the mount 101, and an image side corresponds to the reflecting surface. A numerical aperture NA on the object side of the objective lens 102 is 0.7 or larger, and can be configured so that an image of an area of at least 10 mm×10 mm on an object surface can be formed desirably at once.
The image taking unit 103 has a plurality of individual image taking units 103A to 103D at a portion which takes an image of the mount 101 imaged by the objective lens 102. The image taking unit 103 is retained by the body frame or the lens barrel of the objective lens, which are not illustrated.
The configuration of the individual image taking units 103A to 103D will be described with reference to
The reflecting surface of the reflecting member 31 and the image taking surface of the image taking element 33 are arranged so as to be optically conjugated with respect to the re-imaging unit 32. The object side corresponds to the reflecting surface, and the image side corresponds to the image forming surface. Also, an optical axis of the objective lens 102 and an optical axis of the re-imaging unit 32 are orthogonal to each other via the reflecting member 31. A (two-dimensional) image taking element such as a CCD or a CMOS sensor may be used as the image taking element 33. The individual image taking units 103B to 103D have the same configuration.
The number of the individual image taking units mounted on the apparatus 100 is determined as needed depending on a surface area of the field of view of the objective lens 102. The arrangement and the configuration of the individual image taking units to be mounted are also determined as needed depending on the shape of the field of view of the objective lens 102 and the shape and the configuration of the image taking element 33. In this embodiment, as an example, 2×2 individual image taking units 103A to 103D are arranged on an X-Y plane. The individual image taking units 103A to 103D may include a plurality of reflecting members 31, or may have a configuration in which the reflecting member 31 and the re-imaging unit 32 are not provided and the images imaged by the objective lens 102 are directly taken by the image taking element 33.
Here, in the case where the postures of the respective image taking element 33 are described in the following description, a coordinate system illustrated in
In general, the individual image taking unit includes an area such as a substrate on which the image taking elements 33 are mounted in the periphery of the image taking surface of the image taking element 33, and hence it is difficult to arrange the plurality of the image taking elements 33 adjacently without gap. Therefore, it is not easy to arrange the individual image taking units 103A to 103D having the image taking element 33 adjacent to each other, and is obliged to arrange apart from each other as illustrated in
The retaining member 332 is fixed to the moving members 333A to 333C, and the changing mechanisms 334A to 334C move the moving members 333A to 333C, so that the posture of the image taking element 33 can be changed. A mechanism using a linear actuator having a linear motor, an air cylinder, a stepping motor, or an ultrasonic wave motor, and the like may be used as the changing mechanism 334A to 334C and the correcting mechanism 335. A rotating function around the Zsa axis may be added to the correcting mechanism 335.
The stage 104 is a position changing unit configured to change the position of the mount 101 by moving in the state of supporting the mount 101. The stage 104 includes a supporting portion configured to support the mount 101, an XY stage configured to move the supporting portion in the XY direction, and a Z stage configured to move the supporting portion in the Z direction (these members are not illustrated). The XY stage and the Z stage move the supporting portion in accordance with the control target value output from the control unit 106.
The XY stage (not illustrated) is configured to allow the mount 101 to move between a range which can be preliminary measured by the preliminary measuring unit 105 of the apparatus 100 (preliminary measurement range) and a range in which image can be taken by the image taking unit 103 (image taking executing range). In the image taking executing range, the relative position between the mount 101 and the image taking unit 103 is changed by moving the XY stage as illustrated in
The preliminary measuring unit 105 has a function to perform measurement for acquiring a present area of the sample 11 included in the mount 101 in the preliminary measurement range and a function to perform measurement for acquiring information on a surface to be imaged 15 of the sample 11. The measurement for acquiring the information on the surface to be imaged 15 is, for example, measurement of waviness on the upper surface of the cover glass included in the mount 101. A specific configuration of this case may be the same as that disclosed in US2013/0169788, and detailed description will be omitted here.
Alternatively, a configuration further including measurement of the thickness of the cover glass included in the mount 101 and the measurement result thereof and a measurement result of the waviness on the upper surface of the cover glass are used to acquired information on waviness on a lower surface of the cover glass close to the upper surface of the sample 11 is also applicable. Alternatively, a configuration including a measuring function for measuring the amount of change of contrast and the amount of transmissive light with respect to illumination light having a specific wavelength from the image taking result obtained by taking images of the plurality of the different positions in the Z direction of the sample 11 included in the mount 101 is also applicable.
The control unit 106 controls the respective configurations of the apparatus 100 and generates the image data for observation by using the image taking result of the image taking unit 103. The control unit 106 is composed of a multi-purpose computer or a work station providing a high-speed arithmetic processing such as a CPU, a memory and a hard disk, and a specific graphic board, or a combination thereof. The control unit 106 includes an interface, which is not illustrated, which allows the user to change the setting of the apparatus 100 or to input drawing information of the chart 108 described later.
The generating unit 61 has a function to generate observation image data by processing the image data of the mount 101 acquired by the image taking unit 103. Specifically, positions of a plurality of divided image data acquired by a plurality of times of image taking while moving the stage 104 in the XY direction are aligned and these divided image data are connected to generate the observation image data so as to be displayed in the display unit 107.
The calculating unit 62 obtains a control target value of the mechanism control unit 64 for controlling the stage 104 of the apparatus 100 on the basis of the preliminary measurement result measured by the preliminary measuring unit 105. A configuration in which the mount 101 is preliminary measured by an apparatus other than the apparatus 100 and, the result is acquired to calculate the control target value is also applicable. Specifically, the present area of the sample 11 included in the mount 101 is acquired by using the preliminary measurement result of the preliminary measuring unit 105. On the basis of the present area of the sample 11, the surface to be imaged 15 for generating the observation image data is selectively determined. The calculating unit 62 divides the surface to be imaged 15 into divided areas that the single image taking element 33 can take the image at once, and determines the order of image taking among the respective divided areas and the position to which the stage 104 is to be moved for taking the images of the respective divided areas. Here, a control target value table in which the order of movement of the stage 104 and the position thereof are shown is generated.
The mechanism control unit 64 is capable of controlling the movement of the stage 104 on the basis of the acquired control target value table, and acquiring only the image data of the area to be imaged. Accordingly, the image data of the area required for the pathological diagnosis may be selectively acquired, and hence the capacity of the observation image data may be reduced, so that handling of the observation image data is facilitated. Normally, the surface to be imaged 15 is determined so as to be equal to the area where the sample 11 is present.
The calculating unit 62 acquires information on the surface to be imaged 15 of the sample 11 included in the mount 101 from the measurement result of the preliminary measuring unit 105. On the basis of the magnifications of the objective lens 102 and the re-imaging unit 32, an imaging surface (imaging curve) on which the optical flux from the surface to be imaged 15 of the sample 11 images via the objective lens 102 is calculated. Approximate planes of the imaging surfaces are calculated for the respective divided areas, and control target values of the changing mechanism 334A to 334C of the image taking element 33 required for aligning the respective image taking surfaces of the individual image taking units 103A to 103D with the acquired approximate planes are determined. On the basis of the control target value, the mechanism control unit 64 controls the changing mechanism 334A to 334C to change the posture of the image taking element 33, so that a desired image with less out-of focus can be acquired.
Accordingly, the calculating unit 62 divides the surface to be imaged 15 of the sample 11 into the divided areas 12, and calculates control target value of the stage 104 so that the respective divided areas move efficiently to XY positions of the image taking areas 14A to 14D. Approximate planes of the imaging surfaces projected from the respective divided areas 12 are calculated by the calculating unit 62, and control target values of the changing mechanism 334A to 334C of the image taking element 33 required for aligning the respective image taking surfaces with the calculated approximate planes are determined. The mechanism control unit 64 controls the stage 104 and the changing mechanism 334A to 334C so that the relationship between the surface to be imaged 15 and the image taking areas 14A to 14D of the image taking unit 103 becomes as the posture control example 16 on the basis of the control target value that the calculating unit 62 acquires.
An example of the configuration of the movement mechanism 330 of the image taking element 33 is illustrated in
In
The different axis movement component will be described with reference to
This is caused by a presence of a physical distance such as the length of the moving members 333A and 333C, the thickness of the substrate 331 and the thickness of the retaining member 332 between the resilient portion 336A and 336C and the image taking surface as illustrated in
Problems occurring because of the presence of the different axis movement component will be described with reference to
As illustrated in
As illustrated in
Accordingly, in order to solve such a problem, the control unit 106 includes the calculating unit 63. The calculating unit 63 calculates a correction value with which the mechanism control unit 64 controls a correcting mechanism 335 so that the image taking unit 103 can take the image at a posture in which a misalignment due to the different axis movement component is alleviated. Specifically, the calculating unit 63 calculates the target image data which is expected to be acquired when the image taking unit 103 takes an image of the chart 108 in the target posture from the known drawing information of the chart 108 and the control target value of the image taking unit 103.
In contrast, the calculating unit 63 acquires the reached image data obtained by the image taking unit 103 as a result of image taking of the chart 108 having known drawing information in a reached posture reached after the image taking unit 103 is controlled in accordance with the control target value. The calculating unit 63 compares the target image data and the reached image data, analyzes the different axis movement component superimposed with the reached posture of the image taking surface of the image taking unit 103, and calculates the correction value.
On the basis of the correction value, the apparatus 100 is controlled via the mechanism control unit 64 to bring the posture of the image taking unit 103 into a state in which the different axis movement component is reduced, so that a desirable image with less blurring can be obtained while effectively using the pixel area of the image taking element.
The mechanism control unit 64 controls the changing mechanism 334A to 334C and the correcting mechanisms 335A to 335C configured to move the stage 104 and the image taking unit 103 on the basis of the result of calculation of the calculating unit 62 and the result of calculation of the calculating unit 63.
The display unit 107 has a function to display the observation image suitable for pathological diagnosis on the basis of the observation image data that the generating unit 61 has generated. The display unit 107 may be composed of a monitor such as a CRT or liquid crystal.
A member on which chart patterns are drawn by a general fine machining such as laser processing or photo-etching may be used as the chart 108. The chart 108 is installed on the stage 104, and is configured to be arranged within the image taking executing range while the mount 101 is arranged within the preliminary measurement range. Alternatively, the chart 108 may be configured to be arranged in the image taking executing range after the control target value is calculated by the calculating unit 62 until the mount 101 is moved to the image taking executing range. The image of the chart 108 is taken by the image taking unit 103 in a state of having reached the reached posture and the calculating unit 63 acquires the reached image data.
The drawing information of the chart 108 is memorized in a memory 65 of the control unit 106. The calculating unit 63 calculates the target image data which is expected to be acquired when the image of the chart 108 is taken in a state in which the image taking unit 103 reaches the target posture from the drawing information and the control target value of the image taking unit 103.
The patterns 811 and 821 out of the patterns 811 to 813, and 821 to 823, which are arranged at the lowest positions, may be determined so as to match the height of the assumed lowest surface of the sample 11 (the upper surface of the slide glass). The patterns 813 and 823 arranged at the highest positions may be determined so as to match the assumed uppermost surface of the sample 11. The patterns 811 and 821 arranged on the outermost side in the Y direction may be determined on the basis of the range of pixel area of the image taking element 33 of the individual image taking unit 103A in the Ypa direction.
In addition,
The target image data 331D is capable of acquiring from the positional information on the respective patterns memorized in the control unit 106 and the image taking area 337P on the basis of the control target value of the image taking unit 103 by the calculating unit 63. As illustrated in
Specifically, the difference between the center of the pattern data 812D and the center of the pattern data 812PD are determined as the different axis movement component. At this time, the same difference may be obtained by using the pattern data 823D and the pattern data 823PD to calculate the different axis movement component from the result of the difference therebetween. When comparing the target image data 331D and the reached image data 331PD, the areas used for comparison is preferably set to areas 831 and 832 so that two or more pattern data is not included. The blurred pattern data 813PD and 822PD acquired from the pattern 813 and 822 are reflected on the reached image data 331PD. However, these pattern data 813PD and 822PD are preferably removed from the object used for the analysis of the different axis movement component.
The example of the case where the image taking surface of the individual image taking unit 103A is moved in the θXsa direction has been described. This disclosure is also applicable to the case where the image-taking surface is moved in the θysa direction and the both direction by configuring the drawing portion 80A as illustrated in
The specific number of the patterns and the positional relationship of the drawing portions 80A to 80D are determined as needed depending on the shape and configuration of the objective lens 102 or the image taking unit 103.
Subsequently, the image acquisition method using the apparatus 100 will be described with reference to a flowchart in
The acquisition of the drawing information of the chart 108 may be performed with a different apparatus as described above, or the corresponding function may be mounted on the control unit 106 of the apparatus 100. Alternatively, without using the different apparatus as described above, a configuration in which the apparatus 100 is used, the image of the chart 108 is taken under a stable temperature environment in a reference posture, and the drawing information corresponding to this image is prepared and memorized is also applicable. In addition, without taking the image by using the apparatus 100 or the different apparatus, a configuration in which the drawing information is memorized on the basis of design data when manufacturing the chart 108 is also applicable. The drawing information may be entered as needed via the interface, which is not illustrated, without being memorized in the memory of the apparatus 100.
In this manner, the step S61 is a preparatory step of an image acquisition action of the apparatus 100. Whether or not the action of step S61 is to be performed may be selected every time when the image is acquired, or step S61 may be performed only at the time of adjustment for the first time, such as at the manufacture.
In the next step S62, the mechanism control unit 64 controls the stage 104 so as to move the mount 101 to the preliminary measurement range. In the preliminary measurement range, the preliminary measuring unit 105 performs a preliminary measurement of the mount 101.
In the nest step S63, the calculating unit 62 determines the control procedure for the stage 104 and the movement mechanism 330. Specifically, the calculating unit 62 determines the surface to be imaged 15 that generates the observation image data of the sample 11 included in the mount 101 on the basis of the preliminary measurement result, and calculates the imaging surface via the objective lens 102 and the re-imaging system 32 for taking the image of the surface to be imaged 15. On the basis of the results described above, the calculating unit 62 determines a control procedure for the position changing unit 104 and the changing mechanism 334A to 334C by the mechanism control unit 64 for acquiring data of the image of the surface to be imaged 15 as a control target value table.
In contrast, the image taking unit 103 takes an image of the chart 108 within an image taking range. On the basis of the results described above, the calculating unit 63 determines the control procedure for the correcting mechanism 335 by the mechanism control unit 64 as the control correction value table. A method of determining the control procedure in step S63 will be described later with reference to
In the next step S64, the mechanism control unit 64 controls the stage 104 so that the mount 101 moves to the image taking range. Then, the image taking unit 103 takes an image of the mount 101. Here, in accordance with the procedure determined in step S63 (the control target value table that moves the stage 104), the stage 104 is controlled so that the relative position between the mount 101 and the image taking unit 103 changes.
At the same time as the respective movement of the stage 104, the posture of the image taking unit 103 is controlled in accordance with the procedure determined in step S63 (the control target value table and the control correction value table that move the image taking unit 103). At the completion of the control described above, the image taking unit 103 takes an image of the mount 101, and the control unit 106 acquires the divided image data from the image taking unit 103.
In the final step S65, positions of a plurality of divided image data acquired in step S64 are aligned and these divided image data are connected to generate the observation image data so as to be displayed in the display unit 107. The action of connecting the divided image data may be performed in parallel to the acquisition of the image data in step S64.
In step S631, the calculating unit 62 calculates the surface to be imaged 15 that generates the observation image data of the sample 11 included in the mount 101 and the imaging surface via the objective lens 102 and the re-imaging system 32 on the basis of the preliminary measurement result. Subsequently, in order to control the movements of the stage 104 and the changing mechanism 334A to 334C of the image taking unit 103, the control target value table as shown in Table 1 will be prepared. Table 1 includes control target values for moving the stage 104 and the image taking unit 103 for aligning the unit areas (image taking areas) that the respective individual image taking units 103A to 103D can take images with the divided areas in the surface to be imaged 15 in a predetermined order for each control procedure.
In the following step S632, the calculating unit 63 determines whether or not a standard control correction value table is to be updated as a previous step for updating the control correction value table at a posture of the image taking unit 103. In the standard control correction value table, relationship information between the control target values (standard control target values) in the range assumed for the changing mechanism 334A to 334C and the correction values (standard control correction value) required when being controlled on the basis of the respective standard control correction values are listed as Table 2, for example. In Table 2, an example in which the standard control correction value required was (Xsa1T[n], Ysa1T[n]) in the case where the respective changing mechanism 334A to 334C of the individual image taking unit 103A were controlled with the standard control target value (Zsa1T[n], Zsa2T[0], Zsa3T[0]) is shown.
In the case where the calculating unit 63 determines that update of the standard control correction value table is not necessary, the procedure goes to step S637, and the calculating unit 63 updates the control correction value table on the basis of the current standard control correction value table. In the case where the calculating unit 63 determines that update of the standard control correction value table is necessary, the procedure goes to step S633, and the calculating unit 63 starts acquisition of the correction value to be listed in the standard control correction value table. In the subsequent step S633, the calculating unit 63 determines whether or not the standard control correction value to be listed in the standard control correction value table is to be newly calculated.
In the case where the calculating unit 63 determines that new calculation of the standard control correction value is necessary, the calculating unit 63 determines a standard control target value for newly calculating the standard control correction value from the standard control correction value table. Then, the procedure goes to step S634, where acquisition of the correction value for the determined standard control target value is started. In contrast, in the case where the calculating unit 63 determines that new calculation of the standard control correction value is not necessary, the procedure goes to step S637, and the calculating unit 63 updates the control correction value table on the basis of the current standard control correction value table.
In step S634, the calculating unit 63 acquires the target image data. Specifically, the calculating unit 63 acquires target image data acquired when the image taking unit 103 takes an image of a chart in the target posture with the standard control target value selected from the drawing information of the chart 108 stored in step S61 and the standard control target value selected in step S633.
In the next step S635, the mechanism control unit 64 controls the image taking unit 103 on the basis of the standard control target value determined in step S633. Then, an image of the chart 108 is taken in the reached posture to acquire the reached image data.
In the following step S636, target image data acquired in step S634 and reached image data acquired in step S635 are compared, and analyzes the different axis movement component superimposed with the reached posture of the image taking unit 103. Then, a correction value to be allocated to the standard control target value selected in step S633 is calculated. Subsequently, the posture of the image taking element 33 is corrected on the basis of the calculated standard control correction value, and, a sequence in which a flow of steps S635 to S636 is performed again in this state, and a re-calculated correction value is reflected on the standard control correction value calculated before may be performed. Alternatively, a sequence may be repeated until an amount of change of the re-calculated correction value and the standard control correction value calculated before becomes a predetermined value or lower.
In step S637, on the basis of the current standard control correction value table, the control correction value table for correcting the posture of the image taking unit 103 as shown in Table 3 is updated. In Table 3, control correction values of the respective correction mechanism for correcting the posture of the image taking unit 103 are listed according to the order of control recorded in the control target value table (Table 1).
In order to update the control correction value table on the basis of the standard control correction value table, a general interpolation may be used. For example, in the examples in Table 1 to Table 3, the control correction value Xsa[N] in the Xsa direction listed in the Nth order of control is obtained by Expression (1).
X
sa
[N]=X
sa1T
[N]+X
sa2T
[N]+X
sa3T
[N] (1)
where: the control correction value of the changing mechanism 334A is Xsa1T[N], the control correction value of the changing mechanism 334B is Xsa2T[N], and the control correction value of the changing mechanism 334C is Xsa3T [N].
Here, the control correction values Xsa1T[N], Xsa2T[N], Xsa3T[N] are expressed respectively by Expressions (2) to (4).
X
sa1T
[N]=X
sa1T
[n]+a
1*(Xsa1T[n+1]−Xsa1T[n]) 0<a1<1 (2)
X
sa2T
[N]=X
sa2T
[n]+a
2*(Xsa2T[n+1]−Xsa2T[n]) 0<a2<1 (3)
X
sa3T
[N]=X
sa3T
[n]+a
3*(Xsa3T[n+1]−Xsa3T[n]) 0<a3<1 (4)
Since the target control values Zsa1[N], Zsa2[N], and Zsa3[N] of the changing mechanisms 334A to 334C for achieving the target posture can be expressed by Expression (5) to (7), a coefficient a1, a2, and a3 can be obtained.
Z
sa1
[N]=Z
sa1
[n]+a
1*(Zsa1[n+1]−Zsa1[n]) (5)
Z
sa2
[N]=Z
sa2
[n]+a
2*(Zsa2[n+1]−Zsa2[n]) (6)
Z
sa3
[N]=Z
sa3
[n]+a
3*(Zsa3[n+1]−Zsa3[n]) (7)
The control correction values in the Ysa direction can be obtained in the same manner.
The update of the standard control correction table by the flow in steps S633 to S636 may be performed selectively depending on whether it is necessary or not every time when acquiring the image, or may be performed only at the time of adjustment for the first time such as the time of manufacture. Alternatively, it is also possible to make the standard control target value of the standard control correction value table match the control target value of the image taking unit 103 listed in the control target value table prepared in step S632, and uses the updated standard control correction value table as the control correction value table as-is. In this case, the update of the control correction value table by the flow in steps S633 to S636 is performed every time when the image acquisition action is performed. However, a correction value in compliance with the actually controlled target posture, and corresponding to the different axis movement component having lower regularity can be obtained.
Alternatively, an abnormal control target value is selected from the control target values for the image taking unit 103 listed in the control target value table prepared in step S632 is added to the standard control correction value table. Then, it is also applicable to calculate the correction value for the abnormal control target value selectively by each of the image acquisition actions, and reflect the result to the control correction value table as-is.
Here, the abnormal control target value is a control target value which specifically requests a posture having low regularity in comparison with other control target values, and can be determined from specifications of a control mechanism and the correcting mechanism or a calculated history of the movement component when the control target value is calculated in step S632. Alternatively, when where the control target value is calculated in step S632, in the case where the difference in comparison result is a predetermined value or larger in comparison with the plurality of the standard posture target values of the standard control correction table, the corresponding value may be determined as the abnormal posture target value.
In this manner, in the image acquiring apparatus configured to be capable of changing the position of the image taking element in the direction of the optical axis and the inclination of the image taking element with respect to the optical axis so as to follow the waviness of the surface to be imaged 15 of the sample 11 included in the mount 101, the displacement caused by the different axis movement component can be corrected.
Accordingly, images having less burring in the respective postures can be acquired, and the pixels of the image taking element can be effectively used.
Specifying the posture of a substance moving with the different axis movement component such as the image taking element 33 superimposed therewith by the measurement of the angle as needed is difficult in many cases. Therefore, providing the measuring device in the direction other than the direction of the drive axis may become disadvantageous in terms of space and cost. According to the apparatus 100, it is not necessary to provide the measuring device for measuring the angle, and the posture of the image taking element 33 can be corrected easily.
The example of enabling the acquisition of the image in which the influence of the waviness of the surface to be imaged 15 of the sample 11 included in the mount 101 is suppressed by bringing the image taking surface to approach to the imaging surface by performing change control and correction control on the posture of the image taking elements 33 that the respective individual image taking unit have has been described. For example, however, the same effects are also achieved by performing the movement control and the correction control on the position and the posture of the stage 104, the reflecting member 31, and the re-imaging system 32 or by combining these controls. For example, with a configuration in which the imaging surface is brought to approach to the image taking surface by the movement control on the posture of the reflecting member 31 and the different axis movement component generated at the time is corrected by performing the movement control on the position of the image taking element 33, the arrangement and the configuration of the respective members may be optimized. Also, by using a parallel link mechanism, an integral configuration of the changing mechanism 334A to 334C and the correcting mechanism 335 is also applicable.
Although the individual image taking unit group in which the image taking elements are arranged two dimensionally has been described, the individual image taking unit group in which the image taking elements are arranged one dimensionally or three dimensionally may also be used. Although two dimensional image taking element is used as the image taking element, a one-dimensional image taking element (line-sensor) may also be used.
The system 200 includes the apparatus 100, a display device 201, and an image server (image memory device) 202. The apparatus 100, the display device 201, and the image server 202 are connected by a general-purpose LAN cable 204 via a network 203. Alternatively, a configuration in which between the image server 202 and the apparatus 100 or between the apparatus 100 and the display device 201 is connected with a general-purpose I/F cable is also applicable.
The image server 202 has a function to store the observation image data generated by the apparatus 100. The apparatus 100 has a function (not illustrated) to acquire the observation image data from the image server 202 and to re-edit the observation image data for displaying the image or information suitable for the pathological diagnosis in addition to the function described in the first embodiment. Other configurations are the same as those of the apparatus 100 described in conjunction with
The display device 201 is equivalent to the display unit 107, and has a function to display the observation image suitable for the pathological diagnosis on the basis of the observation image data that the apparatus 100 has generated. The display device 201 includes an interface, which is not illustrated, which allows the user to change the setting of the apparatus 100 or to input drawing information of the chart 108. A monitor which constitutes part of the display device 201 may be configured as a touch panel.
In the system 200 configured in this manner, components can be arranged remotely, so that the user is capable of acquiring images or displaying images by a remote control.
An image acquiring method of the system 200 will be described with reference to a flowchart illustrated in
In the next step S73, the calculating unit 62 determines the control procedure which moves the stage 104 and the image taking unit 103 as shown in Table 1 on the basis of the preliminary measurement result. This procedure is the same as in steps S631 to S636 described in conjunction with
In the next step S74, the mechanism control unit 64 controls the stage 104 and the changing mechanism 334A to 334C so that the mount 101 moves to the image taking range of the image taking unit 103. The correcting mechanism 335 is controlled on the basis of the correction value acquired in step S73. Subsequently, the image taking unit 103 takes an image of the mount 101, and the generating unit 61 acquires the divided image data, which is an image taking result of the image taking unit 103.
In the next step S75, the generating unit 61 determines whether or not a series of the image acquisition control listed in the control target value table updated in step S73. In other words, the generating unit 61 determines whether or not there is an area whereof the image is to be taken among the divided areas in the surface to be imaged 15. When it is determined that the image acquisition control is to be completed, the procedure goes to step S76, and the generating unit 61 generates observation image data. When it is determined that the image acquisition control is not to be completed, the procedure goes to step S72, where the image acquisition process in compliance with the control target value table is continued. In this case, a control target value corresponding to the divided area whereof the image is to be taken next is selected from the control target value table to acquire a correction value corresponding to the control target value.
In the final step S76, positions of a plurality of divided image data acquired in the flow from steps S72 to S75 are aligned, the divided data are connected to generate the observation image data so as to be displayed in the display unit 107. The action of connecting the divided image data may be performed in parallel to the acquisition of the divided image data in the flow of the steps S72 to S75.
In the case of the second embodiment, since the steps S72 to S74 are repeated, the stage 104 is moved to the range where the images of the chart 108 and the mount 101 can be taken every time where the posture control of the image taking unit 103 is performed. However, acquisition of a correction value corresponding to the different axis movement component which has low reproducibility is achieved in compliance with the posture which is actually reached. In combination with the image acquisition method of the first embodiment, the flow of the steps S72 to S74 may be performed selectively only for the posture control of the abnormal control target value which is expected to have low reproducibility.
In this manner, in the image acquiring apparatus configured to be capable of changing the position of the image taking element in a direction of the optical axis and the inclination of the image taking element with respect to the optical axis so as to follow the waviness of the surface to be imaged 15 of the sample 11 included in the mount 101, the displacement caused by the different axis movement component can be corrected. Consequently, the pixel of the image taking element may be effectively used. In addition, images with less blurring may be obtained further stably at each postures.
In a third embodiment, a recording medium (or a memory medium) in which a software program code which realizes the entire part or part of the functions of the respective embodiments described above is recorded is supplied to the system or the apparatus. A program is executed by computer (or a CPU or an MPU) of the system 200 or the apparatus 100 by reading and executing the program code stored in the recording medium. In this case, the program code which is read out form the recording medium realizes the function of the above-described embodiment, and the recording medium itself which records the program code constitutes part of this disclosure.
The computer executes the read-out program code, so that an operating system (OS) or the like working on the computer performs part or the entire part of the actual process. The case where the functions of the above-described embodiment are realized by the process described above is also included in this disclosure.
It is assumed that the program code read out from the recording medium is written in the memory provided on a function enhancement card inserted into the computer or a function enhancement unit connected to the computer. Subsequently, the case where the CPU or the like provided on the function enhancement card or the function enhancement unit executes part or the entire part of the actual process on the basis of the instruction of the program code whereby the above-described functions of the embodiments are realized is also included in this disclosure.
In the case where this disclosure is applied to the recording medium described above, the program code corresponding to the flowchart described above is stored in the recording medium.
Although the preferred embodiments of this disclosure have been described, this disclosure is not limited to those embodiments, and various modifications or variations may be made within the scope of this disclosure. The configurations described in conjunction with the first to third embodiments may be combined with each other. Therefore, to configure a new system by combining various technologies as needed in the above-described embodiments may be found easily by those skilled in the art, so that systems achieved by various combinations are also included within the scope of this disclosure.
For example, in the image acquisition action of the image acquiring apparatus according to the respective embodiments described above, NA of the objective lens 102 may be set to different values when taking an image of the mount 101 and when taking an image of the chart 108, which are effective way for achieving highly accurate detection of the different axis movement component. Specifically, when taking the image of the chart 108 from which acquisition of imaged data of the intended pattern is wanted, the NA is set to a higher value than the NA selected when taking the image of the mount 101 from which the image data is acquired over the entire part of the range to be imaged. This allows acquisition of the image at a high-resolution, a different axis movement component can be detected at a high degree of accuracy.
In contrast, when taking an image of the chart 108 from which imaged data of the intended pattern is wanted in various postures, the NA is set to a larger depth of focus (lower NA) than that selected when taking an image of the mount 101 from which data is acquired in the posture following the approximate imaging surface of a range whereof the image is to be taken. In this configuration, the pattern of the chart 108 may be formed at only a single level, whereby the process can be simplified together with the process of image processing thereof and detecting the different axis movement component.
Adjustment of the NA is considered to be effective also when correcting distortions in order to detect a different axis movement component with further higher degree of accuracy. In order to detect the different axis movement component, detection of positions of centers of gravity of the respective patterns in reached image data 331PD is effective. At this time, the contrast of the reached image data 331PD may be adjusted by changing the NA, so that the accuracy of detection of the positions of the centers of gravity of the respective patterns may be enhanced.
As a configuration for adjusting the NA, an NA diaphragm which allows an arrangement of a plurality of field-of-view shielding plates having different apertures depending on the application or an iris diaphragm composed of a plurality of field-of-view shielding blades may be used. Furthermore, the imaging position of the objective lens 102 is set to different values when taking an image of the mount 101 and when taking an image of the chart 108. In this configuration, in the same manner as the NA adjustment described above, the pattern of the chart 108 may be formed at only a single level, whereby the process can be simplified together with the process of image processing thereof and detecting the different axis movement component.
A mechanism using a linear actuator having a linear motor, an air cylinder, a stepping motor, or an ultrasonic wave motor, and the like may be used as a configuration of adjusting the imaging position of the objective lens 102. Such a configuration is used at a connecting portion between a body frame, which is not illustrated, and a lens barrel of the objective lens 102, or a connecting portion between a lens and a mirror in the objective lens 102 and the lens barrel.
In the embodiments described above, the surface to be imaged 15 in the sample 11 is determined and the observation image data regarding the surface to be imaged 15 is acquired. However, this disclosure is not limited thereto, and a configuration in which the stage 104 is moved in the Z direction after the image taking following the waviness of the surface to be imaged 15, and images of a plurality of surfaces different in position in the Z direction are taken to acquire a three-dimensional image is also applicable.
In the case where the positional misalignment due to the different axis movement component is reduced with the method described in the embodiments described above, this disclosure is not limited to a configuration in which the posture of the image taking element 33 is changed as in the embodiments described thus far, and for example, and a configuration in which the posture of the stage 104 is changed is also applicable. The method described above is not limited to the positional misalignment generated by a change of the posture for causing the image taking surface of the image taking element 33 to follow the surface to be imaged 15, but may be used for reducing the positional misalignment due to the different axis movement component caused in association with the movement of the stage 104.
Furthermore, a configuration in which a plurality of charts 108 are arranged on the stage 104 and a plurality of sets of correction value groups including a plurality of correction values are acquired by using respective charts is also applicable. For example, a final correction value is acquired from a mean value in the plurality of the sets of the correction value groups. In this case, a center of gravity (center of line if there are two) of the polygon obtained by connecting centers of gravity of the plurality of the charts 108 can be matched with the center of the mount 101 or the sample 11 placed on the stage 104 or the position where a portion in the vicinity of the center is arranged.
In this configuration, a difference of the correction values caused by the difference between the position of the mount 101 and the position of the chart 108 may be reduced.
According to the image acquiring apparatus as an aspect of the present disclosure, in the image acquiring apparatus configured to be capable of changing the posture of the imaging taking unit so as to follow the waviness of the surface to be imaged of the object, a difference due to the different axis movement component which superimposes when changing the posture of the image taking unit is corrected to allow the pixels of the imaging taking unit to be used effectively.
While the present invention has been described with reference to exemplary embodiments, it is to be understood that the invention is not limited to the disclosed exemplary embodiments. The scope of the following claims is to be accorded the broadest interpretation so as to encompass all such modifications and equivalent structures and functions.
This application claims the benefit of Japanese Patent Application No. 2014-156794 filed Jul. 31, 2014, which is hereby incorporated by reference herein in its entirety.
Number | Date | Country | Kind |
---|---|---|---|
2014-156794 | Jul 2014 | JP | national |