The present invention relates to a sample observation technique. As an example, the present invention relates to a device having a function of observing a defect, an abnormality, and so on (sometimes collectively referred to as defect) and a circuit pattern in a sample such as a semiconductor wafer.
In semiconductor wafer manufacturing, it is important to quickly start a manufacturing process and shift early to a high-yield mass production system. For this purpose, various inspection devices, observation devices, measuring devices, and so on are introduced in a production line. A sample observation device (also referred to as defect observation device) has a function of imaging a semiconductor wafer surface defect position with high resolution and outputting the image based on defect coordinates in defect position information inspected and output by an inspection device. The defect coordinates are coordinate information representing the position of a defect on a sample surface. In the sample observation device, a scanning electron microscope (SEM) or the like is used as an imaging device. Such sample observation devices are also called review SEMs and widely used.
Observation work automation is desired in semiconductor manufacturing lines. The review SEM includes, for example, automatic defect review (ADR) and automatic defect classification (ADC) functions. The ADR function is to perform, for example, processing to automatically collect images at sample defect positions indicated by defect coordinates in defect position information. The ADC function is to perform, for example, processing to automatically classify the defect images collected by the ADR function.
There are multiple types of circuit pattern structures formed on semiconductor wafers. Likewise, semiconductor wafer defects are various in type, occurrence position, and so on. As for the ADR function, it is important to capture and output a high-picture quality image with high defect visibility, circuit pattern visibility, and the like. Accordingly, in the related art, visibility enhancement is performed using an image processing technique with respect to a raw captured image that is a signal obtained from a detector of a review SEM and turned into an image.
By one method related thereto, the correspondence relationship between images different in image quality is pre-learned and an image of the other image quality is estimated based on the trained model when an image similar to one image quality is input. Machine learning or the like can be applied to the learning.
As an example of the related art related to the learning, JP-A-2018-137275 (Patent Document 1) describes a method for estimating a high-magnification image from a low-magnification image by pre-learning the relationship between the images captured at low and high magnifications.
In applying a method as described above for pre-learning the relationship between a captured image and an image of ideal image quality (also referred to as target image) with regard to the ADR function of a sample observation device, it is necessary to prepare the captured image (particularly plurality of captured images) and the target image for learning. However, it is difficult to prepare the image of ideal image quality in advance. For example, an actual captured image has noise, and it is difficult to prepare a noise-free image of ideal image quality based on the captured image.
In addition, the image quality of the captured image changes depending on, for example, the imaging environment or sample state difference. Accordingly, in order to perform more accurate learning, it is necessary to prepare a plurality of captured images of various image qualities. However, this requires a lot of effort. In addition, when learning is performed using a captured image, a sample needs to be prepared and imaged in advance, which imposes a heavy burden on a user.
There is a need for a mechanism capable of responding to, for example, a case where it is difficult to prepare multiple captured images or an image of ideal image quality and a mechanism capable of acquiring images of various image qualities suitable for sample observation.
An object of the present invention is to provide a technique for reducing work such as capturing an actual image with regard to sample observation device technique.
A typical embodiment of the present invention has the following configuration. A sample observation device according to the embodiment includes an imaging device and a processor. The processor: stores design data on a sample in a storage resource; creates a first learning image as a plurality of input images; creates a second learning image as a target image; learns a model related to image quality conversion with the first and second learning images; acquires, as an observation image, a second captured image output by inputting a first captured image obtained by imaging the sample with the imaging device to the model in observing the sample; and creates at least one of the first and second learning images based on the design data.
According to the typical embodiment of the present invention, provided is a technique for reducing work such as capturing an actual image with regard to sample observation device technique. Tasks, configurations, effects, and so on other than those described above are shown in the forms for carrying out the invention.
Hereinafter, embodiments of the present invention will be described in detail with reference to the drawings. In the drawings, the same parts are designated by the same reference numerals in principle, and repeated description thereof will be omitted. In the embodiments and the drawings, the representation of each component may not represent the actual position, size, shape, range, and so on to facilitate understanding of the invention. In the description, in describing processing by a program, the program, a function, a processing unit, and so on may be mainly described, but the main hardware therefor is a processor or a controller, a device, a computer, a system, or the like configured by the processor or the like. The computer executes processing in accordance with a program read out on a memory with the processor while appropriately using resources such as the memory and a communication interface. As a result, a predetermined function, processing unit, and so on are realized. The processor is configured by, for example, a semiconductor device such as a CPU and a GPU. The processor is configured by a device or circuit capable of performing a predetermined operation. The processing can also be implemented by a dedicated circuit without being limited to software program processing. FPGA, ASIC, CPLD, and so on can be applied to the dedicated circuit. The program may be pre-installed as data in the target computer or may be installed after being distributed as data from a program source to the target computer. The program source may be a program distribution server on a communication network, a non-transient computer-readable storage medium (e.g. memory card), or the like. The program may be configured by a plurality of modules. A computer system may be configured by a plurality of devices. The computer system may be configured by a client server system, a cloud computing system, an IoT system, or the like. Various data and information are represented and implemented in a structure such as a table and a list, but the present invention is not limited thereto. Representations such as identification information, identifiers, IDs, names, and numbers are mutually replaceable.
Regarding a sample observation device, in image quality conversion based on machine learning (i.e. image estimation), preparing an image of target image quality is important for improving the performance of an image quality conversion engine (including learning model). In the embodiments, an image that matches user preference is used as a target image even in a case where an image that is difficult to realize with an actual captured image is a target image. In addition, in the embodiments, the performance of the image quality conversion engine is maintained even in a case where the image quality of an image fluctuates depending on the state of an observation sample and the like.
In the embodiments, a target image for learning (second learning image) is created based on a parameter in which a target image quality is designated by a user and design data. As a result, it is also possible to realize an image quality that is difficult to realize with an actual captured image and target image preparation is facilitated. In addition, in the embodiments, images of various image qualities (first learning images) are created based on design data. In the embodiments, those images are used as input images to optimize the model of the image quality conversion engine. In other words, a model parameter is set and adjusted to an appropriate value. As a result, robustness is improved against fluctuations in the image quality of an input image.
The sample observation device and method of the embodiments pre-create at least one of an image of a target image quality (second learning image) and input images of various image qualities (first learning images) based on sample design data and optimizes the model by learning. As a result, in observing a sample, a first captured image of the image quality obtained by actually imaging the sample is converted into a second captured image of ideal image quality by the model and the image is obtained as an observation image.
The sample observation device of the embodiments is a device for observing, for example, a circuit pattern or defect formed on a sample such as a semiconductor wafer. This sample observation device performs processing with reference to defect position information created and output by an inspection device. This sample observation device learns a model for estimating the second learning image, which is a target image of ideal image quality (image quality reflecting user preference), from the first learning images (plurality of input images), which are images captured by an imaging device or images created based on design data without imaging.
The sample observation device and method of the related art example are techniques for preparing multiple actually captured images and learning a model using the images as input and target images. On the other hand, the sample observation device and method of the embodiments are provided with a function of creating at least one of the first learning image and the second learning image based on design data. As a result, the work of imaging for learning can be reduced.
The sample observation device and so on of a first embodiment will be described with reference to
In the first embodiment, each of the first learning image, which is an input image, and the second learning image, which is a target image, is an image created based on design data and is not an actually captured image.
Hereinafter, a device for observing, for example, a semiconductor wafer defect using a semiconductor wafer as a sample will be described as an example of the sample observation device. This sample observation device includes an imaging device that images a sample based on defect coordinates indicated by defect position information from an inspection device. An example of using an SEM as an imaging device will be described below. The imaging device is not limited to an SEM and may be a non-SEM device such as an imaging device using charged particles such as ions.
It should be noted that regarding the image qualities of the first learning image and the second learning image, the image quality (i.e. image properties) is a concept including a picture quality and other properties (e.g. partial extraction of circuit pattern). The picture quality is a concept including, for example, image magnification, field of view range, image resolution, and S/N. In the relationship between the image quality of the first learning image and the image quality of the second learning image, the high-low relationship of, for example, picture quality is a relative definition. For example, the second learning image is higher in picture quality than the first learning image. In addition, image quality-defining conditions, parameters, and so on are applied not only in a case where the image is obtained by performing imaging with an imaging device but also in a case where the image is created and obtained by image processing or the like.
The sample observation device and method of the first embodiment include a design data input unit inputting sample circuit pattern layout design data, a first learning image creation unit creating (i.e. generating) a plurality of first learning images of the same layout (i.e. same region) from the design data by changing a first processing parameter in a plurality of ways, a second learning image creation unit creating (i.e. generating) a second learning image from the design data using a second processing parameter designated by a user in accordance with user preference, a learning unit learning a model for estimating and outputting the second learning image using the plurality of first learning images as an input (i.e. learning unit learning model using first and second learning images), and an estimation unit inputting a first captured image of the sample imaged by an imaging device to the model and obtaining a second captured image as an output. The first learning image creation unit changes a parameter value in a plurality of ways with regard to at least one of the elements of sample circuit pattern shading value, shape deformation, image resolution, image noise, and so on to create the plurality of first learning images of the same region from the design data. The second learning image creation unit creates the second learning image from the design data using a parameter designated by the user on a GUI as a parameter different from a parameter for the first learning image.
[1-1. Sample Observation Device]
The sample observation device 1 is a device or system that has an automatic defect review (ADR) function. In this example, defect position information 8 is created as a result of pre-inspecting a sample at the external inspection device 7. The defect position information 8 output and provided from the inspection device 7 is pre-stored in the storage medium device 4. The higher control device 3 reads out the defect position information 8 from the storage medium device 4 and refers to the defect position information 8 during defect observation-related ADR processing. The SEM 101 that is the imaging device 2 captures an image of a semiconductor wafer that is a sample 9. The sample observation device 1 obtains an observation image (particularly, plurality of images by ADR function) that is an image of ideal image quality reflecting user preference based on the image captured by the imaging device 2.
The manufacturing execution system (MES) 10 is a system that manages and executes a process for manufacturing a semiconductor device using the semiconductor wafer that is the sample 9. The MES 10 has design data 11 related to the sample 9. In this example, the design data 11 pre-acquired from the MES 10 is stored in the storage medium device 4. The higher control device 3 reads out the design data 11 from the storage medium device 4 and refers to the design data 11 during processing. The format of the design data 11 is not particularly limited insofar as the design data 11 is data representing a structure such as the circuit pattern of the sample 9.
The defect classification device 5 is a device or system that has an automatic defect classification (ADC) function. The defect classification device 5 performs ADC processing based on information or data that is the result of the defect observation processing by the sample observation device 1 using the ADR function and obtains a result in which defects (corresponding defect images) are classified. The defect classification device supplies the information or data that is the classification result to, for example, another network-connected device (not illustrated). It should be noted that the present invention is not limited to the configuration illustrated in
The higher control device 3 includes, for example, a control unit 102, a storage unit 103, an arithmetic unit 104, an external storage medium input-output unit 105 (i.e. input-output interface unit), a user interface control unit 106, and a network interface unit 107. These components are connected to a bus 114 and are capable of mutual communication, input, and output. It should be noted that although the example of
The control unit 102 corresponds to a controller that controls the entire sample observation device 1. The storage unit 103 stores various information and data including a program and is configured by a storage medium device including, for example, a magnetic disk, a semiconductor memory, or the like. The arithmetic unit 104 performs an operation in accordance with a program read out of the storage unit 103. The control unit 102 and the arithmetic unit 104 include a processor and a memory. The external storage medium input-output unit (i.e. input-output interface unit) 105 performs data input and output in relation to the external storage medium device 4.
The user interface control unit 106 is a part that provides and controls a user interface including a graphical user interface (GUI) for performing information and data input and output in relation to a user (i.e. operator). An input-output terminal 5 is connected to the user interface control unit 106. Another input or output device (e.g. display device) may be connected to the user interface control unit 106. The defect classification device 5, the inspection device 7, and so on are connected to the network interface unit 107 via a network (e.g. LAN). The network interface unit 107 is a part that has a communication interface controlling communication with an external device such as the defect classification device 5 via a network. A DB server or the like is another example of the external device.
A user inputs information (e.g. instruction or setting) to the sample observation device 1 (particularly, higher control device 3) using the input-output terminal 5 and confirms information output from the sample observation device 1. A PC or the like can be applied to the input-output terminal 5, and the input-output terminal 5 includes, for example, a keyboard, a mouse, and a display. The input-output terminal 5 may be a network-connected client computer. The user interface control unit 106 creates a GUI screen (described later) and displays the screen on the display device of the input-output terminal 5.
The arithmetic unit 104 is configured by, for example, a CPU, a ROM, and a RAM and operates in accordance with a program read out of the storage unit 103. The control unit 102 is configured by, for example, a hardware circuit or a CPU. In a case where the control unit 102 is configured by a CPU or the like, the control unit 102 also operates in accordance with the program read out of the storage unit 103. The control unit 102 realizes each function based on, for example, program processing. Data such as a program is stored in the storage unit 103 after being supplied from the storage medium device 4 via the external storage medium input-output unit 105. Alternatively, data such as a program may be stored in the storage unit 103 after being supplied from a network via the network interface unit 107.
The SEM 101 of the imaging device 2 includes, for example, a stage 109, an electron source 110, a detector 111, an electron lens (not illustrated), and a deflector 112. The stage 109 (i.e. sample table) is a stage on which the semiconductor wafer that is the sample 9 is placed, and the stage is movable at least horizontally. The electron source 110 is an electron source for irradiating the sample 9 with an electron beam. The electron lens (not illustrated) converges the electron beam on the sample 9 surface. The deflector 112 is a deflector for performing electron beam scanning on the sample 9. The detector 111 detects electrons and particles such as secondary and backscattered electrons generated from the sample 9. In other words, the detector 111 detects the state of the sample 9 surface as an image. In this example, a plurality of detectors are provided as the detector 111 as illustrated in the drawing.
The information (i.e. image signal) detected by the detector 111 of the SEM 101 is supplied to the bus 114 of the higher control device 3. The information is processed by, for example, the arithmetic unit 104. In this example, the higher control device 3 controls the stage 109 of the SEM 101, the deflector 112, the detector 111, and so on. It should be noted that a drive circuit or the like for driving, for example, the stage 109 is not illustrated. Observation processing is realized with respect to the sample 9 by the computer system that is the higher control device 3 processing the information (i.e. image) from the SEM 101.
This system may have the following form. The higher control device 3 is a server such as a cloud computing system, and the input-output terminal 5 operated by a user is a client computer. For example, in a case where a lot of computer resources are required for machine learning, machine learning processing may be performed in a server group such as a cloud computing system. A processing function may be shared between the server group and the client computer. The user operates the client computer, and the client computer transmits a request to the server. The server receives the request and performs processing in accordance with the request. For example, the server transmits data on a screen (e.g. web page) reflecting the result of the requested processing to the client computer as a response. The client computer receives the response data and displays the screen (e.g. web page) on a display device.
[1-2. Functional Blocks and Flows]
The learning image creation processing step S11 has a design data input unit 200, parameter designation 205 by GUI, a second learning image creation unit 220, and a first learning image creation unit 210 as functional blocks. The design data input unit 200 inputs design data 250 from the outside (e.g. MES 10) (e.g. reads the design data 11 from the storage medium device 4 of
In the model learning processing S12, a model 260 is trained such that the target image 252 that is the second learning image (estimated second learning image) is output no matter which of the plurality of input images 251 that are the first learning images (images of various image qualities) is input.
[1-3. Defect Position Information]
The sample observation device 1 of the first embodiment has an ADR function to automatically collect a high-definition image showing a defect part on the surface of the sample 9 based on such defect coordinates. However, the defect coordinates in the defect position information 8 from the inspection device 7 include an error. In other words, an error may occur between the defect coordinates in the coordinate system of the inspection device 7 and the defect coordinates in the coordinate system of the sample observation device 1. Examples of the cause of the error include imperfect alignment of the sample 9 on the stage 109.
Accordingly, the sample observation device 1 captures a low-magnification image with a wide field of view (i.e. image of relatively low picture quality, first image) under a first condition centering on the defect coordinates of the defect position information 8 and re-detects the defect part based on the image. Then, the sample observation device 1 estimates a high-magnification image with a narrow field of view (i.e. image of relatively high picture quality, second image) under a second condition regarding the re-detected defect part using a pre-trained model and acquires the image as an observation image.
The wafer 301 includes the plurality of regular dies 302. Accordingly, in a case where, for example, another die 302 adjacent to the die 302 that has a defect part is imaged, it is possible to acquire an image of a non-defective die that includes no defect part. In the defect detection processing in the sample observation device 1, for example, such a non-defective die image can be used as a reference image. Further, in the defect detection processing, for example, shading (example of feature quantity) comparison as a defect determination is performed between the inspection target image (observation image) and the reference image and a part different in shading can be detected as a defect part.
[1-4. Learning Phase 1]
In the learning phase S1, the processor acquires first learning images 404 by inputting data obtained by cutting out a part of the region of design data 400 and a first processing parameter 401 to the drawing engine 403. The first learning images 404 are a plurality of input images for learning. Here, this image is also indicated by the symbol f. i is 1 to M, and M is an image count. The plurality of first learning images are indicated as f={f1, f2, . . . , fi, . . . , fM}.
The first processing parameter 401 is a parameter (i.e. condition) for creating (i.e. generating) the first learning image 404. In the first embodiment, the first processing parameter 401 is a parameter preset in this system. The first processing parameter 401 is a parameter set for creating the plurality of first learning images of different image qualities by assuming a change in the image quality of the captured image attributable to the imaging environment or the state of the sample 9. The first processing parameter 401 is a parameter set that is set by changing a parameter value in a plurality of ways using the parameter of at least one of the elements of circuit pattern shading value, shape deformation, image resolution, image noise, and so on.
The design data 400 is layout data on the circuit pattern shape of the sample 9 to be observed. For example, in a case where the sample 9 is a semiconductor wafer or a semiconductor device, the design data 400 is a file in which edge information on the design shape of the semiconductor circuit pattern is written as coordinate data. In the related art, a format such as GDS-II and OASIS is known as such a design data file. By using the design data 400, it is possible to obtain pattern layout information without actually imaging the sample 9 with the SEM 101.
The drawing engine 403 creates both the first learning image 404 and a second learning image 407 as images based on the layout information on the pattern in the design data 400.
In the first embodiment (
[1-5. Design Data)]
The layout information on the pattern in the design data 400 will be described with reference to
An image 505 in
Information (region) 502 in
Examples of how an image is acquired in the drawing engine 403 include drawing in order from the lower layer based on the pattern layout information acquired from the design data 400 and a processing parameter. The drawing engine 403 trims the region to be drawn (e.g. region 501) from the design data 500 and draws a pattern-less region (e.g. region 506) based on the processing parameter (first processing parameter 401). Next, the drawing engine 403 draws the region 503 of the lower layer pattern and, finally, draws the region 504 of the upper layer pattern to obtain an image such as the information 502. The first learning image 404 of
[1-6. Learning Phase 2]
Returning to
Next, the processor obtains estimated second learning images 406 as an output by estimation by inputting the first learning images 404, which are a plurality of input images, to the image quality conversion engine 405. The estimated second learning image 406 is an image estimated by a model. Here, this image is also indicated by the symbol g′. j is 1 to N, and N is an image count. The plurality of estimated second learning images are indicated as g′={g′1, g′2, . . . , g′j, . . . , g′N}.
It should be noted that in the first embodiment, the number M of the first learning images (f) 404 and the number N of the estimated second learning images (g′) 406 are equal to each other, but the present invention is not limited thereto.
A deep learning model such as a model represented by a convolutional neural network (CNN) may be applied as the machine learning model of the image quality conversion engine 405.
Next, in an operation 408, the processor inputs the second learning image (g) 407 and the plurality of estimated second learning images (g′) 406 to calculate an estimation error 409 regarding the difference therebetween. The calculated estimation error 409 is fed back to the image quality conversion engine 405. The processor updates the parameter of the model of the image quality conversion engine 405 such that the estimation error 409 decreases.
The processor optimizes the image quality conversion engine 405 by repeating the learning processing as described above. The optimized image quality conversion engine 405 means that the accuracy in estimating the estimated second learning image 406 from the first learning image 404 is high. It should be noted that an image difference or an output by a CNN identifying the second learning image 407 and the estimated second learning image 406 may be used for the estimation error 409. As a modification example, in the latter case, the operation 408 is an operation by learning using a CNN.
A task of this processing is how to acquire the first learning image 404 and the second learning image 407. In order to optimize the image quality conversion engine 405 to be robust against a change in the image quality of a captured image attributable to the state of the sample 9 or imaging condition difference, it is necessary to ensure a variation in the image quality of the first learning image 404. In this regard, in this processing, a change in image quality that may occur is assumed, the first processing parameter 401 is changed in a plurality of ways, and the first learning image 404 is created from the design data 400. As a result, it is possible to ensure a variation in the image quality of the first learning image 404.
In addition, in order to optimize the image quality conversion engine 405 so as to be capable of outputting an image of an image quality reflecting user preference, it is necessary to use an image of an image quality reflecting user preference as the second learning image 407 that is a target image. However, in a case where it is difficult to realize an image quality that matches user preference (i.e. image quality suitable for observation) with an image obtained by imaging the sample 9, it is difficult to prepare such a target image. In this regard, in this processing, the second learning image 407 is created by inputting the design data 400 and the second processing parameter 402 to the drawing engine 403. As a result, an image of an image quality that is difficult to realize with a captured image can also be acquired as the second learning image 407. In addition, in this processing, both the first learning image 404 and the second learning image 407 are created based on the design data 400. Accordingly, in the first embodiment, it is basically unnecessary to prepare and image the sample 9 in advance and the image quality conversion engine 405 can be optimized by learning.
It should be noted that in the sample observation device 1 of the first embodiment, it is unnecessary to use an image captured by the SEM 101 for the learning processing, but there is no limitation on using an image captured by the SEM 101 in the learning or sample observation processing. For example, as a modification example, some captured images may be added and used as an auxiliary in the learning processing.
It should be noted that the first processing parameter 401 is pre-designed as a parameter reflecting a fluctuation that may occur in a target process. This target process is the manufacturing process of a manufacturing process corresponding to the type of the target sample 9. The fluctuation is a fluctuation in environment, state, or condition related to an image quality (e.g. resolution, pattern shape, noise, and so on).
In the first embodiment, the first processing parameter 401 related to the first learning image 404 is pre-designed in this system, but the present invention is not limited thereto. In a modification example, the first processing parameter as well as the second processing parameter may allow variable setting by a user on a GUI screen. For example, a parameter set or the like to be used as the first processing parameter may allow selection from candidates and setting. In particular, on the GUI screen in a modification example, a fluctuation range (or statistical value of dispersion or the like) may be settable for each employed parameter regarding the first processing parameter for ensuring an image quality variation. As a result, a user can make trials and adjustments by variable first processing parameter setting while taking the trade-off between the processing time and accuracy into consideration.
[1-7. Effect, and the Like]
As described above, according to the sample observation device and method of the first embodiment, it is possible to reduce work such as capturing an actual image. In the first embodiment, the first learning image and the second learning image can be created using the design data without using an actual captured image. As a result, it is unnecessary to prepare and image a sample prior to sample observation and it is possible to optimize the model of the image quality conversion engine offline, that is, without imaging. Accordingly, for example, learning can be performed at design data completion and the first captured image can be captured and the second captured image can be estimated as soon as a semiconductor wafer as a target sample is completed. In other words, the efficiency of the entire work can be improved.
According to the first embodiment, it is possible to optimize the image quality conversion engine capable of conversion into an image quality matching user preference. In addition, the image quality conversion engine can be optimized to be highly robust against a change in sample state or imaging conditions. As a result, using this image quality conversion engine in observing a sample, it is possible to stably and highly accurately output an image of an image quality matching user preference as an observation image.
According to the first embodiment, multiple images can be prepared even in a case where deep learning is used as machine learning. According to the first embodiment, a target image corresponding to user preference can be created. According to the first embodiment, an input image corresponding to various imaging conditions is created from design data, a target image is created by a user performing parameter designation, and thus the above effects can be achieved.
The sample observation device and so on according to a second embodiment will be described with reference to
In the second embodiment, a task of this processing is how to acquire an ideal target image matching user preference. It is not easy to image a sample while changing the imaging conditions of an imaging device and figure out the imaging conditions of an image matching user preference. Further, under any imaging conditions, it may be impossible to obtain an image of ideal image quality anticipated by a user. In other words, not all evaluation values such as image resolution, signal/noise ratio (S/N), and contrast can be as desired as electron microscopic imaging has its own physical limitations.
In this regard, in the second embodiment, design data is input to a drawing engine, drawing is performed using the second processing parameter reflecting user preference, and a target image of ideal image quality can be created as a result. The ideal target image created from the design data is used as the second learning image.
As for the configuration of the learning phase S1 in the second embodiment that is different from
[2-1. Learning Phase]
In
It should be noted that in the second embodiment, the imaging 612 by the imaging device 2 is not limited to an electron microscope such as the SEM 101 and an optical microscope, an ultrasonic inspection device, or the like may be used.
However, in a case where a plurality of images of various image qualities assuming a change in image quality that may occur are acquired as the first learning images 604 by the imaging 612, in the related art, a plurality of samples corresponding thereto are necessary, which causes a heavy work burden on a user. Accordingly, in the second embodiment, the processor may create and acquire a plurality of input images of variously changed image qualities as the first learning images by applying image processing in which a parameter value is variously changed with respect to one first learning image 604 obtained by the imaging 612.
Next, the processor acquires a second learning image 607 (g) by inputting design data 600 and a second processing parameter 602, which is a processing parameter reflecting user preference, to a drawing engine 603. The drawing engine 603 corresponds to the second learning image creation unit.
[2-2. Effect, and the Like]
As described above, according to the second embodiment, the second learning image, which is a target image, is created based on design data, and thus the work of imaging for target image creation can be reduced.
In addition, other effects include the following. In the first embodiment described above (
The sample observation device and so on according to a third embodiment will be described with reference to
The first learning image creation unit changes the first processing parameter of at least one of the elements of circuit pattern shading value, shape deformation, image resolution, image noise, and so on in a plurality of ways to create a plurality of input images of the same region as the first learning images from the design data.
In the third embodiment, a task of this processing is to use images of various image qualities as the first learning images. If only an image of single image quality is used as the first learning image, it is difficult to ensure robustness against a change in image quality attributable to the sample state or imaging condition difference, and thus the versatility of the image quality conversion engine is low. In the third embodiment, in creating the first learning image from the same design data, the first processing parameter is changed assuming a change in image quality that may occur, and thus it is possible to ensure a variation in the image quality of the first learning image.
As for the configuration of the learning phase S1 in the third embodiment that is different from
[3-1. Learning Phase]
In
It should be noted that the image acquired by the imaging 712 may lack visibility due to the effect of insufficient contrast, noise, or the like. Accordingly, in the third embodiment, the processor may apply image processing such as contrast correction and noise removal to the image obtained by the imaging 712 and use the image as the second learning image 707. In addition, the processor of the sample observation device 1 may use an image acquired from another external device as the second learning image 707.
Next, the processor acquires first learning images 704 (f), which are a plurality of input images, by inputting design data 700 and a first processing parameter 701 to a drawing engine 703.
It should be noted that the first processing parameter 701 is a parameter set for acquiring the first learning images 704, which are a plurality of input images of different image qualities, by changing a parameter value in a plurality of ways regarding the parameter of at least one of the elements of circuit pattern shading value, shape deformation, image resolution, and image noise by assuming a change in the image quality of the captured image attributable to the imaging environment or the state of the sample 9.
In general, in a case where an image of satisfactory image quality is acquired by electron microscopic imaging, the time required for processing such as imaging increases. For example, the imaging takes a relatively long time as electron beam scanning, addition processing on a plurality of image frames, and so on are required. Accordingly, in that case, it is difficult to achieve picture quality and processing time at the same time and there is a trade-off relationship between picture quality and processing time.
In this regard, in this processing in the third embodiment, the imaging 712 of an image of satisfactory picture quality is performed in advance and the image is used for learning of an image quality conversion engine 705 as the second learning image 707. As a result, conversion is possible from an image of relatively poor picture quality to an image of relatively satisfactory picture quality although the imaging processing time is relatively short. As a result, it is possible to achieve picture quality and processing time at the same time. In other words, it is easy to adjust the balance between picture quality and processing time depending on the user.
In addition, in the case of a modification example in which an image acquired by another device is used as the second learning image 707, in the sample observation phase S2, the image quality of the image captured by the sample observation device 1 can be converted into the image quality of the image acquired by the other device.
It should be noted that in
[3-2. Effect, and the Like]
As described above, according to the third embodiment, the first learning images, which are a plurality of input images, are created based on design data, and thus the work of imaging for creating a plurality of input images can be reduced.
The sample observation device and so on according to a fourth embodiment will be described with reference to
[4-1. Learning Phase]
In
It should be noted that in a case where the function of the image alignment 802 as described above is provided, in the case of the second embodiment of
In the fourth embodiment, the first learning image and the second learning image can be aligned by this processing and there is no misalignment or misalignment is reduced regarding the first learning image and the second learning image. As a result, it is possible to optimize the model without taking misalignment between the first learning image and the second learning image into consideration and the stability of the optimization processing is improved.
The sample observation device and so on according to a fifth embodiment will be described with reference to the drawings starting from
[5-1. Learning Phase]
Each image in the plurality of first learning images 904 (f1 to fM) acquired by the drawing engine 903 can be treated as a two-dimensional array. For example, in a certain rectangular image, the screen horizontal direction (x direction) can be the first dimension, the screen vertical direction (y direction) can be the second dimension, and then the position of each pixel in the image region can be designated in the two-dimensional array. Further, as for the configuration of the first learning images 904 (images of two-dimensional array), which are a plurality of input images, each image may be expanded into a three-dimensional array by connecting the directions corresponding to the image count V as three-dimensional directions. For example, the first image group f1 (=f1−1 to f1−V) is configured by one three-dimensional array.
The plurality of first learning images 904 (f1 to fM) can be identified and specified as follows corresponding to the image count M and the image count V. i is used as a variable (index) for identifying a plurality in the direction corresponding to the image count M, and m is used as a variable (index) for identifying a plurality in the direction corresponding to the image count V. Of the first learning images 904, a certain image can be specified by designating (i, m). For example, the image can be specified as the mth image fi−m of the ith image group fi={fi−1, . . . , fi−V}, which is the first learning images 904.
In addition, each image in the plurality of second learning images 907 {g−1, . . . , g−U} acquired by the drawing engine 903 can be treated as a two-dimensional array. Further, as for the configuration of the second learning images 907 (images of two-dimensional array), which are a plurality of target images, each image may be expanded into a three-dimensional array by connecting the directions corresponding to the image count U as three-dimensional directions. Of the second learning images 907 (g−1 to g−U), which are a plurality of target images, one image can be specified as, for example, an image g-k using a variable (referred to as k) for identifying a plurality in a direction corresponding to the image count U.
Next, by inputting each image (e.g. image group f1) of the first learning image 904 to an image quality conversion engine 905, the corresponding image group (e.g. g′1) is obtained as an estimated second learning image 906. Regarding this estimated second learning image 906 as well, the processor may perform division into a plurality of elements in a direction (e.g. three-dimensional direction) corresponding to an image count (referred to as W) different from the image count N to create, for example, the image group g′1 {g′1−1, . . . , g′1−W}. These estimated second learning images 906 may also be configured by three-dimensional arrays.
The plurality of second learning images 906 (g′1 to g′N) can be, for example, identified and specified as follows corresponding to the image count N and the image count W. j is used as a variable (index) in the direction corresponding to the image count N, and n is used as a variable (index) in the direction corresponding to the image count W. Of the plurality of second learning images 906, a certain image can be specified by designating (j, n). For example, the image can be specified as the nth image g′j−n of the jth image group g′j={g′j−1, . . . , g′j−W}, which is the second learning images 906.
It should be noted that in the fifth embodiment, in the example of
In addition, in the fifth embodiment, with respect to the first to fourth embodiments described above, the model of the image quality conversion engine 905 is changed to a configuration inputting and outputting a multidimensional image corresponding to the image counts (V, W) in the three-dimensional direction of the first learning image 904 and the estimated second learning image 906. For example, in a case where a CNN is applied to the image quality conversion engine 905, simply the input and output layers in the CNN may be changed to the configuration corresponding to the image counts (V, W) in the three-dimensional direction.
In the fifth embodiment, it is possible to apply, for example, a plurality of images of a plurality of types that can be acquired by the plurality of detectors 111 (
It should be noted that as for the plurality of images of input and output with respect to the model (first and estimated second learning images) in the configurations of
[5-2. Detector]
The four detectors are disposed so as to be capable of selectively detecting electrons that have specific emission angles (elevation and azimuth angles). For example, the detector at the position P1 is capable of detecting electrons emitted from sample 9 along the positive direction of the y axis. The detector at the position P4 is capable of detecting electrons emitted from the sample 9 along the positive direction of the x axis. The detector at the position P5 is capable of detecting mainly electrons emitted from the sample 9 in the z-axis direction.
As described above, with the configuration in which the plurality of detectors are disposed at the plurality of positions along the different axes, it is possible to acquire an image with contrast as if light was emitted from a facing direction with respect to each detector. Accordingly, more detailed defect observation is possible. The configuration of the detector 111 is not limited thereto, and different numbers, positions, orientations, and so on may be configured.
[5-3. First Learning Image Created by Drawing Engine]
For example, a secondary electron image or a backscattered electron image can be obtained depending on the type of the electron ejected from the sample 9. Secondary electron is also abbreviated as SE. Backscattered electron is also abbreviated as BSE.
In addition, depending on the configuration of the SEM 101, it is possible to obtain a tilt image obtained by observing a measurement target from any inclination direction. The example of an image 1190 in
In the fifth embodiment, such images of a plurality of types are used as the first learning images 904, and thus more information than in a configuration in which one image is used as the first learning image can be input to the model of the image quality conversion engine 905. Accordingly, it is possible to improve the performance of the model of the image quality conversion engine 905, particularly robustness allowing a response to various image qualities. The plurality of estimated second learning images 906 with different image qualities can be obtained as outputs of the model of the image quality conversion engine 905.
In addition, in the case of a configuration in which a plurality of different image quality conversion engines are prepared for each output image in order to use a plurality of images of different image qualities as outputs of the image quality conversion engine 905, it is necessary to optimize the plurality of image quality conversion engines. In addition, in using the image quality conversion engines, processing time increases as it is necessary to input a captured image into each image quality conversion engine and process the image. On the other hand, in the fifth embodiment, simply one image quality conversion engine 905 is sufficient in order to use a plurality of images of different image qualities (estimated second learning images 906) as outputs of the image quality conversion engine 905. In the fifth embodiment, the second learning image 907 is created based on the same design data 900, and thus the image quality conversion engine 905 is capable of creating each output image (estimated second learning image 906) from the same feature quantity. In this processing, using one image quality conversion engine 905 capable of outputting a plurality of images, processing during optimization and processing during image quality conversion are expedited and efficiency and convenience are improved.
An image 1110 of
An image 1120 in
An image 1130 in
An image 1140 in
An image 1150 in
An image 1160 in
An image 1170 in
In
The image 1190 in
An image 1200 in
The processor estimates and creates such a tilt image from, for example, two-dimensional pattern layout data in design data. At this time, examples of how the tilt image is estimated and created include inputting a pattern height design value to generate a pseudo-pattern three-dimensional shape and estimate the image observed from the tilt direction.
As described above, the processor of the sample observation device 1 creates images of various different image qualities as variations and uses the images as the first learning images 904, which are a plurality of input images, by taking into consideration image quality fluctuations assumed in imaging the sample 9 due to the effect of the state of the sample 9, imaging conditions, or the like such as charging and pattern shape fluctuation. As a result, it possible to optimize the model of the image quality conversion engine 905 to be robust against the image quality fluctuation of an input image. In addition, the model can be optimized with high accuracy by setting the detector 111 of the imaging device 2 (e.g. detector used among detectors) in accordance with conditions in observing the sample 9 or by making a tilt image.
[5-4. Second Learning Image Created by Drawing Engine]
Next,
In
An image 1320 in
The images from an image 1330 in
The image 1340 in
In
In a case where image processing is applied to a captured image, correct information extraction may be impossible due to the effect of image noise or the like or a parameter may need to be adjusted in accordance with the application process. In the fifth embodiment, when a post-image processing application image is acquired from design data, noise or the like has no effect, and thus information can be acquired with ease. In the fifth embodiment, an image to which image processing for acquiring information from an image to be obtained by imaging is applied is learned as the second learning image 907 to optimize the model of the image quality conversion engine 905. As a result, it is possible to use the image quality conversion engine 905 instead of image processing.
It should be noted that although the edge images in this example are a plurality of direction-specific edge images in the two directions of x and y, the present invention is not limited thereto and similar application is possible regarding another direction (e.g. in-plane diagonal direction) as well.
<Sample Observation Phase>
An example of the sample observation phase S2 of
Next, in step S204, the processor moves the stage 109 such that the observation target region on the sample 9 is included in the imaging field of view. In other words, the processor positions the imaging optical system in the observation target region. The processing of steps S204 to S207 is loop processing repeated for each observation target region (e.g. defect position indicated by defect position information 8). Next, in step S205, the processor irradiates the sample 9 with an electron beam under the control of the SEM 101 and acquires the first captured image 253 (F) of the observation target region by detecting, for example, secondary or backscattered electrons with the detector 111 and performing conversions into an image.
Next, in step S206, the processor acquires a second captured image 254 (G′) by estimation as an output by inputting the first captured image 253 (F) to the image quality conversion engine 405 (model 260 of estimation unit 240 of
Then, in step S207, the processor may apply image processing corresponding to the purpose of observation to the second captured image 254. Examples of this image processing application include dimension measurement, alignment with design data, and defect detection and identification. Each example will be described later. It should be noted that such image processing may be performed by a device other than the sample observation device 1 (e.g. defect classification device 5 of
<A. Dimension Measurement>
An example of the dimension measurement processing as an example of the image processing in step S207 is as follows. FIG. 16 illustrates the example of the dimension measurement processing. In this dimension measurement, the dimension of the circuit pattern of the sample 9 is measured using the second captured image 254 (F′). The processor of the higher control device 3 uses an image quality conversion engine 1601 pre-optimized using an edge image (
Next, the processor performs dimension measurement processing 1603 with respect to the image 1602. In this dimension measurement processing 1603, the processor performs pattern dimension measurement by inter-edge distance measurement. The processor obtains an image 1604, which is the result of the dimension measurement processing 1603. In the examples of the images 1602 and 1604, lateral width measurement is performed for each inter-edge region. The examples include a breadth (X) 1606 of an inter-edge region 1605.
Further, the edge image as described above is effective for two-dimensional pattern shape evaluation based on a pattern contour line as well as a one-dimensional pattern dimension represented by the line width and hole diameter described above. For example, in a lithography process in semiconductor manufacturing, an optical proximity effect may lead to two-dimensional pattern shape deformation. Examples of the shape deformation include a rounded corner portion and an undulating pattern.
In performing pattern dimension and shape measurement and evaluation from an image, it is necessary to specify a pattern edge position with as high accuracy as possible by image processing. However, an image obtained by imaging also includes information other than pattern information such as noise. Accordingly, in order to specify an edge position with high accuracy, it is necessary in the related art to manually adjust an image processing parameter. On the other hand, in this processing, the image quality conversion engine (model) pre-optimized by learning converts a captured image into an edge image, and thus an edge position can be specified with high accuracy without manual image processing parameter adjustment. In the model learning, learning and optimization are performed using images of various image qualities in which edges, noise, and so on are taken into consideration as input-output images. Accordingly, it is possible to perform high-accuracy dimension measurement using a suitable edge image (second captured image 254) as described above.
<B. Alignment with Design Data>
An example of the processing of alignment with design data as an example of the image processing in step S207 is as follows. In an electron microscope such as the SEM 101, it is necessary to estimate and correct (i.e. address) an imaging position deviation amount. An electron beam irradiation position needs to be moved in order to move the field of view of the electron microscope. There are two methods therefor, one is a stage shift by which a sample-transporting stage is moved, the other is an image shift by which a deflector changes the trajectory of an electron beam, and each entails a stop position error.
As a method for imaging position deviation amount estimation, it is conceivable to perform alignment (i.e. matching) between a captured image and design data (region therein). Meanwhile, in a case where the image quality of the captured image is poor, the alignment itself may fail. Accordingly, in the embodiment, the imaging position of the first captured image 253 is specified by performing alignment between design data (region therein) and the second captured image 254, which is an output when the captured image (first captured image 253) is input to the image quality conversion engine (model 270). Several image conversion methods are conceivable as methods effective for the alignment. For example, in one method, an image higher in picture quality than the first captured image 253 is estimated as the second captured image 254. As a result, an improvement in alignment success rate can be anticipated. In addition, in another method, it is conceivable to estimate a direction-specific edge image as the second captured image 254.
Next, the processor draws the region of the sample 9 in design data 1704 with a drawing engine 1708 and creates an edge image (image group) 1705 for each layer and edge direction. The edge image (image group) 1705 is an edge image (design image) created from the design data 1704. Similarly to the edge image 1703, examples thereof include images in the upper layer x and y directions and the lower layer x and y directions.
Next, the processor calculates 1706 each correlation map between the edge image 1705 created from the design data 1704 and the edge image 1703 created from the captured image 1700. In this correlation map calculation 1706, the processor creates a correlation map for each set of images corresponding in layer and direction with each image of the edge image 1703 and each image of the edge image 1705. As the plurality of correlation maps, for example, correlation maps in the upper layer x and y directions and the lower layer x and y directions can be obtained. Next, the processor calculates and obtains a final correlation map 1707 by combining the plurality of correlation maps into one by performing weighted addition or the like.
In this final correlation map 1707, the position of maximum correlation value is the position of alignment (matching) between the captured image (corresponding observation target region) and the design data (corresponding region therein). In the weighted addition, for example, the weight is inversely proportional to the amount of the edge in the image. As a result, correct alignment can be anticipated without sacrificing the degree of matching of an image with a small edge amount.
As described above, the captured image and the design data can be aligned with high accuracy using the pattern shape-indicating edge image. However, the captured image also includes information other than pattern information as described above, and thus image processing parameter adjustment needs to be performed in order to highly accurately specify an edge position from the captured image by image processing. In this processing, the pre-optimized image quality conversion engine converts the first captured image into an edge image, and thus an edge position can be specified with high accuracy without manual parameter adjustment. As a result, the alignment between the captured image and the design data can be realized with high accuracy.
<C. Defect Detection and Defect Type Identification>
An example of the processing of defect detection and defect type identification (classification) as an example of the image processing in step S207 is as follows.
Next, the processor acquires a cut-out image 1805 by performing processing 1804 to cut out the same region as the image 1801 obtained based on the captured image from the alignment result image 1803 based on the design data.
Next, the processor performs defect position specifying processing 1806 by calculating the difference between the cut-out image 1805 and the image 1801 obtained based on the captured image to obtain an image (defect image) 1807 including a specified defect position as the result thereof.
Subsequently, the processor may further apply processing 1808 (i.e. classification processing) to perform defect type identification using the defect image 1807. As a method for the defect identification, a feature quantity may be calculated from an image by image processing and identification may be performed based on the feature quantity or identification may be performed using a pre-optimized CNN for defect identification.
In general, the reference image and the first captured image obtained by imaging include noise, and thus it is necessary in the related art to perform manual image processing parameter adjustment in order to perform defect detection and identification with high accuracy. On the other hand, in this processing, the image quality conversion engine converts the first captured image into the high-S/N image 1801 (second captured image 254), and thus the effect of noise can be reduced. In addition, the reference image 1800 created from the design data is noise-free, and thus it is possible to specify a defect position without taking reference image noise into consideration. In this manner, it is possible to reduce the effect of noise in the first captured and reference images, which is a hindrance in specifying a defect position.
<GUI>
Next, a GUI screen example that can be similarly applied to each of the embodiments will be described. It should be noted that the configurations of the first to third embodiments and so on can be combined and, in the combined configurations, a suitable configuration to be appropriately used by a user can be selected from the configurations of the first to third embodiments and so on. The user can select, for example, a model in accordance with the type of sample observation or the like.
In addition, the lower table is provided with a column in which the user can set an acquisition method and a processing parameter regarding the first learning image and the second learning image related to the learning phase S1 described above. In a column 1901, a first learning image acquisition method can be set by selection from the options of “imaging” and “design data use”. In a column 1902, a second learning image acquisition method can be set by selection from the options of “imaging” and “design data use”. In the example of
In a case where “design data use” is selected in the second learning image acquisition method, in the corresponding processing parameter column, the user can designate and set a processing parameter to be used in the engine. In a column 1903, as an example of the parameter, the values of parameters such as pattern shading value, image resolution, and circuit pattern shape deformation can be designated.
In addition, in a column 1904, the user can select an image quality of an ideal image from the options. The image quality of the ideal image (target image, second learning image) can be selected from, for example, an ideal SEM image, an edge image, a tilt image, and the like. In a case where a preview button 1905 is pressed after the image quality of the ideal image is selected, a preview image of the selected image quality can be confirmed on, for example, the screen of
In the screen example of
Although single design data (region of sample 9) and an image created corresponding thereto are displayed in this example, similarly, an image can be displayed by designating another region with an image ID or predetermined operation. In a case where an SEM image is selected as an ideal image in the column 1904 of
Although the present invention has been specifically described above based on the embodiments, the present invention is not limited to the embodiments and can be variously modified without departing from the gist.
Number | Date | Country | Kind |
---|---|---|---|
2021-116563 | Jul 2021 | JP | national |