The description herein relates to lithographic apparatuses and processes, and more particularly to a tool to determine stochastic variations in printed patterns (e.g., in a mask or resist layer on a wafer) which can be used to detect defects (e.g., on a mask or a wafer) and optimize a patterning process such as mask optimization and source optimization.
A lithographic apparatus is a machine that applies a desired pattern onto a target portion of a substrate. The lithographic apparatus can be used, for example, in the manufacture of integrated circuits (ICs). For example, an IC chip in a smart phone, can be as small as a person's thumbnail, and may include over 2 billion transistors. Making an IC is a complex and time-consuming process, with circuit components in different layers and including hundreds of individual steps. Errors in even one step have the potential to result in problems with the final IC and can cause device failure. High process yield and high wafer throughput can be impacted by the presence of defects, especially if operator intervention is required for reviewing the defects. Inspection tools (such as an optical or electron microscope (SEM)) are used in identification of the defects to help in maintaining high yield and low cost.
In some embodiments, there is provided a non-transitory computer-readable media comprising instructions that, when executed by a computer, cause the computer to execute a method for training a machine learning model to determine a source of error contribution to multiple features of a pattern printed on a substrate. The method includes: obtaining training data having multiple datasets, wherein each dataset has error contribution values representative of an error contribution from one of multiple sources to the features, and wherein each dataset is associated with an actual classification that identifies a source of the error contribution of the corresponding dataset; and training, based on the training data, a machine learning model to predict a classification of a reference dataset of the datasets such that a cost function that determines a difference between the predicted classification and the actual classification of the reference dataset is reduced.
In some embodiments, there is provided a non-transitory computer-readable media comprising instructions that, when executed by a computer, cause the computer to execute a method for determining a source of error contribution to multiple features of a pattern printed on a substrate. The method includes: inputting, to a machine learning model, a specified dataset having error contribution values representative of an error contribution from one of multiple sources to the features; and executing the machine learning model to determine a classification associated with the specified dataset, wherein the classification identifies a specified source of the multiple sources as the source of error contribution for the error contribution values in the specified dataset.
Furthermore, in some embodiments, there is provided a method for training a machine learning model to determine a source of error contribution to multiple features of a pattern printed on a substrate. The method includes: obtaining training data having multiple datasets, wherein each dataset has error contribution values representative of an error contribution from one of multiple sources to the features, and wherein each dataset is associated with an actual classification that identifies a source of the error contribution of the corresponding dataset; and training, based on the training data, a machine learning model to predict a classification of a reference dataset of the datasets such that a cost function that determines a difference between the predicted classification and the actual classification of the reference dataset is reduced.
Furthermore, in some embodiments, there is provided a method for determining a source of error contribution to multiple features of a pattern printed on a substrate. The method includes: inputting, to a machine learning model, a specified dataset having error contribution values representative of an error contribution from one of multiple sources to the features; and executing the machine learning model to determine a classification associated with the specified dataset, wherein the classification identifies a specified source of the multiple sources as the source of error contribution for the error contribution values in the specified dataset.
Furthermore, in some embodiments, there is provided an apparatus for training a machine learning model to determine a source of error contribution to multiple features of a pattern printed on a substrate. The apparatus includes a memory storing a set of instructions; and at least one processor configured to execute the set of instructions to cause the apparatus to perform a method of: obtaining training data having multiple datasets, wherein each dataset has error contribution values representative of an error contribution from one of multiple sources to the features, and wherein each dataset is associated with an actual classification that identifies a source of the error contribution of the corresponding dataset; and training, based on the training data, a machine learning model to predict a classification of a reference dataset of the datasets such that a cost function that determines a difference between the predicted classification and the actual classification of the reference dataset is reduced.
Furthermore, in some embodiments, there is provided a non-transitory computer-readable media comprising instructions that, when executed by a computer, cause the computer to execute a method for training a machine learning model to determine error contributions to a feature of a pattern printed on a substrate. The method includes obtaining training data having multiple datasets, wherein the datasets include a first dataset having (a) a first image data of one or more features of a pattern to be printed on a substrate and (b) a first error contribution data comprising error contributions from multiple sources to the one or more features; and training, based on the training data, a machine learning model to predict error contribution data for the first dataset such that a cost function that is indicative of a difference between the predicted error contribution data and the first error contribution data is reduced.
Furthermore, in some embodiments, there is provided a non-transitory computer-readable media comprising instructions that, when executed by a computer, cause the computer to execute a method for determining error contribution data comprising error contributions from multiple sources to a feature of a pattern to printed on a substrate. The method includes receiving image data of a set of features of a specified pattern to be printed on a first substrate; inputting the image data to a machine learning model; and executing the machine learning model to determine error contribution data comprising error contributions from multiple sources to the set of features.
Furthermore, in some embodiments, there is provided a method for training a machine learning model to determine error contributions to a feature of a pattern printed on a substrate. The method includes obtaining training data having multiple datasets, wherein the datasets include a first dataset having (a) a first image data of one or more features of a pattern to be printed on a substrate and (b) a first error contribution data comprising error contributions from multiple sources to the one or more features; and training, based on the training data, a machine learning model to predict error contribution data for the first dataset such that a cost function that is indicative of a difference between the predicted error contribution data and the first error contribution data is reduced.
Furthermore, in some embodiments, there is provided a method for determining error contribution data comprising error contributions from multiple sources to a feature of a pattern to printed on a substrate. The method includes receiving image data of a set of features of a specified pattern to be printed on a first substrate; inputting the image data to a machine learning model; and executing the machine learning model to determine error contribution data comprising error contributions from multiple sources to the set of features.
Furthermore, in some embodiments, there is provided an apparatus for training a machine learning model to determine error contributions to a feature of a pattern printed on a substrate. The apparatus includes a memory storing a set of instructions; and at least one processor configured to execute the set of instructions to cause the apparatus to perform a method of: obtaining training data having multiple datasets, wherein the datasets include a first dataset having (a) a first image data of one or more features of a pattern to be printed on a substrate and (b) a first error contribution data comprising error contributions from multiple sources to the one or more features; and training, based on the training data, a machine learning model to predict error contribution data for the first dataset such that a cost function that is indicative of a difference between the predicted error contribution data and the first error contribution data is reduced.
Furthermore, in some embodiments, there is provided an apparatus for determining error contribution data comprising error contributions from multiple sources to a feature of a pattern to printed on a substrate. The apparatus includes a memory storing a set of instructions; and at least one processor configured to execute the set of instructions to cause the apparatus to perform a method of: receiving image data of a set of features of a specified pattern to be printed on a first substrate; inputting the image data to a machine learning model; and executing the machine learning model to determine error contribution data comprising error contributions from multiple sources to the set of features.
Furthermore, in some embodiments, there is provided a computer program product comprising a non-transitory computer readable medium having instructions recorded thereon, the instructions when executed by a computer system implementing the aforementioned methods.
Embodiments will now be described, by way of example only, with reference to the accompanying drawings in which:
Embodiments will now be described in detail with reference to the drawings, which are provided as illustrative examples so as to enable those skilled in the art to practice the embodiments. Notably, the figures and examples below are not meant to limit the scope to a single embodiment, but other embodiments are possible by way of interchange of some or all of the described or illustrated elements. Wherever convenient, the same reference numbers will be used throughout the drawings to refer to same or like parts. Where certain elements of these embodiments can be partially or fully implemented using known components, only those portions of such known components that are necessary for an understanding of the embodiments will be described, and detailed descriptions of other portions of such known components will be omitted so as not to obscure the description of the embodiments. In the present specification, an example showing a singular component should not be considered limiting; rather, the scope is intended to encompass other embodiments including a plurality of the same component, and vice-versa, unless explicitly stated otherwise herein. Moreover, applicants do not intend for any term in the specification or claims to be ascribed an uncommon or special meaning unless explicitly set forth as such. Further, the scope encompasses present and future known equivalents to the components referred to herein by way of illustration.
A lithographic apparatus is a machine that applies a desired pattern onto a target portion of a substrate. This process of transferring the desired pattern to the substrate is called a patterning process. The patterning process can include a patterning step to transfer a pattern from a patterning device (such as a mask) to the substrate. Also, there can then be one or more related pattern processing steps, such as resist development by a development apparatus, baking of the substrate using a bake tool, etching the pattern onto the substrate using an etch apparatus, etc. Various variations (e.g., stochastic variations, errors or noises due to any of inspection tool, mask or resist) can potentially limit lithography implementation for semiconductor high volume manufacturing (HVM). In order to characterize, understand and determine such variation, the industry needs a trustworthy method to measure such variation for variety of design patterns.
Some embodiments derive the stochastic variations using independent component analysis (ICA) methods. In the ICA method, measurement data of a number of features is obtained using a number of sensors. For example, three sets of measurement data are obtained using three different sensors and these three sets of measurement data are input as three signals to the ICA method, which decomposes the three input signals to obtain three output signals corresponding to error contributions from the three sources, such as a mask, resist and inspection tool such as a scanning electron microscope (SEM). However, in some cases, the ICA method may not be able to determine which output signal corresponds to an error contribution from which source, because the error contributions from various sources can be similar and therefore, the ICA method may not be able distinguish between them.
Some embodiments of the present disclosure identify an error contribution source for a given signal of error contribution values. A machine learning (ML) model is trained to distinguish between error contributions from various sources, and the trained ML model is used to determine a classification (e.g., an error contribution source) of a given signal.
While the ICA method may be used to determine the error contributions from multiple sources, the ICA method is characterized by the assumption that the error contributions are a linear mixture of errors from the different sources. In some embodiments, additional noise sources, e.g., noise from sources other than those determined using ICA, may be present, and if these noise sources are not removed when using the ICA method, the error contributions determined by the ICA method may not be accurate. Accordingly, the ICA method may be constrained by the above assumption. The embodiments of the present disclosure implement a ML model to determine the error contributions from a set of sources. For example, the ML model is trained using images of various features and error contribution measurements associated with those features to predict error contributions from the set of sources for a given feature. The error contribution measurements for training the ML model may be obtained using methods that are not constrained by the assumption that the error contributions are a linear mixture of errors from the set of sources. For the prediction, an image of a feature (e.g., contact hole) is provided as an input to the ML model, and the ML model predicts error contributions from various sources for the input feature. By training the ML model based on the error contribution data determined using methods that are not constrained by the assumption that the error contributions are a linear mixture of the set of sources, the error contribution data predicted by the ML model may not be affected by the presence of additional noise sources, thereby improving the accuracy in determination of the error contributions.
As a brief introduction,
Although specific reference may be made in this text to the manufacture of ICs, it should be explicitly understood that the description herein has many other possible applications. For example, it may be employed in the manufacture of integrated optical systems, guidance and detection patterns for magnetic domain memories, liquid-crystal display panels, thin-film magnetic heads, etc. The skilled artisan will appreciate that, in the context of such alternative applications, any use of the terms “reticle”, “wafer” or “die” in this text should be considered as interchangeable with the more general terms “mask”, “substrate” and “target portion”, respectively.
In the present document, the terms “radiation” and “beam” are used to encompass all types of electromagnetic radiation, including ultraviolet radiation (e.g. with a wavelength of 365, 248, 193, 157 or 126 nm) and EUV (extreme ultra-violet radiation, e.g. having a wavelength in the range 5-20 nm).
The term “optimizing” and “optimization” as used herein mean adjusting a lithographic projection apparatus such that results or processes of lithography have more desirable characteristics, such as higher accuracy of projection of design layouts on a substrate, larger process windows, etc.
Further, the lithographic projection apparatus may be of a type having two or more substrate tables (or two or more patterning device tables). In such “multiple stage” devices the additional tables may be used in parallel, or preparatory steps may be carried out on one or more tables while one or more other tables are being used for exposures. Twin stage lithographic projection apparatuses are described, for example, in U.S. Pat. No. 5,969,441, incorporated herein by reference.
The patterning device referred to above comprises or can form design layouts. The design layouts can be generated utilizing CAD (computer-aided design) programs, this process often being referred to as EDA (electronic design automation). Most CAD programs follow a set of predetermined design rules in order to create functional design layouts/patterning devices. These rules are set by processing and design limitations. For example, design rules define the space tolerance between circuit devices (such as gates, capacitors, etc.) or interconnect lines, so as to ensure that the circuit devices or lines do not interact with one another in an undesirable way. The design rule limitations are typically referred to as “critical dimensions” (CD). A critical dimension of a circuit can be defined as the smallest width of a line or hole or the smallest space between two lines or two holes. Thus, the CD determines the overall size and density of the designed circuit. Of course, one of the goals in integrated circuit fabrication is to faithfully reproduce the original circuit design on the substrate (via the patterning device).
The term “mask” or “patterning device” as employed in this text may be broadly interpreted as referring to a generic patterning device that can be used to endow an incoming radiation beam with a patterned cross-section, corresponding to a pattern that is to be created in a target portion of the substrate; the term “light valve” can also be used in this context. Besides the classic mask (transmissive or reflective; binary, phase-shifting, hybrid, etc.), examples of other such patterning devices include:
a programmable mirror array. An example of such a device is a matrix-addressable surface having a viscoelastic control layer and a reflective surface. The basic principle behind such an apparatus is that (for example) addressed areas of the reflective surface reflect incident radiation as diffracted radiation, whereas unaddressed areas reflect incident radiation as undiffracted radiation. Using an appropriate filter, the said undiffracted radiation can be filtered out of the reflected beam, leaving only the diffracted radiation behind; in this manner, the beam becomes patterned according to the addressing pattern of the matrix-addressable surface. The required matrix addressing can be performed using suitable electronic means. More information on such mirror arrays can be gleaned, for example, from U.S. Pat. Nos. 5,296,891 and 5,523,193, which are incorporated herein by reference.
a programmable LCD array. An example of such a construction is given in U.S. Pat. No. 5,229,872, which is incorporated herein by reference.
Major components are a radiation source 12A, which may be a deep-ultraviolet excimer laser source or other type of source including an extreme ultra violet (EUV) source (as discussed above, the lithographic projection apparatus itself need not have the radiation source), illumination optics which define the partial coherence (denoted as sigma) and which may include optics 14A, 16Aa and 16Ab that shape radiation from the source 12A; a patterning device 14A; and transmission optics 16Ac that project an image of the patterning device pattern onto a substrate plane 22A. An adjustable filter or aperture 20A at the pupil plane of the projection optics may restrict the range of beam angles that impinge on the substrate plane 22A, where the largest possible angle defines the numerical aperture of the projection optics NA=sin(Θmax).
In an optimization process of a system, a figure of merit of the system can be represented as a cost function. The optimization process boils down to a process of finding a set of parameters (design variables) of the system that minimizes the cost function. The cost function can have any suitable form depending on the goal of the optimization. For example, the cost function can be weighted root mean square (RMS) of deviations of certain characteristics (evaluation points) of the system with respect to the intended values (e.g., ideal values) of these characteristics; the cost function can also be the maximum of these deviations (i.e., worst deviation). The term “evaluation points” herein should be interpreted broadly to include any characteristics of the system. The design variables of the system can be confined to finite ranges or be interdependent due to practicalities of implementations of the system. In case of a lithographic projection apparatus, the constraints are often associated with physical properties and characteristics of the hardware such as tunable ranges, or patterning device manufacturability design rules, and the evaluation points can include physical points on a resist image on a substrate, as well as non-physical characteristics such as dose and focus.
In a lithographic projection apparatus, a source provides illumination (i.e. light); projection optics direct and shapes the illumination via a patterning device and onto a substrate. The term “projection optics” is broadly defined here to include any optical component that may alter the wavefront of the radiation beam. For example, projection optics may include at least some of the components 14A, 16Aa, 16Ab and 16Ac. An aerial image (AI) is the radiation intensity distribution at substrate level. A resist layer on the substrate is exposed and the aerial image is transferred to the resist layer as a latent “resist image” (RI) therein. The resist image (RI) can be defined as a spatial distribution of solubility of the resist in the resist layer. A resist model can be used to calculate the resist image from the aerial image, an example of which can be found in commonly assigned U.S. patent application Ser. No. 12/315,849, disclosure of which is hereby incorporated by reference in its entirety. The resist model is related only to properties of the resist layer (e.g., effects of chemical processes which occur during exposure, PEB and development). Optical properties of the lithographic projection apparatus (e.g., properties of the source, the patterning device and the projection optics) dictate the aerial image. Since the patterning device used in the lithographic projection apparatus can be changed, it is desirable to separate the optical properties of the patterning device from the optical properties of the rest of the lithographic projection apparatus including at least the source and the projection optics.
An exemplary flow chart for simulating lithography in a lithographic projection apparatus is illustrated in
More specifically, it is noted that the source model 31 can represent the optical characteristics of the source that include, but are not limited to, NA-sigma (a) settings as well as any particular illumination source shape (e.g. off-axis radiation sources such as annular, quadrupole, and dipole, etc.). The projection optics model 32 can represent the optical characteristics of the of the projection optics that include aberration, distortion, refractive indexes, physical sizes, physical dimensions, etc. The design layout model 35 can also represent physical properties of a physical patterning device, as described, for example, in U.S. Pat. No. 7,587,704, which is incorporated by reference in its entirety. The objective of the simulation is to accurately predict, for example, edge placements, aerial image intensity slopes and CDs, which can then be compared against an intended design. The intended design is generally defined as a pre-OPC design layout which can be provided in a standardized digital file format such as GDSII or OASIS or other file format.
From this design layout, one or more portions may be identified, which are referred to as “clips”. In some embodiments, a set of clips is extracted, which represents the complicated patterns in the design layout (typically about 50 to 1000 clips, although any number of clips may be used). As will be appreciated by those skilled in the art, these patterns or clips represent small portions (e.g. circuits, cells or patterns) of the design and especially the clips represent small portions for which particular attention or verification is needed. In other words, clips may be the portions of the design layout or may be similar or have a similar behavior of portions of the design layout where critical features are identified either by experience (including clips provided by a customer), by trial and error, or by running a full-chip simulation. Clips usually contain one or more test patterns or gauge patterns.
An initial larger set of clips may be provided a priori by a customer based on known critical feature areas in a design layout which require particular image optimization. Alternatively, in some embodiments, the initial larger set of clips may be extracted from the entire design layout by using some kind of automated (such as, machine vision) or manual method that identifies the critical feature areas.
Stochastic variations of the patterning process (e.g., resist process) potentially limit EUV lithography implementation for semiconductor high volume manufacturing (HVM) because of the combination of “few” photons per millijoule dose and preferred low dose processes, for example, in terms of shrink potential of features and exposure-dose specification, which in turn affects product yield or wafer throughput of the patterning process or both. In some embodiments, stochastic variations of a resist layer may manifest in different failure modes described by, for example, line width roughness (LWR), line edge roughness (LER), local CD non-uniformity, closed holes or trenches, or broken lines at the extreme condition. Such stochastic variations impact and limit successful high volume manufacturing (HVM). In order to characterize, understand and predict stochastic variation, the industry needs a trustworthy method to measure such variation for variety of design patterns.
Existing approaches of measuring stochastic variation involve different measurement techniques for different features. For example, line/space are measured in one direction (e.g., x or y), a contact hole or an array of contact holes patterns printed on a substrate may be measured in two directions (e.g., x and y). As an example of measures, a pattern measure is a line width roughness (LWR) (an example of one directional measurement), and a repeating dense contact array measure is local CD uniformity (LCDU) (an example of a two directional measurement). Various stochastic contributors cause variations in the LWR/LCDU of the features.
In order to control, reduce and predict the stochastic contributors, the semiconductor industry needs a robust solution to measure them accurately. Currently, the industry measures LWR for a line and LCDU for repeating a contact array to estimate the stochastic contributors. Moreover, these measures only focus on pattern level (e.g., one number per pattern) but not on edge point level (e.g., points along a contour of a pattern) where a hot spot occurs.
In some embodiments, a metrology tool such as scanning electron microscope (SEM) is used to characterize the stochastic contributors associated with the desired pattern. In SEM image data captured by the SEM tool, noise is embedded therein. In some embodiments, SEM images may be analyzed to determine a CD of a feature (e.g., CD of a contact hole), delta CD, which is a deviation of the CD from a mean of the CD distribution, and LCDU of the contact holes. In some embodiments, the term “local” (e.g., in LCDU) may refer to a particular area (e.g., a unit cell or a particular die. In some embodiments, the CD of a contact hole or the LCDU may be affected by a number of contributors including: (i) SEM noise (or SEM error contribution), δCDSEM, (ii) mask noise (or mask error contribution), δCDMASK. and (iii) resist noise (or resist error contribution), δCDRESIST. In the equation below, the CD of a measured contact hole can be expressed as:
CD=
wherein
The mask noise can originate from an error during mask manufacturing. The resist noise (also called shot noise) can originate from chemical layers in the resist together with a photon shot noise of a light source of a lithographic apparatus used in the printing the pattern on the substrate, and the SEM related noise can originate from the SEM (e.g., shot noise from electron flank). In existing technology, the decomposition of the noises can be performed based on a linear nested model. For example, contact hole's local critical dimension uniformity (LCDU) has various contributions including SEM noise, mask noise, and resist noise. In some embodiments, LCDU data can be provided to the linear nested model to decompose the three contributions.
In some embodiments, in order to prepare the data for the decomposition method using existing technologies, dedicated experiments are performed for taking the measurements, which include printing the design pattern on the substrate, using the same SEM metrology recipe twice to capture images of the pattern printed on the substrate, and enabling local alignment in the recipe to reduce SEM measurement location offset among different measurement repeats. Similar measurements may be performed among different dies. In some embodiments, anchor features (e.g., at a center of an area to be scanned) are usually included in the SEM's field of view (FOV) to help align SEM image among different measurements (and different dies).
In the disclosure, the term “repeat” used with reference to measurement of a substrate refers to multiple measurements taken at a specified location of the substrate using a specified metrology recipe. For example, repeat data refers to acquiring a plurality of images at a first location (e.g., center of a specified die) on the substrate at a specified metrology recipe (e.g., landing energy, probe current, scan rate, etc.). In some embodiments, at least two repeat data are generated from the plurality of images.
The disadvantages of existing technology include, but are not limited to, following. Dedicated experiments may need to be performed for obtaining the measurements, which is time consuming, cost prohibitive, consumes significant computing and manufacturing resources. The measurement process includes at least two repeats. Next, there exists a large (x, y, z) placement offset between any two measurement repeats. For example, when running the SEM metrology recipe multiple times, the recipe has to perform global and local alignment (e.g., wafer alignment) for each recipe run. Even with local alignment (which the reduces measurement throughput), the typical (x, y) placement error is approximately 10 nm. There exists a large variation in a time lag difference associated with the same die location, therefore large SEM-shrinkage uncertainty associated with resist of the substrate being measured. For example, when running the SEM metrology recipe twice, it is also hard to control the time lapse between a first measurement repeat and a second measurement repeat among different dies. The time lapse increases the shrinkage uncertainty between two measurement repeats. This shrinkage uncertainty will degrade the accuracy of the decomposition result such as the SEM noise, the mask noise, and the resist. There exist longer data acquisition time and higher chance of wafer damage. For example, to acquire a good quality SEM image at the defined locations on substrate, the metrology tool has to perform focus adjustment, global and local alignment for each recipe run. This leads longer acquisition time, and more chance for wafer damage. When running focus and local alignment with SEM beam, SEM beam can damage the wafer surface.
The present disclosure decomposes LWR/LCDU/CD distribution using independent component analysis (ICA) methods. Some of the advantages of the disclosed methods include eliminating the need for performing dedicated experiments and multiple repeats and minimizing a number of SEM images required for the decomposition (typically with significantly lesser number of SEM images than required by the prior known methods). Further, the disclosed methods perform the decomposition with less metrology measurement time and less wafer damage compared to existing methods. In some embodiments, the method uses a large FOV and high-throughput SEM tool (such as HMI), which can acquire SEM images covering large wafer area with short time. While the following embodiments for deriving error contributors are described with reference to CD distribution and LCDU data, the embodiments are not restricted to CD distribution and LCDU data, they can also be used to derive error contributions by decomposing LWR data of the features.
A simple application of ICA is the “cocktail party problem”, where the underlying speech signals (e.g., a first source signal 301 and a second source signal 302) are separated from a sample data consisting of people talking simultaneously in a room. The sample data can be different observations of different people talking simultaneously. For example, a first observation can be a first mixed signal 305 of both the source signals 301 and 302 output by a first sensor 311 (e.g., microphone) located at a first place in a room, and a second observation can be a second mixed signal 306 of both the source signals 301 and 302 output by a second sensor 312 (e.g., microphone) located in a second place different from the first place. A decomposer module 320, which is implemented based on the ICA method, can analyze the mixed signals 305 and 306 as linearly mixed signals, determine a mixing matrix (A) 313, and decompose the linearly mixed signals using unmixing matrix 314 to determine the original source signals 301 and 302.
In some embodiments, the ICA determines the mixing matrix as follows. In ICA, n mixed signals (e.g., mixed signals 305 and 306) are represented as n linear mixtures x1, . . . , xn of n independent components, s (e.g., source signals 301 and 302).
xj=aj
1
s
1
+aj
2
s
2
+ . . . +aj
n
s
n, for all j (2)
In some embodiments, a linear mixture is a linear function of a set of coefficients and explanatory variables (independent variables), whose value is used to predict the outcome of a dependent variable. In the above Eq. 2, the dependent variable can be xj, the set of coefficients can be aj1-ajn, and the explanatory variables can be s1-sn.
Let x denote the vector whose elements are the linear mixtures x1-xn, and likewise let s denote the vectors with elements s1-sn. Let A denote the matrix with the coefficients aij. Using this vector-matrix notation, the above mixing model may be written as
x=As (3)
or
x=Σ
i=1
n
aisi (4)
In some embodiments, the statistical model in Eq. 4 is called independent component analysis, or ICA model. The ICA model is a generative model, which means that it describes how the observed data are generated by a process of mixing the components si. The independent components are latent variables, meaning that they cannot be directly observed. Also, the mixing matrix (A) 313 is assumed to be unknown. All that is observed is the random vector x, and both A and s may be estimated using it. This must be done under as general assumptions as possible.
The ICA model performs a number of processes (e.g., linearly mixing the source signals, whitening the mixed signals, which are not described here for the sake of brevity) to determine the mixing matrix (A) 313. Then, after estimating the mixing matrix (A) 313, the inverse 314 of mixing matrix (A) 313, e.g., W, is obtained, which is then used to obtain the source component, s, by:
s=Wx (5)
In some embodiments, the ICA is based on two assumptions that (1) the source signals, si, are independent of each other, and (2) the values in each source signal, si, have non-Gaussian distributions. Further, in ICA, one of the constraints may be that if N sources are present, at least N observations (e.g. sensors or microphones) are needed to recover the original N signals. While the following paragraphs describe using three input signals to derive three error contributors, it should be noted that more than three input signals may be used to derive the three error contributors. In another example, if two error contributors are to be derived, then two or more input signals may be needed. In some embodiments, ICA method can be implemented using one of many algorithms, such as FastICA, infomax, JADE, and kernel-independent component analysis.
In some embodiments, the ICA method can be used in determining the error contributors, such as the δCDMASK, δCDRESIST, and δCDSEM, to the LCDU/CD distribution of contact holes printed on the substrate, which is described below at least with reference to
In some embodiments, a contour of a contact hole 410 may be obtained using threshold values associated with the SEM image 405. For example, the SEM image 405 may be a grayscale image and a threshold value can be a pixel value (e.g., corresponding to a white band in the greyscale image), such as 30%, 50%, or 70% as shown in the graph 415. The graph 415 shows CD values of the contour of a contact hole for various threshold values (e.g., white band values). In some embodiments, if the value of white colored pixel is “1” and black colored pixel is “0,” a threshold value of 30% of white band can be 30% of “1,” which is “0.3.” The position of a contour (e.g., contour height), and therefore, the CD of the contour may be obtained for that threshold value. In some embodiments, the threshold values correspond to the sensors described with respect to the ICA method in
The position of the contour, and therefore, the CD of the contour are typically affected by the error contributors. Accordingly, the CD value of the contour for a first threshold value 421 (e.g., 30%) can be used as or in deriving a mixed signal that can be input to an ICA method to be decomposed for obtaining the error contributors to the CD distribution. In some embodiments, instead of using the CD value, a δCD value may be used as the mixed signal that is input to the ICA method. In some embodiments, δCD value of a contact hole can be a difference between the mean CD value and the CD value of the contact hole. In some embodiments, the mean CD value is a mean of CD values of a number of contact holes. Further, in some embodiments, δCD value may be determined with the mean CD value shifted to “0” (which means the mean value is subtracted from the CD values of all the contact holes). In some embodiments, δCD value of a contact hole can be a distance between a specified point on a contour of a contact hole to a reference point on a reference contour of the contact hole. The reference contour may be obtained from a target pattern, which is simulated from a mask pattern of the corresponding contact hole.
In some embodiments, the relationship between δCD value of a contact hole and the error contributors may be expressed as:
δCD=δCDMASK+δCDRESIST+δCDSEM (6)
For decomposing the error contributors using the ICA, in some embodiments, the δCD can be represented as a linear mixture of the error contributors as follows:
δCD=a11*δCDMASK+a12*δCDRESIST+a13*δCDSEM (7)
where a11-a13 are set of coefficients of the linear mixture and part of the mixing matrix (A) 313 of the ICA.
The δCD value may be used as an input to the ICA method. However, in some embodiments, since there are three error contributors at least three different SCD values may be needed for the decomposition process as the ICA has a constraint that the number of mixed signals required as input has to be equal to or greater than the number of source components that need to be derived or decomposed. Accordingly, the δCD values are obtained for three different threshold values of the white band, e.g., a first δCD value, δCD30%, is obtained based on the CD value at the first threshold value 421 (e.g., 30% of the white band), a second δCD value, δCD50%, is obtained based on the CD value at a second threshold value 422 (e.g., 50% of the white band), and a third δCD value, δCD70%, is obtained based on the CD value at a third threshold value 423 (e.g., 70% of the white band). The three δCD values can be represented as three different linear mixtures of the error contributors as follows:
is the mixing matrix 313, and
δCDMASK, δCDRESIST, and δCDSEM are a function of the error contributors in Eqs. 8-10. For example, δCDMASK can be considered as a mean of δCDMASK(30%), δCDMASK(50%), and δCDMASK(70%) values, or δCDMASK can be considered as one of δCD(30%), δCD(50%), and δCD(70%) values.
While the above δCD values, δCD30%, δCD50%, and δCD70%, are determined with respect to one measurement point, a number of such δCD values are obtained for multiple measurement points for each of the three thresholds resulting in three different signals in which a first signal includes multiple δCD30% values, a second signal includes multiple δCD50% values, and a third signal includes multiple δCD70%, values.
In some embodiments, the measurement points or metrology points (e.g., a point where the CD value is measured) can be on the same contact hole or on different contact holes.
In some embodiments, the decomposer module 320 is implemented using an ICA method, which is discussed in detail at least with reference to
The decomposer module 320 may process the first, second, and third set of δCD values of the contact holes 515a-525a (e.g., based on the ICA method as described above at least with reference to
Accordingly, the decomposer module 320 may determine the values of each of the error contributors based on Eq. 12. The decomposer module 320 may output three signals or datasets corresponding to the δCDMASK, δCDRESIST, and δCDSEM error contributions. For example, a first output signal or dataset may include values that correspond to δCDMASK error contribution 601, a second output signal or dataset may include values that correspond to δCDRESIST error contribution 602, and a third output signal or dataset may include values that correspond to δCDSEM error contribution 603. In
In some embodiments, the decomposer module 320 may determine a particular error contribution in terms of a single value instead of or in addition to as a vector. For example, the decomposer module 320 may determine a mean of the values in the first dataset 601 as the δCDMASK error contribution.
In some embodiments, the error contribution values 601-603 may be used for improving/optimizing various aspects of a patterning process such as source optimization or mask optimization or optimal proximity correction process. For example, based on the δCDMASK error contribution or the δCDRESIST error contribution, one or more parameters of a mask/patterning device or a lithographic apparatus used to print the pattern may be adjusted so that the pattern printed on the substrate satisfies a specified criterion. The parameters that may be adjusted may include adjustable parameters of the source, the patterning device, the projection optics, dose, focus, characteristics of the design layout/pattern, etc. Typically, optimizing or improving a patterning process includes adjusting one or more parameters until one or more cost functions associated with process is minimized or satisfies a specified criterion. Some examples of optimization are described at least with reference to
While the above decomposition process uses the CD distribution data such as the first, second, and third set of δCD values 515a-525a as inputs 615-625, for determining the error contributors, in some embodiments, the decomposition process may also obtain the error contributors using LCDU data as inputs 615-625.
As illustrated in the graph of
Each LCDU dataset can be expressed as a linear mixture of the three error contributors as shown in the equation below (e.g., like the CD distribution linear mixtures of Eq. 8-10).
LCDU1=a11*LCDUMASK+a12*LCDURESIST+a13*LCDUSEM (13)
LCDU2=a21*LCDUMASK+a22*LCDURESIST+a23*LCDUSEM (14)
LCDU3=a31*LCDUMASK+a32*LCDURESIST+a33*LCDUSEM (15)
The above LCDU datasets 715-725 may be provided as inputs 615-625, respectively, to the decomposer module 320. The decomposer module 320 processes the first, second, and third LCDU datasets (e.g., based on the ICA method as described above at least with reference to
In another example, a white band value in a SEM image can be used as a sensor (e.g., as described at least with reference to
In another example, a focus level can be used as a sensor and different sets of LCDU data may be obtained for different focus levels as the input signals 615-625. For example, a first LCDU dataset including LCDU values for multiple dose values at a first focus level, a second LCDU dataset includes LCDU values for multiple dose values at a second focus level, and a third LCDU dataset includes LCDU values for multiple dose values at a third focus level may be obtained.
At operation 810, multiple measurement values 811 of a feature of the pattern are obtained using the image 801. For example, the measurement values 811 may include CD distribution data (e.g., CD or δCD values) or LCDU data of a number of contact holes for different sensor values. Different parameters can be used as sensors. For example, a threshold value associated with the image 801, such as a white band of the image 801 can be used as a sensor, and the measurement values 811 for different threshold values of the white band may include the first set of δCD values 515a obtained at the first threshold value 421 (e.g., 30% of the white band), the second set of δCD values 520a obtained at the second threshold value 422 (e.g., 50% of the white band), and a third set of δCD values 525a obtained at the third threshold value 423 (e.g., 70% of the white band) as described at least with reference to
In another example, a dose level can be used a sensor and the measurement values 811 for different dose levels may include the first LCDU dataset 715 obtained for the first dose level, the second LCDU dataset 720 obtained for the second dose level, and the third LCDU dataset 725 obtained for the third dose level as described at least with reference to
At operation 815, each of the measurement values 811 is correlated to a linear mixture of the multiple error contributions to generate a number of linear mixtures 816. In some embodiments, the error contributions are derived using ICA method (e.g., as described at least with reference to
At operation 820, the error contributions 821 are derived from the linear mixtures 816. In some embodiments, the linear mixtures 816 are decomposed using the ICA method as described at least with reference to
At operation 860, an inverse of the mixing matrix A 613 is determined, e.g., as shown in Eq. 12, to obtain an unmixing matrix 614.
At operation 865, the error contributions 821 are derived from the linear mixtures 816 using the unmixing matrix 614, e.g., as shown in Eq. 12.
In some embodiments, due to the presence of a noise (e.g., error contributions from multiple sources such as mask, resist and SEM), the contour 906 is subject to distortions giving rise to different contour heights, such as 906a, 906b and 906c. In some embodiments, the distorted contours 906a-906c may be identified by thresholding the SEM image to different threshold values and the CD values of the contour 906 can be obtained for different thresholds. For example, contour 906a may be identified be thresholding the SEM image 405 to a first threshold value (e.g., 30% of the white band value as described at least with reference to
At operation 910, CD values are obtained for different threshold values. For example, a specified threshold value 1051 may be the first threshold value 421 (e.g., 30% of the white band value) as shown in graph 415 of
A CD value may be obtained using any of a number of methods.
At the end of operation 910, different CD values (e.g., three different CD values) corresponding to different thresholds (e.g., three different thresholds 420-422) may be obtained for a particular cut-line (or measurement point). In some embodiments, the operations 905 and 910 are repeated for a finite number of iterations (e.g., user defined number) to obtain the CD values for each threshold for the finite number of measurement points (e.g., cutlines). The measurement points can be in on the same contact hole or different contact holes. At the conclusion of the finite number of iterations of 905 and 910, different sets of CD values are created. For example, the first set of CD values 515 corresponding to the first threshold value 421 of 30% as shown in
At operation 915, a mean value 916 of the CD values is determined. The CD values may include those obtained in operation 910, such as the first, second and third set of CD values 515-525.
At operation 920, the mean value 916 may be shifted to a specified value (e.g., “0”). In some embodiments, shifting the mean value 916 to the specified value may include subtracting a difference between the mean value 916 and the specified value from each of the CD values.
At operation 925, δCD value is obtained for each of the CD values in the first, second and third set of CD values 515-525. For example, the first set of δCD values 515a of
In some embodiments, after obtaining the first, second and third sets of δCD values 515a-525a, the process 900 may return to operation 815 of process 800.
When the substrate PSub is irradiated with electron beam EBP, secondary electrons are generated from the substrate PSub. The secondary electrons are deflected by the E×B deflector EBD2 and detected by a secondary electron detector SED. A two-dimensional electron beam image can be obtained by detecting the electrons generated from the sample in synchronization with, e.g., two-dimensional scanning of the electron beam by beam deflector EBD1 or with repetitive scanning of electron beam EBP by beam deflector EBD1 in an X or Y direction, together with continuous movement of the substrate PSub by the substrate table ST in the other of the X or Y direction.
A signal detected by secondary electron detector SED is converted to a digital signal by an analog/digital (A/D) converter ADC, and the digital signal is sent to an image processing system IPU. In some embodiments, the image processing system IPU may have memory MEM to store all or part of digital images for processing by a processing unit PU. The processing unit PU (e.g., specially designed hardware or a combination of hardware and software) is configured to convert or process the digital images into datasets representative of the digital images. Further, image processing system IPU may have a storage medium STOR configured to store the digital images and corresponding datasets in a reference database. A display device DIS may be connected with the image processing system IPU, so that an operator can conduct necessary operation of the equipment with the help of a graphical user interface.
The charged particle beam generator 81 generates a primary charged particle beam 91. The condenser lens module 82 condenses the generated primary charged particle beam 91. The probe forming objective lens module 83 focuses the condensed primary charged particle beam into a charged particle beam probe 92. The charged particle beam deflection module 84 scans the formed charged particle beam probe 92 across the surface of an area of interest on the sample 90 secured on the sample stage 89. In some embodiments, the charged particle beam generator 81, the condenser lens module 82 and the probe forming objective lens module 83, or their equivalent designs, alternatives or any combination thereof, together form a charged particle beam probe generator which generates the scanning charged particle beam probe 92.
The secondary charged particle detector module 85 detects secondary charged particles 93 emitted from the sample surface (maybe also along with other reflected or scattered charged particles from the sample surface) upon being bombarded by the charged particle beam probe 92 to generate a secondary charged particle detection signal 94. The image forming module 86 (e.g., a computing device) is coupled with the secondary charged particle detector module 85 to receive the secondary charged particle detection signal 94 from the secondary charged particle detector module 85 and accordingly forming at least one scanned image. In some embodiments, the secondary charged particle detector module 85 and image forming module 86, or their equivalent designs, alternatives or any combination thereof, together form an image forming apparatus which forms a scanned image from detected secondary charged particles emitted from sample 90 being bombarded by the charged particle beam probe 92.
As noted above, SEM images may be processed to extract contours that describe the edges of objects, representing device structures, in the image. These contours are then quantified via metrics, such as CD. Thus, typically, the images of device structures are compared and quantified via simplistic metrics, such as an edge-to-edge distance (CD) or simple pixel differences between images. Typical contour models that detect the edges of the objects in an image in order to measure CD use image gradients. Indeed, those models rely on strong image gradients. But, in practice, the image typically is noisy and has discontinuous boundaries. Techniques, such as smoothing, adaptive thresholding, edge-detection, erosion, and dilation, may be used to process the results of the image gradient contour models to address noisy and discontinuous images, but will ultimately result in a low-resolution quantification of a high-resolution image. Thus, in most instances, mathematical manipulation of images of device structures to reduce noise and automate edge detection results in loss of resolution of the image, thereby resulting in loss of information. Consequently, the result is a low-resolution quantification that amounts to a simplistic representation of a complicated, high-resolution structure.
So, it is desirable to have amathematical representation of the structures (e.g., circuit features, alignment mark or metrology target portions (e.g., grating features), etc.) produced or expected to be produced using a patterning process, whether, e.g., the structures are in a latent resist image, in a developed resist image or transferred to a layer on the substrate, e.g., by etching, that can preserve the resolution and yet describe the general shape of the structures. In the context of lithography or other pattering processes, the structure may be a device or a portion thereof that is being manufactured and the images may be SEM images of the structure. In some instances, the structure may be a feature of semiconductor device, e.g., integrated circuit. In some instances, the structure may be an alignment mark, or a portion thereof (e.g., a grating of the alignment mark), that is used in an alignment measurement process to determine alignment of an object (e.g., a substrate) with another object (e.g., a patterning device) or a metrology target, or a portion thereof (e.g., a grating of the metrology target), that is used to measure a parameter (e.g., overlay, focus, dose, etc.) of the patterning process. In some embodiments, the metrology target is a diffractive grating used to measure, e.g., overlay.
In some embodiments, the measurement data (e.g., stochastic variations) related to the printed pattern, determined according to the method of
In order to ensure that the projected image of the design layout is in accordance with requirements of a given target circuit design, proximity effects need to be predicted and compensated for, using sophisticated numerical models, corrections or pre-distortions of the design layout. The article “Full-Chip Lithography Simulation and Design Analysis—How OPC Is Changing IC Design”, C. Spence, Proc. SPIE, Vol. 5751, pp 1-14 (2005) provides an overview of current “model-based” optical proximity correction processes. In a typical high-end design almost every feature of the design layout has some modification in order to achieve high fidelity of the projected image to the target design. These modifications may include shifting or biasing of edge positions or line widths as well as application of “assist” features that are intended to assist projection of other features.
Application of model-based OPC to a target design involves good process models and considerable computational resources, given the many millions of features typically present in a chip design. However, applying OPC is generally not an “exact science”, but an empirical, iterative process that does not always compensate for all possible proximity effect. Therefore, effect of OPC, e.g., design layouts after application of OPC and any other RET, need to be verified by design inspection, i.e. intensive full-chip simulation using calibrated numerical process models, in order to minimize the possibility of design flaws being built into the patterning device pattern. This is driven by the enormous cost of making high-end patterning devices, which run in the multi-million-dollar range, as well as by the impact on turn-around time by reworking or repairing actual patterning devices once they have been manufactured.
Both OPC and full-chip RET verification may be based on numerical modeling systems and methods as described, for example in, U.S. Pat. No. 10,815,573 and an article titled “Optimized Hardware and Software For Fast, Full Chip Simulation”, by Y. Cao et al., Proc. SPIE, Vol. 5754, 405 (2005).
One RET is related to adjustment of the global bias of the design layout. The global bias is the difference between the patterns in the design layout and the patterns intended to print on the substrate. For example, a circular pattern of 25 nm diameter may be printed on the substrate by a 50 nm diameter pattern in the design layout or by a 20 nm diameter pattern in the design layout but with high dose.
In addition to optimization to design layouts or patterning devices (e.g., OPC), the illumination source can also be optimized, either jointly with patterning device optimization or separately, in an effort to improve the overall lithography fidelity. The terms “illumination source” and “source” are used interchangeably in this document. Since the 1990s, many off-axis illumination sources, such as annular, quadrupole, and dipole, have been introduced, and have provided more freedom for OPC design, thereby improving the imaging results, As is known, off-axis illumination is a proven way to resolve fine structures (i.e., target features) contained in the patterning device. However, when compared to a traditional illumination source, an off-axis illumination source usually provides less radiation intensity for the aerial image (AI). Thus, it becomes desirable to attempt to optimize the illumination source to achieve the optimal balance between finer resolution and reduced radiation intensity.
Numerous illumination source optimization approaches can be found, for example, in an article by Rosenbluth et al., titled “Optimum Mask and Source Patterns to Print A Given Shape”, Journal of Microlithography, Microfabrication, Microsystems 1(1), pp. 13-20, (2002). The source is partitioned into several regions, each of which corresponds to a certain region of the pupil spectrum. Then, the source distribution is assumed to be uniform in each source region and the brightness of each region is optimized for process window. However, such an assumption that the source distribution is uniform in each source region is not always valid, and as a result the effectiveness of this approach suffers. In another example set forth in an article by Granik, titled “Source Optimization for Image Fidelity and Throughput”, Journal of Microlithography, Microfabrication, Microsystems 3(4), pp. 509-522, (2004), several existing source optimization approaches are overviewed, and a method based on illuminator pixels is proposed that converts the source optimization problem into a series of non-negative least square optimizations. Though these methods have demonstrated some successes, they typically require multiple complicated iterations to converge. In addition, it may be difficult to determine the appropriate/optimal values for some extra parameters, such as γ in Granik's method, which dictates the trade-off between optimizing the source for substrate image fidelity and the smoothness requirement of the source.
For low k1 photolithography, optimization of both the source and patterning device is useful to ensure a viable process window for projection of critical circuit patterns. Some algorithms (e.g. Socha et. al. Proc. SPIE vol. 5853, 2005, p.180) discretize illumination into independent source points and mask into diffraction orders in the spatial frequency domain, and separately formulate a cost function (which is defined as a function of selected design variables) based on process window metrics such as exposure latitude which could be predicted by optical imaging models from source point intensities and patterning device diffraction orders. The term “design variables” as used herein comprises a set of parameters of a lithographic projection apparatus or a lithographic process, for example, parameters a user of the lithographic projection apparatus can adjust, or image characteristics a user can adjust by adjusting those parameters. It should be appreciated that any characteristics of a lithographic projection process, including those of the source, the patterning device, the projection optics, or resist characteristics can be among the design variables in the optimization. The cost function is often a non-linear function of the design variables. Then standard optimization techniques are used to minimize the cost function.
Relatedly, the pressure of ever decreasing design rules have driven semiconductor chipmakers to move deeper into the low k1 lithography era with existing 193 nm ArF lithography. Lithography towards lower k1 puts heavy demands on RET, exposure tools, and the need for litho-friendly design. 1.35 ArF hyper numerical aperture (NA) exposure tools may be used in the future. To help ensure that circuit design can be produced on to the substrate with workable process window, source-patterning device optimization (referred to herein as source-mask optimization or SMO) is becoming a significant RET for 2× nm node.
A source and patterning device (design layout) optimization method and system that allows for simultaneous optimization of the source and patterning device using a cost function without constraints and within a practicable amount of time is described in a commonly assigned International Patent Application No. PCT/US2009/065359, filed on Nov. 20, 2009, and published as WO2010/059954, titled “Fast Freeform Source and Mask Co-Optimization Method”, which is hereby incorporated by reference in its entirety.
Another source and mask optimization method and system that involves optimizing the source by adjusting pixels of the source is described in a commonly assigned U.S. patent application Ser. No. 12/813,456, filed on Jun. 10, 2010, and published as U.S. Patent Application Publication No. 2010/0315614, titled “Source-Mask Optimization in Lithographic Apparatus”, which is hereby incorporated by reference in its entirety.
In a lithographic projection apparatus, as an example, a cost function is expressed as
CF(z1,z2, . . . ,zN)=Σp=1pwpƒp2(z1,z2, . . . ,zN) (Eq. 1)
wherein (z1, z2, . . . , zN) are N design variables or values thereof. ƒp(z1, z2, . . . , zN) can be a function of the design variables (z1, z2, . . . , zN) such as a difference between an actual value and an intended value of a characteristic at an evaluation point for a set of values of the design variables of (z1, z2, . . . , zN). wp is a weight constant associated with ƒp(z1, z2, . . . , zN). An evaluation point or pattern more critical than others can be assigned a higher wp value. Patterns or evaluation points with larger number of occurrences may be assigned a higher wp value, too. Examples of the evaluation points can be any physical point or pattern on the substrate, any point on a virtual design layout, or resist image, or aerial image, or a combination thereof ƒp(z1, z2, . . . , zN) can also be a function of one or more stochastic effects such as the LWR, which are functions of the design variables (z1, z2, . . . , zN). The cost function may represent any suitable characteristics of the lithographic projection apparatus or the substrate, for instance, failure rate of a feature, focus, CD, image shift, image distortion, image rotation, stochastic effects, throughput, CDU, or a combination thereof. CDU is local CD variation (e.g., three times of the standard deviation of the local CD distribution). CDU may be interchangeably referred to as LCDU. In one example, the cost function represents (i.e., is a function of) CDU, throughput, and the stochastic effects. In one example, the cost function represents (i.e., is a function of) EPE, throughput, and the stochastic effects. In one example, the design variables (z1, z2, . . . , zN) comprise dose, global bias of the patterning device, shape of illumination from the source, or a combination thereof. Since it is the resist image that often dictates the circuit pattern on a substrate, the cost function often includes functions that represent some characteristics of the resist image. For example, ƒp(z1, z2, . . . , zN) of such an evaluation point can be simply a distance between a point in the resist image to an intended position of that point (i.e., edge placement error EPEp(z1, z2, . . . , zN)). The design variables can be any adjustable parameters such as adjustable parameters of the source, the patterning device, the projection optics, dose, focus, etc. The projection optics may include components collectively called as “wavefront manipulator” that can be used to adjust shapes of a wavefront and intensity distribution or phase shift of the irradiation beam. The projection optics preferably can adjust a wavefront and intensity distribution at any location along an optical path of the lithographic projection apparatus, such as before the patterning device, near a pupil plane, near an image plane, near a focal plane. The projection optics can be used to correct or compensate for certain distortions of the wavefront and intensity distribution caused by, for example, the source, the patterning device, temperature variation in the lithographic projection apparatus, thermal expansion of components of the lithographic projection apparatus. Adjusting the wavefront and intensity distribution can change values of the evaluation points and the cost function. Such changes can be simulated from a model or actually measured. Of course, CF(z1, z2, . . . , zN) is not limited the form in Eq. 1. CF(z1, z2, . . . , zN) can be in any other suitable form.
It should be noted that the normal weighted root mean square (RMS) of ƒp(z1, z2, . . . , zN) is defined as
therefore, minimizing the weighted RMS of ƒp(z1, z2, . . . , zN) is equivalent to minimizing the cost function CF(z1, z2, . . . , zN)=Σp=1pwpƒp2(z1, z2, . . . , zN), defined in Eq. 1. Thus the weighted RMS of ƒp(z1, z2, . . . , zN) and Eq. 1 may be utilized interchangeably for notational simplicity herein.
Further, if considering maximizing the PW (Process Window), one can consider the same physical location from different PW conditions as different evaluation points in the cost function in (Eq. 1). For example, if considering N PW conditions, then one can categorize the evaluation points according to their PW conditions and write the cost functions as:
Where ƒp
The design variables may have constraints, which can be expressed as (z1, z2, . . . , zN)∈Z, where Z is a set of possible values of the design variables. One possible constraint on the design variables may be imposed by yield or a desired throughput of the lithographic projection apparatus. The desired yield or throughput may limit the dose and thus has implications for the stochastic effects (e.g., imposing a lower bound on the stochastic effects). Higher throughput generally leads to lower dose, shorter longer exposure time and greater stochastic effects. Higher yield generally leads to a restricted design which may be sensitive to stochastic risk. Consideration of substrate throughput, yield and minimization of the stochastic effects may constrain the possible values of the design variables because the stochastic effects are function of the design variables. Without such a constraint imposed by the desired throughput, the optimization may yield a set of values of the design variables that are unrealistic. For example, if the dose is among the design variables, without such a constraint, the optimization may yield a dose value that makes the throughput economically impossible. However, the usefulness of constraints should not be interpreted as a necessity. The throughput may be affected by the failure rate-based adjustment to parameters of the patterning process. It is desirable to have lower failure rate of the feature while maintaining a high throughput. Throughput may also be affected by the resist chemistry. Slower resist (e.g., a resist that requires higher amount of light to be properly exposed) leads to lower throughput. Thus, based on the optimization process involving failure rate of a feature due to resist chemistry or fluctuations, and dose requirements for higher throughput, appropriate parameters of the patterning process may be determined.
The optimization process therefore is to find a set of values of the design variables, under the constraints (z1, z2, . . . , zN)∈Z, that minimize the cost function, i.e., to find
A general method of optimizing the lithography projection apparatus, according to some embodiments, is illustrated in
In a lithographic projection apparatus, the source, patterning device and projection optics can be optimized alternatively (referred to as Alternative Optimization) or optimized simultaneously (referred to as Simultaneous Optimization). The terms “simultaneous”, “simultaneously”, “joint” and “jointly” as used herein mean that the design variables of the characteristics of the source, patterning device, projection optics or any other design variables, are allowed to change at the same time. The term “alternative” and “alternatively” as used herein mean that not all of the design variables are allowed to change at the same time.
In
As shown in the non-limiting example flowchart of
The pattern selection algorithm, as discussed before, may be integrated with the simultaneous or alternative optimization. For example, when an alternative optimization is adopted, first a full-chip SO can be performed, the ‘hot spots’ or ‘warm spots’ are identified, then an MO is performed. In view of the present disclosure numerous permutations and combinations of sub-optimizations are possible in order to achieve the desired optimization results.
In an exemplary optimization process, no relationship between the design variables (z1, z2, . . . , zN) and ƒp(z1, z2, . . . , zN) is assumed or approximated, except that ƒp(z1, z2, . . . , zN) is sufficiently smooth (e.g. first order derivatives
n=1, 2, . . . N) exist), which is generally valid in a lithographic projection apparatus. An algorithm, such as the Gauss-Newton algorithm, the Levenberg-Marquardt algorithm, the gradient descent algorithm, simulated annealing, the genetic algorithm, can be applied to find ({tilde over (z)}1, {tilde over (z)}2, . . . , {tilde over (z)}N).
Here, the Gauss-Newton algorithm is used as an example. The Gauss-Newton algorithm is an iterative method applicable to a general non-linear multi-variable optimization problem. In the i-th iteration wherein the design variables (z1, z2, . . . , zN) take values of (z1i, z2i, zNi), the Gauss-Newton algorithm linearizes (z1, z2, . . . , zN) in the vicinity of (z1i, z2i, . . . , zNi), and then calculates values (z1(i+1), z2(i+1), . . . , zN(i+1)) in the vicinity of (z1i, z2i, . . . , zNi) that give a minimum of CF(z1, z2, . . . , zN). The design variables (z1, z2, . . . , zN) take the values of (z1(i+1), z2(i+1), . . . , zN(i+1)) in the (i+1)-th iteration. This iteration continues until convergence (i.e. CF(z1, z2, . . . , zN) does not reduce any further) or a preset number of iterations is reached.
Specifically, in the i-th iteration, in the vicinity of (z1i, z2i, . . . , zNi),
Under the approximation of Eq. 3, the cost function becomes:
which is a quadratic function of the design variables (z1, z2, . . . , zN). Every term is constant except the design variables (z1, z2, . . . , zN).
If the design variables (z1, z2, . . . , zN) are not under any constraints, (z1(i+1), z2(i+1), . . . , zN(i+1)) can be derived by solving by N linear equations:
wherein n=1, 2, . . . N.
If the design variables (z1, z2, . . . , zN) are under the constraints in the form of J inequalities (e.g. tuning ranges of (z1, z2, . . . , zN)) Σn=1NAnjzn≤Bj, for j=1, 2, . . . J; and K equalities (e.g. interdependence between the design variables) Σn=1NCnkzn≤Dk, for k=1, 2, . . . K; the optimization process becomes a classic quadratic programming problem, wherein Anj, Bj, Cnk, Dk are constants. Additional constraints can be imposed for each iteration. For example, a “damping factor” ΔD can be introduced to limit the difference between (z1(i+1), z2(i+1), . . . , zN(i+1)) and (z1i, z2i, . . . , zNi), so that the approximation of Eq. 3 holds. Such constraints can be expressed as zni−ΔD≤zn≤zni+ΔD. (z1(i+1), z2(i+1), . . . , zN(i+1)) can be derived using, for example, methods described in Numerical Optimization (2nd ed.) by Jorge Nocedal and Stephen J. Wright (Berlin New York: Vandenberghe. Cambridge University Press).
Instead of minimizing the RMS of ƒp(z1, z2, . . . , zN), the optimization process can minimize magnitude of the largest deviation (the worst defect) among the evaluation points to their intended values. In this approach, the cost function can alternatively be expressed as
wherein CLp is the maximum allowed value for ƒp(z1, z2, . . . , zN). This cost function represents the worst defect among the evaluation points. Optimization using this cost function minimizes magnitude of the worst defect. An iterative greedy algorithm can be used for this optimization.
The cost function of Eq. 5 can be approximated as:
wherein q is an even positive integer such as at least 4, preferably at least 10. Eq. 6 mimics the behavior of Eq. 5, while allowing the optimization to be executed analytically and accelerated by using methods such as the deepest descent method, the conjugate gradient method, etc.
Minimizing the worst defect size can also be combined with linearizing of ƒp(z1, z2, . . . , zN). Specifically, ƒp(z1, z2, . . . , zN) is approximated as in Eq. 3. Then the constraints on worst defect size are written as inequalities ELp≤ƒp(z1, z2, . . . , zN)≤EUp, wherein ELp and EUp are two constants specifying the minimum and maximum allowed deviation for the ƒp(z1, z2, . . . , zN). Plugging Eq. 3 in, these constraints are transformed to, for p=1, . . . P,
Since Eq. 3 is generally valid only in the vicinity of (z1i, z2i, . . . , zNi), in case the desired constraints ELp≤ƒp(z1, z2, . . . , zN)≤EUp cannot be achieved in such vicinity, which can be determined by any conflict among the inequalities, the constants ELp and EUp can be relaxed until the constraints are achievable. This optimization process minimizes the worst defect size in the vicinity of (z1i, z2i, . . . , zNi). Then each step reduces the worst defect size gradually, and each step is executed iteratively until certain terminating conditions are met. This will lead to optimal reduction of the worst defect size.
Another way to minimize the worst defect is to adjust the weight wp in each iteration. For example, after the i-th iteration, if the r-th evaluation point is the worst defect, wr can be increased in the (i+1)-th iteration so that the reduction of that evaluation point's defect size is given higher priority.
In addition, the cost functions in Eq.4 and Eq.5 can be modified by introducing a Lagrange multiplier to achieve compromise between the optimization on RMS of the defect size and the optimization on the worst defect size, i.e.,
where λ is a preset constant that specifies the trade-off between the optimization on RMS of the defect size and the optimization on the worst defect size. In particular, if λ=0, then this becomes Eq.4 and the RMS of the defect size is only minimized; while if λ=1, then this becomes Eq.5 and the worst defect size is only minimized; if 0<λ<1, then both are taken into consideration in the optimization. Such optimization can be solved using multiple methods. For example, the weighting in each iteration may be adjusted, similar to the one described previously. Alternatively, similar to minimizing the worst defect size from inequalities, the inequalities of Eq. 6′ and 6″ can be viewed as constraints of the design variables during solution of the quadratic programming problem. Then, the bounds on the worst defect size can be relaxed incrementally or increase the weight for the worst defect size incrementally, compute the cost function value for every achievable worst defect size, and choose the design variable values that minimize the total cost function as the initial point for the next step. By doing this iteratively, the minimization of this new cost function can be achieved.
Optimizing a lithographic projection apparatus can expand the process window. A larger process window provides more flexibility in process design and chip design. The process window can be defined as a set of focus and dose values for which the resist image are within a certain limit of the design target of the resist image. Note that all the methods discussed here may also be extended to a generalized process window definition that can be established by different or additional base parameters in addition to exposure dose and defocus. These may include, but are not limited to, optical settings such as NA, sigma, aberrations, polarization, or optical constants of the resist layer. For example, as described earlier, if the PW also consists of different mask bias, then the optimization includes the minimization of MEEF (Mask Error Enhancement Factor), which is defined as the ratio between the substrate EPE and the induced mask edge bias. The process window defined on focus and dose values only serve as an example in this disclosure. A method of maximizing the process window, according to some embodiments, is described below.
In a first step, starting from a known condition (ƒ0, ε0) in the process window, wherein ƒ0 is a nominal focus and ε0 is a nominal dose, minimizing one of the cost functions below in the vicinity
If the nominal focus ƒ0 and nominal dose ε0 are allowed to shift, they can be optimized jointly with the design variables (z1, z2, . . . , zN). In the next step, (ƒ0±Δƒ, ε0±Δε) is accepted as part of the process window, if a set of values of (z1, z2, . . . , zN, ƒ, ε) can be found such that the cost function is within a preset limit.
Alternatively, if the focus and dose are not allowed to shift, the design variables (z1, z2, . . . , zN) are optimized with the focus and dose fixed at the nominal focus ƒ0 and nominal dose co. In an alternative example, (ƒ0±Δƒ, ε0±Δε) is accepted as part of the process window, if a set of values of (z1, z2, . . . , zN) can be found such that the cost function is within a preset limit.
The methods described earlier in this disclosure can be used to minimize the respective cost functions of Eqs. 7, 7′, or 7″. If the design variables are characteristics of the projection optics, such as the Zernike coefficients, then minimizing the cost functions of Eqs. 7, 7′, or 7″ leads to process window maximization based on projection optics optimization, i.e., LO. If the design variables are characteristics of the source and patterning device in addition to those of the projection optics, then minimizing the cost function of Eqs. 7, 7′, or 7″ leads to process window maximizing based on SMLO, as illustrated in
The method starts by defining the pixel groups of the illumination source and the patterning device tiles of the patterning device (step S802). Generally, a pixel group or a patterning device tile may also be referred to as a division of a lithographic process component. In one exemplary approach, the illumination source is divided into 117 pixel groups, and 94 patterning device tiles are defined for the patterning device, substantially as described above, resulting in a total of 211 divisions.
In step S804, a lithographic model is selected as the basis for photolithographic simulation. Photolithographic simulations produce results that are used in calculations of photolithographic metrics, or responses. A particular photolithographic metric is defined to be the performance metric that is to be optimized (step S806). In step S808, the initial (pre-optimization) conditions for the illumination source and the patterning device are set up. Initial conditions include initial states for the pixel groups of the illumination source and the patterning device tiles of the patterning device such that references may be made to an initial illumination shape and an initial patterning device pattern. Initial conditions may also include mask bias, NA, and focus ramp range. Although steps S802, S804, S806, and S808 are depicted as sequential steps, it will be appreciated that in other embodiments, these steps may be performed in other sequences.
In step S810, the pixel groups and patterning device tiles are ranked. Pixel groups and patterning device tiles may be interleaved in the ranking. Various ways of ranking may be employed, including: sequentially (e.g., from pixel group 1 to pixel group 117 and from patterning device tile 1 to patterning device tile 94), randomly, according to the physical locations of the pixel groups and patterning device tiles (e.g., ranking pixel groups closer to the center of the illumination source higher), and according to how an alteration of the pixel group or patterning device tile affects the performance metric.
Once the pixel groups and patterning device tiles are ranked, the illumination source and patterning device are adjusted to improve the performance metric (step S812). In step S812, each of the pixel groups and patterning device tiles are analyzed, in order of ranking, to determine whether an alteration of the pixel group or patterning device tile will result in an improved performance metric. If it is determined that the performance metric will be improved, then the pixel group or patterning device tile is accordingly altered, and the resulting improved performance metric and modified illumination shape or modified patterning device pattern form the baseline for comparison for subsequent analyses of lower-ranked pixel groups and patterning device tiles. In other words, alterations that improve the performance metric are retained. As alterations to the states of pixel groups and patterning device tiles are made and retained, the initial illumination shape and initial patterning device pattern changes accordingly, so that a modified illumination shape and a modified patterning device pattern result from the optimization process in step S812.
In other approaches, patterning device polygon shape adjustments and pairwise polling of pixel groups or patterning device tiles are also performed within the optimization process of S812.
In some embodiments, the interleaved simultaneous optimization procedure may include to alter a pixel group of the illumination source and if an improvement of the performance metric is found, the dose is stepped up and down to look for further improvement. In some embodiments, the stepping up and down of the dose or intensity may be replaced by a bias change of the patterning device pattern to look for further improvement in the simultaneous optimization procedure.
In step S814, a determination is made as to whether the performance metric has converged. The performance metric may be considered to have converged, for example, if little or no improvement to the performance metric has been witnessed in the last several iterations of steps S810 and S812. If the performance metric has not converged, then the steps of S810 and S812 are repeated in the next iteration, where the modified illumination shape and modified patterning device from the current iteration are used as the initial illumination shape and initial patterning device for the next iteration (step S816).
The optimization methods described above may be used to increase the throughput of the lithographic projection apparatus. For example, the cost function may include an ƒp(z1, z2, . . . , zN) that is a function of the exposure time. Optimization of such a cost function is preferably constrained or influenced by a measure of the stochastic effects or other metrics. Specifically, a computer-implemented method for increasing a throughput of a lithographic process may include optimizing a cost function that is a function of one or more stochastic effects of the lithographic process and a function of an exposure time of the substrate, in order to minimize the exposure time.
In some embodiments, the cost function includes at least one ƒp(z1, z2, . . . , zN) that is a function of one or more stochastic effects. The stochastic effects may include the failure of a feature, measurement data (e.g., SEPE) determined as in method of
Computer system 100 may be coupled via bus 102 to a display 112, such as a cathode ray tube (CRT) or flat panel or touch panel display for displaying information to a computer user. An input device 114, including alphanumeric and other keys, is coupled to bus 102 for communicating information and command selections to processor 104. Another type of user input device is cursor control 116, such as a mouse, a trackball, or cursor direction keys for communicating direction information and command selections to processor 104 and for controlling cursor movement on display 112. This input device typically has two degrees of freedom in two axes, a first axis (e.g., x) and a second axis (e.g., y), that allows the device to specify positions in a plane. A touch panel (screen) display may also be used as an input device.
According to some embodiments, portions of one or more methods described herein may be performed by computer system 100 in response to processor 104 executing one or more sequences of one or more instructions contained in main memory 106. Such instructions may be read into main memory 106 from another computer-readable medium, such as storage device 110. Execution of the sequences of instructions contained in main memory 106 causes processor 104 to perform the process steps described herein. One or more processors in a multi-processing arrangement may also be employed to execute the sequences of instructions contained in main memory 106. In some embodiments, hard-wired circuitry may be used in place of or in combination with software instructions. Thus, the description herein is not limited to any specific combination of hardware circuitry and software.
The term “computer-readable medium” as used herein refers to any medium that participates in providing instructions to processor 104 for execution. Such a medium may take many forms, including but not limited to, non-volatile media, volatile media, and transmission media. Non-volatile media include, for example, optical or magnetic disks, such as storage device 110. Volatile media include dynamic memory, such as main memory 106. Transmission media include coaxial cables, copper wire and fiber optics, including the wires that comprise bus 102. Transmission media can also take the form of acoustic or light waves, such as those generated during radio frequency (RF) and infrared (IR) data communications. Common forms of computer-readable media include, for example, a floppy disk, a flexible disk, hard disk, magnetic tape, any other magnetic medium, a CD-ROM, DVD, any other optical medium, punch cards, paper tape, any other physical medium with patterns of holes, a RAM, a PROM, and EPROM, a FLASH-EPROM, any other memory chip or cartridge, a carrier wave as described hereinafter, or any other medium from which a computer can read.
Various forms of computer readable media may be involved in carrying one or more sequences of one or more instructions to processor 104 for execution. For example, the instructions may initially be borne on a magnetic disk of a remote computer. The remote computer can load the instructions into its dynamic memory and send the instructions over a telephone line using a modem. A modem local to computer system 100 can receive the data on the telephone line and use an infrared transmitter to convert the data to an infrared signal. An infrared detector coupled to bus 102 can receive the data carried in the infrared signal and place the data on bus 102. Bus 102 carries the data to main memory 106, from which processor 104 retrieves and executes the instructions. The instructions received by main memory 106 may optionally be stored on storage device 110 either before or after execution by processor 104.
Computer system 100 also preferably includes a communication interface 118 coupled to bus 102. Communication interface 118 provides a two-way data communication coupling to a network link 120 that is connected to a local network 122. For example, communication interface 118 may be an integrated services digital network (ISDN) card or a modem to provide a data communication connection to a corresponding type of telephone line. As another example, communication interface 118 may be a local area network (LAN) card to provide a data communication connection to a compatible LAN. Wireless links may also be implemented. In any such implementation, communication interface 118 sends and receives electrical, electromagnetic or optical signals that carry digital data streams representing various types of information.
Network link 120 typically provides data communication through one or more networks to other data devices. For example, network link 120 may provide a connection through local network 122 to a host computer 124 or to data equipment operated by an Internet Service Provider (ISP) 126. ISP 126 in turn provides data communication services through the worldwide packet data communication network, now commonly referred to as the “Internet” 128. Local network 122 and Internet 128 both use electrical, electromagnetic or optical signals that carry digital data streams. The signals through the various networks and the signals on network link 120 and through communication interface 118, which carry the digital data to and from computer system 100, are exemplary forms of carrier waves transporting the information.
Computer system 100 can send messages and receive data, including program code, through the network(s), network link 120, and communication interface 118. In the Internet example, a server 130 might transmit a requested code for an application program through Internet 128, ISP 126, local network 122 and communication interface 118. One such downloaded application may provide for the illumination optimization, for example. The received code may be executed by processor 104 as it is received, or stored in storage device 110, or other non-volatile storage for later execution. In this manner, computer system 100 may obtain application code in the form of a carrier wave.
an illumination system IL, to condition a beam B of radiation. In this particular case, the illumination system also comprises a radiation source SO;
a first object table (e.g., mask table) MT provided with a patterning device holder to hold a patterning device MA (e.g., a reticle), and connected to a first positioner to accurately position the patterning device with respect to item PS;
a second object table (substrate table) WT provided with a substrate holder to hold a substrate W (e.g., a resist-coated silicon wafer), and connected to a second positioner to accurately position the substrate with respect to item PS;
a projection system (“lens”) PS (e.g., a refractive, catoptric or catadioptric optical system) to image an irradiated portion of the patterning device MA onto a target portion C (e.g., comprising one or more dies) of the substrate W.
As depicted herein, the apparatus is of a transmissive type (i.e., has a transmissive mask). However, in general, it may also be of a reflective type, for example (with a reflective mask). Alternatively, the apparatus may employ another kind of patterning device as an alternative to the use of a classic mask; examples include a programmable mirror array or LCD matrix.
The source SO (e.g., a mercury lamp or excimer laser) produces a beam of radiation. This beam is fed into an illumination system (illuminator) IL, either directly or after having traversed conditioning means, such as a beam expander Ex, for example. The illuminator IL may comprise adjusting means AD for setting the outer or inner radial extent (commonly referred to as σ-outer and σ-inner, respectively) of the intensity distribution in the beam. In addition, it will generally comprise various other components, such as an integrator IN and a condenser CO. In this way, the beam B impinging on the patterning device MA has a desired uniformity and intensity distribution in its cross-section.
It should be noted with regard to
The beam PB subsequently intercepts the patterning device MA, which is held on a patterning device table MT. Having traversed the patterning device MA, the beam B passes through the lens PL, which focuses the beam B onto a target portion C of the substrate W. With the aid of the second positioning means (and interferometric measuring means IF), the substrate table WT can be moved accurately, e.g. so as to position different target portions C in the path of the beam PB. Similarly, the first positioning means can be used to accurately position the patterning device MA with respect to the path of the beam B, e.g., after mechanical retrieval of the patterning device MA from a patterning device library, or during a scan. In general, movement of the object tables MT, WT will be realized with the aid of a long-stroke module (coarse positioning) and a short-stroke module (fine positioning), which are not explicitly depicted in
The depicted tool can be used in two different modes:
In step mode, the patterning device table MT is kept essentially stationary, and an entire patterning device image is projected in one go (i.e., a single “flash”) onto a target portion C. The substrate table WT is then shifted in the x or y directions so that a different target portion C can be irradiated by the beam PB;
In scan mode, essentially the same scenario applies, except that a given target portion C is not exposed in a single “flash”. Instead, the patterning device table MT is movable in a given direction (the so-called “scan direction”, e.g., the y direction) with a speed v, so that the projection beam B is caused to scan over a patterning device image; concurrently, the substrate table WT is simultaneously moved in the same or opposite direction at a speed V=Mv, in which M is the magnification of the lens PL (typically, M=1/4 or 1/5). In this manner, a relatively large target portion C can be exposed, without having to compromise on resolution.
The lithographic projection apparatus LA includes:
a source collector module SO
an illumination system (illuminator) IL configured to condition a radiation beam B (e.g. EUV radiation).
a support structure (e.g. a mask table) MT constructed to support a patterning device (e.g. a mask or a reticle) MA and connected to a first positioner PM configured to accurately position the patterning device;
a substrate table (e.g. a wafer table) WT constructed to hold a substrate (e.g. a resist coated wafer) W and connected to a second positioner PW configured to accurately position the substrate; and
a projection system (e.g. a reflective projection system) PS configured to project a pattern imparted to the radiation beam B by patterning device MA onto a target portion C (e.g. comprising one or more dies) of the substrate W.
As here depicted, the apparatus LA is of a reflective type (e.g. employing a reflective mask). It is to be noted that because most materials are absorptive within the EUV wavelength range, the mask may have multilayer reflectors comprising, for example, a multi-stack of Molybdenum and Silicon. In one example, the multi-stack reflector has a 40 layer pairs of Molybdenum and Silicon where the thickness of each layer is a quarter wavelength. Even smaller wavelengths may be produced with X-ray lithography. Since most material is absorptive at EUV and x-ray wavelengths, a thin piece of patterned absorbing material on the patterning device topography (e.g., a TaN absorber on top of the multi-layer reflector) defines where features would print (positive resist) or not print (negative resist).
Referring to
In such cases, the laser is not considered to form part of the lithographic apparatus and the radiation beam is passed from the laser to the source collector module with the aid of a beam delivery system comprising, for example, suitable directing mirrors or a beam expander. In other cases the source may be an integral part of the source collector module, for example when the source is a discharge produced plasma EUV generator, often termed as a DPP source.
The illuminator IL may comprise an adjuster for adjusting the angular intensity distribution of the radiation beam. Generally, at least the outer or inner radial extent (commonly referred to as σ-outer and σ-inner, respectively) of the intensity distribution in a pupil plane of the illuminator can be adjusted. In addition, the illuminator IL may comprise various other components, such as facetted field and pupil mirror devices. The illuminator may be used to condition the radiation beam, to have a desired uniformity and intensity distribution in its cross section.
The radiation beam B is incident on the patterning device (e.g., mask) MA, which is held on the support structure (e.g., mask table) MT, and is patterned by the patterning device. After being reflected from the patterning device (e.g. mask) MA, the radiation beam B passes through the projection system PS, which focuses the beam onto a target portion C of the substrate W. With the aid of the second positioner PW and position sensor PS2 (e.g. an interferometric device, linear encoder or capacitive sensor), the substrate table WT can be moved accurately, e.g. so as to position different target portions C in the path of the radiation beam B. Similarly, the first positioner PM and another position sensor PS1 can be used to accurately position the patterning device (e.g. mask) MA with respect to the path of the radiation beam B. Patterning device (e.g. mask) MA and substrate W may be aligned using patterning device alignment marks M1, M2 and substrate alignment marks P1, P2.
The depicted apparatus LA could be used in at least one of the following modes:
1. In step mode, the support structure (e.g. mask table) MT and the substrate table WT are kept essentially stationary, while an entire pattern imparted to the radiation beam is projected onto a target portion C at one time (i.e. a single static exposure). The substrate table WT is then shifted in the X or Y direction so that a different target portion C can be exposed.
2. In scan mode, the support structure (e.g. mask table) MT and the substrate table WT are scanned synchronously while a pattern imparted to the radiation beam is projected onto a target portion C (i.e. a single dynamic exposure). The velocity and direction of the substrate table WT relative to the support structure (e.g. mask table) MT may be determined by the (de-)magnification and image reversal characteristics of the projection system PS.
3. In another mode, the support structure (e.g. mask table) MT is kept essentially stationary holding a programmable patterning device, and the substrate table WT is moved or scanned while a pattern imparted to the radiation beam is projected onto a target portion C. In this mode, generally a pulsed radiation source is employed, and the programmable patterning device is updated as required after each movement of the substrate table WT or in between successive radiation pulses during a scan. This mode of operation can be readily applied to maskless lithography that utilizes programmable patterning device, such as a programmable mirror array of a type as referred to above.
The radiation emitted by the hot plasma 210 is passed from a source chamber 211 into a collector chamber 212 via an optional gas barrier or contaminant trap 230 (in some cases also referred to as contaminant barrier or foil trap) which is positioned in or behind an opening in source chamber 211. The contaminant trap 230 may include a channel structure. Contamination trap 230 may also include a gas barrier or a combination of a gas barrier and a channel structure. The contaminant trap or contaminant barrier 230 further indicated herein at least includes a channel structure, as known in the art.
The collector chamber 211 may include a radiation collector CO which may be a so-called grazing incidence collector. Radiation collector CO has an upstream radiation collector side 251 and a downstream radiation collector side 252. Radiation that traverses collector CO can be reflected off a grating spectral filter 240 to be focused in a virtual source point IF along the optical axis indicated by the dot-dashed line ‘O’. The virtual source point IF is commonly referred to as the intermediate focus, and the source collector module is arranged such that the intermediate focus IF is located at or near an opening 221 in the enclosing structure 220. The virtual source point IF is an image of the radiation emitting plasma 210.
Subsequently the radiation traverses the illumination system IL, which may include a facetted field mirror device 22 and a facetted pupil mirror device 24 arranged to provide a desired angular distribution of the radiation beam 21, at the patterning device MA, as well as a desired uniformity of radiation intensity at the patterning device MA. Upon reflection of the beam of radiation 21 at the patterning device MA, held by the support structure MT, a patterned beam 26 is formed and the patterned beam 26 is imaged by the projection system PS via reflective elements 28, 30 onto a substrate W held by the substrate table WT.
More elements than shown may generally be present in illumination optics unit IL and projection system PS. The grating spectral filter 240 may optionally be present, depending upon the type of lithographic apparatus. Further, there may be more mirrors present than those shown in the figures, for example there may be 1-6 additional reflective elements present in the projection system PS than shown in
Collector optic CO, as illustrated in
Alternatively, the source collector module SO may be part of an LPP radiation system as shown in
The concepts disclosed herein may simulate or mathematically model any generic imaging system for imaging sub wavelength features and may be especially useful with emerging imaging technologies capable of producing increasingly shorter wavelengths. Emerging technologies already in use include EUV (extreme ultra violet), DUV lithography that is capable of producing a 193 nm wavelength with the use of an ArF laser, and even a 157 nm wavelength with the use of a Fluorine laser. Moreover, EUV lithography is capable of producing wavelengths within a range of 20-5 nm by using a synchrotron or by hitting a material (either solid or a plasma) with high energy electrons in order to produce photons within this range.
While the concepts disclosed herein may be used for imaging on a substrate such as a silicon wafer, it shall be understood that the disclosed concepts may be used with any type of lithographic imaging systems, e.g., those used for imaging on substrates other than silicon wafers.
The foregoing paragraphs describe decomposing CD distribution or LCDU data into error contributions from various sources. For example, as described at least with reference to
The present disclosure identifies an error contribution source for a given signal of error contribution values. In some embodiments, a machine learning (ML) model is trained to distinguish between error contributions from various sources, and the trained ML model is used to determine a classification (e.g., an error contribution source), or a label that identifies the error contribution source, of a given signal.
In some embodiments, the signal 2205 can be generated using any of a number of methods, e.g., the ICA method, described at least with reference to
In some embodiments, the training of CNN based classifier model 2250 to determine a classification of an error contribution signal includes adjusting model parameters, such as weights and biases of the CNN, such that a cost function in predicting, determining, or generating the classification is minimized. In some embodiments, adjusting of the model parameter values includes adjusting values of one or more weights of a layer of the CNN, one or more bias of a layer of the CNN, hyperparameters of the CNN and/or a number of layers of the CNN. In some embodiments, the number of layers is a hyperparameter of the CNN which may be pre-selected and may not be changed during the training process. In some embodiments, a series of training process may be performed where the number of layers may be modified.
In some embodiments, training the classifier model 2250 involves determining a value of the cost function and progressively adjusting weights of one or more layers of the CNN such that the cost function is reduced (in some embodiments, minimized or does not reduce beyond a specified threshold). In some embodiments, the cost function is indicative of a difference between a predicted classification 2320 (e.g., of an output vector of CNN) of the input signal 2305 and an actual classification of the input signal 2305 (e.g., specified or provided with the input signal 2305). In some embodiments, the cost function can be a loss function such as binary cross entropy. The cost function is reduced by modifying the values of the CNN model parameters (e.g., weights, bias, stride, etc.). In some embodiments, the cost function is computed as CF=ƒ(predicted classification−CNN(input, cnn_parameters)). In this step, the input to CNN includes an input signal and corresponding actual classification of the input signal, and cnn_parameters, which are weights and biases of CNN, has initial values that may be randomly selected.
In some embodiments, a gradient corresponding to the cost function may be dcost/dparameter, where the cnn_parameters values may be updated based on an equation (e.g., parameter=parameter−learning_rate*gradient). The parameter may be the weight and/or bias, and learning_rate may be a hyper-parameter used to tune the training process and may be selected by a user or a computer to improve convergence (e.g., faster convergence) of the training process.
The classifier model 2250 is trained using labeled training data 2325, which includes multiple error contribution signals such as a first error contribution signal 2305, second error contribution signal 2310 and third error contribution signal 2315, that are representative of error contribution values from multiple sources. Each error contribution signal in the training data 2325 includes (a) error contribution values from a specified source to a set of contact holes printed on the substrate, and (b) a label indicative of the specified source of error contribution (e.g., an actual classification of the error contribution signal). For example, the first error contribution signal 2305 can include (a) a first set of error contribution values associated with a first set of contact holes printed on the substrate, and (b) a label that indicates the source of error contribution as “resist”. Similarly, the second error contribution signal 2310 can include (a) a second set of error contribution values associated with the first set of contact holes printed on the substrate, and (b) a label that indicates the source of error contribution as “mask,” and the third error contribution signal 2315 can include (a) a third set of error contribution values associated with the first set of contact holes printed on the substrate, and (b) a label that indicates the source of error contribution as “SEM.” The training data 2325 can include various such error contribution signals for various contact holes. In some embodiments, the training data 2325 is split into a number of subsets in which each subset includes error contribution signals for a different set of contact holes. For example, a first subset of training data may include three error contribution signals (e.g., one error contribution signal for each source) for a first subset of contact holes, and a second subset of training data includes three error contribution signals (e.g., one error contribution signal for each source) for a second subset of contact holes. The classifier model 2250 is trained by inputting different subsets in different phases of the training.
In some embodiments, training the classifier model 2250 is an iterative process and each iteration may involve inputting different training data (e.g., an error contribution input signal, such as input signal 2305), predicting the classification 2320 for the corresponding error contribution signal, determining the cost function based on the actual classification (e.g., provided in the label) and the predicted classification 2320, and minimizing the cost function. In some embodiments, a first set of iterations is performed with a first subset of training data, then a second set of iterations is performed with a second subset of training data and so on. After several iterations of training (e.g., when cost function is minimized or does not reduce beyond a specified threshold), optimized cnn_parameters values are obtained and further used as the model parameter values of a trained classifier model 2250. The trained classifier model 2250 can then be used to predict a classification for any desired error contribution signal, by using the error contribution signal as input to the trained classifier model 2250, e.g., as described at least with reference to
The training data 2325 can be generated in any of a known number of methods. One such example method of generating the error contribution signals for training the classifier model 2250 is described at least with reference to
At operation 2405, a measurement process is performed to obtain measurement data 2401, such as CD, of a number of contact holes printed on a substrate. The measurements may be obtained from CDU wafer and FEM wafer. The LCDU is decomposed into 3 components: mask noise, resist noise (which includes shot noise) and SEM noise. The measurement process may be designed in accordance with the following principles:
Select “N” contact holes on a reticle
Each contact hole is imaged “M” times in equivalent conditions
Each image (of N*M wafer images of contact holes) is measured “S” times with a SEM
In this experiment N contact holes with the same (intended) dimensions are chosen on a reticle and are typically a part of a contact hole array. The actual sizes of the chosen contact holes on the reticle may vary due to the mask error. The mask errors are translated to the wafer by each exposure and therefore result in a systematic fingerprint of the wafer CD measurements present in each exposure result. The residual random component in the wafer CD variability is due to the resist noise (along with the shot noise) and the SEM noise. To separate the SEM error component all wafer CDs are measured S times (taking S images of each measured location) as summarized in the Table 1.
The CD of a contact hole can be written as:
CDijk=
where
δCDiMASK may be the effect on the substrate of the mask noise present in the reticle contact hole, i, δCDijSN, is the resist noise present along with the shot noise produced by exposure j of contact hole i, and δCDijkSEM is the remaining random noise attributed to SEM error.
After obtaining the measurement data 2401, at operation 2410, the error contributions 2411 are derived from the measurement data 2401 as follows. For example, the following equations represent the error contributors 2411 from sources such as mask, resist and SEM:
The mask noise, δCDiMASK, of i-th contact hole on the reticle is the deviation of substrate CD averaged over all measurements of this contact hole (over all exposures and SEM runs) from the total average CD. The shot noise, δCDijSN, is a factor nested under the mask error factor and its levels dependent on the levels of the mask noise. δCDijSN measures the effect of exposure j of contact hole i. Particularly, for reticle contact hole i, δCDijSN is the deviation of the substrate CD after exposure j from the averaged CD measured for this contact hole (averaged over all exposures and SEM runs). The SEM noise, δCDijkSEM, in the measurement of a particular i-th hole and j-th exposure is the deviation of the k-th measurement from the CD averaged over all measurements of this image.
As can be appreciated, the error contribution values 2411 corresponding to each of the sources is calculated using the Eqs. 3A-5A. The above process 2400 can be used to generate a number of error contributor signals for a number of contact holes, which can be used as the training data 2325 to train the classifier model 2250, e.g., as described at least with reference to
In some embodiments, the training data 2325 is split into a number of subsets in which each subset includes error contribution signals for a different set of contact holes. For example, a first subset of training data 2325 may include three error contribution signals (e.g., one error contribution signal for each source) for a first subset of contact holes, and a second subset of training data 2325 includes three error contribution signals (e.g., one error contribution signal for each source) for a second subset of contact holes.
In some embodiments, the training data 2325 is generated using any of a number of methods, such as the linear nested model described at least with reference to
At operation 2510, the classifier model 2250 is trained based on the training data to predict a classification of each error contributor signal from the training data. In some embodiments, the classifier model 2250 is CNN model. The classifier model 2250 is executed by inputting the first error contribution signal 2305 from the training data 2325. The classifier model 2250 predicts a classification 2320 of the first error contribution signal 2305 (e.g., a source of the error contribution) and computes a cost function that determines a difference between the predicted classification and the actual classification of the first error contribution signal 2305. The training of the classifier model 2250 is an iterative process and is continued (e.g., by inputting different error contribution signals from different subsets of the training data 2325) until the cost function is reduced (e.g., beyond a specified threshold or does not reduce anymore), that is, the predicted classification of any of the error contributor signals from the training data 2325 is similar to the actual classification of the corresponding error contributor signal. Additional details of the training process are described at least with reference to
After the cost function has satisfied a specified criterion (e.g., does not reduce any more, has reduced beyond a specified threshold, or the rate at which it reduced is below a specified threshold), the classifier model 2250 is considered to be trained, and can be used to predict a classification for any desired error contribution signal, e.g., as described at least with reference to
At operation 2555, the classifier model 2250 is executed by inputting a reference error contribution signal, such as the first error contribution signal 2305, to output a predicted classification of the reference error contribution signal, such as the predicted classification 2320 of the first error contribution signal 2305.
At operation 2560, a cost function of the classifier model 2250 is computed, e.g., as the difference between the predicted classification and the actual classification. For example, a cost function 2561 is determined as the difference between the predicted classification 2320 and the actual classification of the first error contribution signal 2305. In some embodiments, the actual classification, which is a source of the error contribution for the first error contribution signal 2305, is provided as a label with the first error contribution signal 2305.
At operation 2565, the classifier model 2250 is adjusted such that the cost function 2561 is reduced. In some embodiments, adjusting the classifier model 2250 to reduce the cost function 2561 includes adjusting model parameters, such as weights and biases of the classifier model 2250 (e.g., parameters of the CNN model).
At operation 2570, a determination is made whether the cost function 2561 is reduced (e.g., does not reduce any more, has reduced beyond a specified threshold, or the rate at which it reduced is below a specified threshold).
If the cost function 2561 is reduced, the classifier model 2250 is considered to be trained and the process returns to operation 2510 of process 2500. However, if the cost function 2561 has not reduced, the operations 2555-2570 are repeated with different error contribution signals from the training data 2325 until the cost function 2561 is reduced. For example, a first set of iterations may be performed by inputting a first subset of training data, which includes three error contribution signals (e.g., one error contribution signal for each source) for a first subset of contact holes, then a second set of iterations is performed with a second subset of training data, which includes three error contribution signals (e.g., one error contribution signal for each source) for a second subset of contact holes, and so on until the cost function 2561 is reduced.
At operation 2610, the error contribution signal 2205 is input to a trained classifier model 2250 to determine a classification 2225, which is indicative of a source of the error contribution values in the error contribution signal 2205. The classifier model 2250 may output the classification 2225 value in any of a number of formats. In some embodiments, the classification 2225 may be output as a probability value (e.g., 0.0 to 1.0) that is indicative of a probability that the error contribution values in the signal 2205 is from a specified source. For example, the classification 2225 value can be “PRESIST=0.98,” which indicates that there is a “98%” probability that the error contribution values in the error contribution signal 2205 is resist noise. In some embodiments, the classification 2225 value can indicate a probability of the error contribution values being from each of the sources. For example, the classification 2225 value can be “PRESIST=0.98,” “PMASK=0.015,” and “PSEM=0.005,” which indicates that there is a “98%” probability that the error contribution values in the signal 2205 is resist noise, “1.5%” probability that the error contribution values in the signal 2205 is mask noise, and a “0.5%” probability that the error contribution values in the signal 2205 is SEM noise. In some embodiments, the classifier model 2250 may be configured to determine the source of the error contribution as the source which has the highest probability.
The present disclosure determines error contributions from multiple sources using a ML model. The ML model is trained to predict error contributions from various sources for a given feature. For example, an image of a feature (e.g., contact hole) is provided as an input to the ML model, and the ML model predicts error contributions from various sources for the input feature. Details of training the ML model are described at least with references to
At operation 2705, multiple datasets in which each dataset includes image data of a feature of a pattern printed on a substrate and error contribution data having error contribution values that are representative of error contributions to the feature from the different sources are obtained as training data 2810. For example, a first dataset 2815, may include a first image data 2816 of a first feature of the pattern (e.g., contact hole) and a first error contribution data 2817 having error contribution values representative of error contributions to the first feature from multiple sources, such as a mask, resist and SEM. The first image data 2816 may include an image of the first feature. The images of the features may be obtained using an inspection tool, such as a SEM. For example, the first error contribution data 2817 may include δCDMASK, CDRESIST, and δCDSEM values as error contributions from the sources mask, resist and SEM, respectively. As described above at least with reference to Eq. 1, δCD is a deviation of a CD value of a given feature from a mean of CD values of a number of features. The error contribution values may be obtained using measurement data of the features, such as CD. For example, the error contribution values may be obtained using a linear nested model, as described at least with reference to
At operation 2710, the training data 2810 is provided as an input to the error contribution model 2805 for training the error contribution model 2805 to predict error contribution data from the training data. The training of the error contribution model 2805 is an iterative process and is continued (e.g., by inputting the same datasets or different subsets of datasets of the training data 2810) until a cost function is reduced (e.g., beyond a specified threshold or does not reduce anymore). Additional details of the training process are described at least with reference to
At operation 2755, the error contribution model 2805 is executed by inputting a reference dataset, such as the first dataset 2815, to output predicted error contribution data 2820 having error contribution values for the reference dataset. In some embodiments, the predicted error contribution data 2820 may be a set of error contribution values, such as δCDMASK, δCDRESIST, and δCDSEM.
At operation 2760, a cost function of the error contribution model 2805 is computed, e.g., as the difference between the predicted error contribution data 2820 and the actual error contribution data associated with the reference dataset. For example, a cost function 2761 is determined as the difference between the predicted set of error contribution values in the predicted error contribution data 2820 and the set of error contribution values from the first error contribution data 2817. In some embodiments, the set of error contribution values from the first error contribution data 2817, is provided as a label with the first image data 2816.
At operation 2765, the error contribution model 2805 is adjusted such that the cost function 2761 is reduced. In some embodiments, adjusting the error contribution model 2805 to reduce the cost function 2761 includes adjusting model parameters, such as weights and biases of the error contribution model 2805.
At operation 2770, a determination is made whether the cost function 2761 has satisfied training criteria (e.g., cost function does not reduce any more, has reduced beyond a specified threshold, or the rate at which it reduced is below a specified threshold).
If the cost function 2761 has satisfied the training criteria, the error contribution model 2805 is considered to be trained and the process returns to operation 2710 of process 2700. However, if the cost function 2761 has not reduced, the operations 2755-2770 are repeated with different datasets or the same datasets from the training data 2810 until the cost function 2761 is reduced. For example, a first set of iterations may be performed by inputting a first subset of training data 2810 for a first subset of contact holes, then a second set of iterations may be performed with a second subset of training data 2810 for a second subset of contact holes, and so on until the cost function 2761 is reduced.
At operation 2910, the error contribution model 2805 is executed with the image data 3005 to generate a prediction of error contribution data 3025. The error contribution data 3025 may include error contribution values that are representative of error contributions from multiple sources to the feature in the image data 3005. For example, the predicted error contribution data 3025 may include a set of error contribution values, such as δCDMASK, δCDRESIST, and δCDSEM, that are error contributions from sources such as mask, resist and SEM, respectively.
While the foregoing paragraphs describe predicting error contributions in terms of δCD, the error contribution model 2805 may also be used to predict error contributions in terms of LCDU. For example, the error contributions from sources such as mask, resist and SEM, to the LCDU of a feature may be represented as such as LCDUMASK, LCDURESIST, and LCDUSEM, respectively. The error contribution model 2805 may be trained using LCDU values instead of the δCD values. For example, in the process 2700 of training the error contribution model 2805, each of the datasets in the training data 2810 may include a number of images and a set of LCDU values as error contribution values. For example, the first dataset 2815 may include a number of images corresponding to a number of features (e.g., contact holes) as the image data 2816, and a set of LCDUMASK, LCDURESIST, and LCDUSEM values that are representative of error contributions to the LCDU of the features from various sources as the error contribution data 2817. In some embodiments, similar to the δCD values, the LCDU error contribution values may be obtained from the linear nested model, as described at least with reference to
Further, while the foregoing paragraphs describe generating a prediction of error contribution values for a feature, the error contribution model 2805 may also be used to predict error contributions for multiple measurement points on the feature. For example, the error contribution model 2805 may predict a first set of error contribution values (e.g., δCD1MASK, δCD1RESIST, and δCD1SEM) for a first measurement point on the feature and a second set of error contribution values (e.g., δCD2MASK, δCD2RESIST, and δCD2SEM) for a second measurement point and so on. The error contribution model 2805 may be trained using multiple sets of error contribution values instead of a single set of error contribution values per feature. For example, in the process 2700 of training the error contribution model 2805, each of the datasets in the training data 2810 may include an image of a feature and multiple sets of error contribution values in which each set of error contribution values corresponds to a single measurement point on the feature. For example, if the number of measurement points, “n,” is “20”, then the first dataset 2815 may include an image of the first feature as the image data 2816, and the error contribution data 2817 may include “20” sets of error contribution values—one set for each of the “20” measurement points. During the prediction process, an image of a feature for which a prediction of the error contribution values is to be generated is input as the image data 3005 to the trained error contribution model 2805. The trained error contribution model 2805 generates a prediction of “n” sets of error contribution values as the error contribution data 3025 in which each set of error contribution values corresponds to one of “n” measurement points on the feature. The error contribution model 2805 may be configured in a number of ways to predict error contribution values for “n” measurement points on the feature. For example, a dense layer in the neural network model used to implement the error contribution model 2805 may be configured to generate n*m values, where n is the number of measurement points on the feature and m is the number of sources contribution to the error (e.g., “3” for sources such as mask, resist and SEM). In another example, an image of the feature may be encoded (e.g., using a neural network encoder) into the n*m values, which may be input as training data to the error contribution model 2805 to train the error contribution model 2805 to generate a prediction of error contribution values for each of n measurement points on a feature.
The embodiments may further be described using the following clauses:
1. A non-transitory computer-readable medium having instructions that, when executed by a computer, cause the computer to execute a method for decomposing error contributions from multiple sources to multiple features of a pattern printed on a substrate, the method comprising: obtaining an image of the pattern on the substrate;
obtaining, using the image, a plurality of measurement values of a feature of the pattern, wherein the measurement values are obtained for different sensor values;
correlating, using a decomposition method, each measurement value of the plurality of measurement values to a linear mixture of the error contributions to generate a plurality of linear mixtures of the error contributions; and deriving, from the linear mixtures and using the decomposition method, each of the error contributions.
2. The computer-readable medium of clause 1, wherein the different sensor values correspond to different threshold values associated with the image in which each threshold value corresponds to a threshold of a pixel value in the image.
3. The computer-readable medium of clause 2, wherein each measurement value corresponds to a critical dimension (CD) value of the feature at one of the different threshold values.
4. The computer-readable medium of clause 2, wherein the error contributions include:
an image acquisition tool error contribution that is associated with an image acquisition tool used to acquire the image,
a mask error contribution that is associated with a mask used to print the pattern on the substrate, and
a resist error contribution that is associated with a resist used to print the pattern, wherein the resist error contribution includes photoresist chemical noise and a shot noise associated with a source of a lithographic apparatus used to print the pattern.
5. The computer-readable medium of clause 4 further comprising:
adjusting, based on the mask error contribution, one or more parameters of at least one of the mask or a source of a lithographic apparatus used to print the pattern.
6. The computer-readable medium of clause 4 further comprising:
adjusting, based on the resist error contribution, one or more parameters of at least one of the mask or a source of a lithographic apparatus used to print the pattern.
7. The computer-readable medium of any of clauses 3-6, wherein obtaining the measurement values includes:
obtaining a first signal having a first plurality of delta CD values from a plurality of measurement points at a first threshold value of the different threshold values,
obtaining a second signal having a second plurality of delta CD values from the plurality of measurement points at a second threshold value of the different threshold values, and
obtaining a third signal having a third plurality of delta CD values from the plurality of measurement points at a third threshold value of the different threshold values.
8. The computer-readable medium of clause 7, wherein each delta CD value is determined per threshold value and per measurement point, and indicates a deviation of a CD value of a given feature from a mean value of a plurality of CD values of the features.
9. The computer-readable medium of clause 7, wherein each delta CD value indicates, at a given threshold value, a distance between a specified point on a contour of a given feature to a reference point on a reference contour of the given feature, wherein the reference contour is a simulated version of the contour of the given feature.
10. The computer-readable medium of clause 7, wherein correlating each measurement value includes:
correlating each of the first plurality of delta CD values in the first signal to a first linear mixture of the image acquisition tool, mask and resist error contributions,
correlating each of the second plurality of delta CD values in the second signal to a second linear mixture of the image acquisition tool, mask and resist error contributions, and
correlating each of the third plurality of delta CD values in the third signal to a third linear mixture of the image acquisition tool, mask and resist error contributions.
11. The computer-readable medium of clause 10, wherein deriving each of the error contributions includes:
deriving, using the first, second, and third linear mixtures, and from each of the first plurality, second plurality, and third plurality of delta CD values: (a) a first output signal having a plurality of the image acquisition tool error contributions, (b) a second output signal having a plurality of the mask error contributions, and (c) a third output signal having a plurality of the resist error contributions.
12. The computer-readable medium of clause 11, wherein each error contribution is determined as a function of the corresponding error contribution at the first, second and third threshold levels.
13. The computer-readable medium of clause 11, wherein deriving each of the error contributions includes:
determining a mixing matrix having a set of coefficients that generates the first, second and third linear mixtures of the error contributions corresponding to each delta CD value from the first plurality, second plurality and third plurality of delta CD values, respectively,
determining an inverse of the mixing matrix, and
using the inverse of the mixing matrix, determining (a) the first output signal having the plurality of the image acquisition tool error contributions, (b) the second output signal having the plurality of the mask error contributions, and (c) the third output signal having the plurality of the resist error contributions, from the first plurality, second plurality and third plurality of delta CD values, respectively.
14. The computer-readable medium of any of clauses 2-3, wherein the obtaining the measurement values includes:
obtaining a first contour of the feature corresponding to a first threshold value of the different threshold values,
obtaining a first CD value of the first contour,
obtaining a second contour of the feature corresponding to a second threshold value of the different threshold values, and
obtaining a second CD value of the second contour.
15. The computer-readable medium of clause 14 further comprising:
obtaining a first delta CD value of the first CD value, wherein the first delta CD indicates a deviation of the first CD value from a mean of a plurality of first CD values measured at a plurality of measurement points at the first threshold value.
16. The computer-readable medium of clause 15, wherein obtaining the first delta CD value includes: obtaining the plurality of first CD values corresponding to the first threshold value at the plurality of measurement points,
obtaining a mean value of the plurality of first CD values,
shifting the mean value to a zero value, and
obtaining the first delta CD value as a difference between the first CD value and the mean value.
17. The computer-readable medium of clause 15, wherein the plurality of measurement points are located on at least one of (a) the feature or (b) a plurality of features of the pattern.
18. The computer-readable medium of any of clauses 15-17, wherein correlating each measurement value includes:
correlating the first delta CD value corresponding to the first threshold value to a first linear mixture of a first error contribution and a second error contribution of the error contributions, and
correlating a second delta CD value corresponding to the second threshold value to a second linear mixture of the first and second error contributions.
19. The computer-readable medium of clause 18, wherein deriving each of the error contributions includes:
deriving, using the decomposition method, the first and second error contributions from the first and second delta CD values, and the first and second linear mixtures.
20. The computer-readable medium of clause 1, wherein the measurement values correspond to a local critical dimension uniformity (LCDU) value of the feature for the different sensor values.
21. The computer-readable medium of any of clauses 1 and 20, wherein the different sensor values correspond to different dose levels associated with a source of a lithographic apparatus used to print the pattern.
22. The computer-readable medium of any of clauses 1 and 20, wherein the different sensor values correspond to different focus levels associated with a source of a lithographic apparatus used to print the pattern.
23. The computer-readable medium of any of clauses 20-21 further comprising:
obtaining, based on a specified focus level, a first LCDU value corresponding to a first dose level, and
obtaining, based on the specified focus level, a second LCDU value corresponding to a second dose level.
24. The computer-readable medium of any of clauses 20 or 22 further comprising:
obtaining, based on a specified dose level, a first LCDU value corresponding to a first focus level, and
obtaining, based on the specified dose level, a second LCDU value corresponding to a second focus level.
25. The computer-readable medium of any of clauses 23 or 24, wherein correlating each measurement value includes:
correlating the first LCDU value to a first linear mixture of a first error contribution of the error contributions and a second error contribution of the error contributions, and
correlating the second LCDU value to a second linear mixture of the first and second error contributions.
26. The computer-readable medium of clause 25, wherein deriving each of the error contributions includes:
deriving, using the decomposition method, the first and second error contributions from the first and second LCDU values, and the first and second linear mixtures.
27. The computer-readable medium of clause 1, wherein the measurement values correspond to a line width roughness (LWR) value of the feature for the different sensor values.
28. A non-transitory computer-readable medium having instructions that, when executed by a computer, cause the computer to execute a method for decomposing error contributions from multiple sources to multiple features associated with a pattern printed on a substrate, the method comprising: obtaining an image of the pattern;
obtaining a plurality of delta critical dimension (CD) values at different heights of contours of the features of the pattern, wherein the plurality of delta CD values includes (a) a first set of delta CD values of the features corresponding to a first contour height, (b) a second set of delta CD values of the features corresponding to a second contour height, and (c) a third set of delta CD values of the features corresponding a third contour height;
correlating, using a decomposition method, (a) the first set of delta CD values to a first linear mixture of a first, second, and third error contributions, (b) the second set of delta CD values to a second linear mixture of the first, second, and third error contributions, (c) the third set of delta CD values to a third linear mixture of the first, second, and third error contributions; and
deriving, from the linear mixtures and using the decomposition method, the first, second, and third error contributions.
29. The computer-readable medium of clause 28, wherein each delta CD value indicates a deviation of a CD value of a feature from a mean value of a plurality of CD values of the features measured at a plurality of measurement points at a specified contour height.
30. The computer-readable medium of clause 28, wherein each delta CD value indicates, at a given contour height, a distance between a specified point on a contour of a feature to a reference point on a reference contour of the feature, wherein the reference contour is a simulated version of the contour of the given feature.
31. The computer-readable medium of clause 28, wherein each contour height is determined by thresholding pixel values of the image to a specified value.
32. The computer-readable medium of clause 28 further comprising:
adjusting, based on one or more of the error contributions, one or more parameters of at least one of a mask or a source of a lithographic apparatus used to print the pattern.
33. A non-transitory computer-readable medium having instructions that, when executed by a computer, cause the computer to execute a method for decomposing error contributions from multiple sources to multiple features associated with a pattern on a substrate, the method comprising:
obtaining local critical dimension uniformity (LCDU) data associated with the pattern, wherein the LCDU data includes, for a specified focus level of a source of a lithographic apparatus used to print the pattern, (a) a first set of LCDU values of the features of the pattern corresponding to a first dose level of the source, (b) a second set of LCDU values of the features corresponding to a second dose level, and (c) a third set of LCDU values of the features corresponding a third dose level;
correlating, using a decomposition method, (a) the first set of LCDU values to a first linear mixture of a first, second, and third error contributions, (b) the second set of LCDU values to a second linear mixture of the first, second, and third error contributions, and (c) the third set of LCDU values to a third linear mixture of the first, second, and third error contributions; and deriving, from the linear mixtures and using the decomposition method, the first, second, and third error contributions.
34. A method for decomposing error contributions from multiple sources to multiple features associated with a pattern to be printed on a substrate, the method comprising:
obtaining an image of the pattern on the substrate;
obtaining, using the image, a plurality of measurement values of a feature of the pattern, wherein the measurement values correspond to different threshold values associated with the image;
correlating, using a decomposition method, each measurement value of the plurality of measurement values to a linear mixture of the error contributions to generate a plurality of linear mixtures of the error contributions; and
deriving, from the linear mixtures and using the decomposition method, each of the error contributions.
35. The method of clause 34, wherein each measurement value corresponds to a critical dimension (CD) value of the feature at one of the different threshold values.
36. The method of clause 35, wherein each threshold value corresponds to a threshold of a pixel value in the image.
37. The method of any of clauses 35-36, wherein the error contributions includes:
a first, second and third error contributions to the CD value, wherein the first error contribution is from a resist used to print the pattern, the second error contribution is from a mask used to print the pattern on the substrate, and the third error contribution is from an image acquisition tool used to acquire the image.
38. A method for decomposing error contributions from multiple sources to one or more features associated with a pattern printed on a substrate, the method comprising:
obtaining local critical dimension uniformity (LCDU) data associated with the pattern, wherein the LCDU data includes, for a specified focus level of a source of a lithographic apparatus used to print the pattern, (a) a first set of LCDU values of the some or more features of the pattern corresponding to a first dose level of the source, (b) a second set of LCDU values of the one or more features corresponding to a second dose level, and (c) a third set of LCDU values of the one or more features corresponding a third dose level;
correlating, using a decomposition method, (a) the first set of LCDU values to a first linear mixture of a first, second, and third error contributions, (b) the second set of LCDU values to a second linear mixture of the first, second, and third error contributions, and (c) the third set of LCDU values to a third linear mixture of the first, second, and third error contributions; and deriving, from the linear mixtures and using the decomposition method, the first, second, and third error contributions.
39. An apparatus for decomposing error contributions from multiple sources to multiple features of a pattern printed on a substrate, the apparatus comprising:
a memory storing a set of instructions; and
at least one processor configured to execute the set of instructions to cause the apparatus to perform a method of:
obtaining an image of the pattern on the substrate;
obtaining, using the image, a plurality of measurement values of a feature of the pattern, wherein the measurement values are obtained for different sensor values;
correlating, using a decomposition method, each measurement value of the plurality of measurement values to a linear mixture of the error contributions to generate a plurality of linear mixtures of the error contributions; and
deriving, from the linear mixtures and using the decomposition method, each of the error contributions.
40. The apparatus of clause 39, wherein the different sensor values correspond to different threshold values associated with the image in which each threshold value corresponds to a threshold of a pixel value in the image.
41. The apparatus of clause 40, wherein each measurement value corresponds to a critical dimension (CD) value of the feature at one of the different threshold values.
42. The apparatus of clause 40, wherein the error contributions include:
an image acquisition tool error contribution that is associated with an image acquisition tool used to acquire the image,
a mask error contribution that is associated with a mask used to print the pattern on the substrate, and
a resist error contribution that is associated with a resist used to print the pattern, wherein the resist error contribution includes photoresist chemical noise and a shot noise associated with a source of a lithographic apparatus used to print the pattern.
43. The apparatus of clause 42 further comprising:
adjusting, based on the mask error contribution, one or more parameters of at least one of the mask or a source of a lithographic apparatus used to print the pattern.
44. The apparatus of clause 42 further comprising:
adjusting, based on the resist error contribution, one or more parameters of at least one of the mask or a source of a lithographic apparatus used to print the pattern.
45. The apparatus of any of clauses 41-44, wherein obtaining the measurement values includes: obtaining a first signal having a first plurality of delta CD values from a plurality of measurement points at a first threshold value of the different threshold values,
obtaining a second signal having a second plurality of delta CD values from the plurality of measurement points at a second threshold value of the different threshold values, and
obtaining a third signal having a third plurality of delta CD values from the plurality of measurement points at a third threshold value of the different threshold values.
46. The apparatus of clause 45, wherein each delta CD value is determined per threshold value and per measurement point, and indicates a deviation of a CD value of a given feature from a mean value of a plurality of CD values of the features.
47. The apparatus of clause 45, wherein each delta CD value indicates, at a given threshold value, a distance between a specified point on a contour of a given feature to a reference point on a reference contour of the given feature, wherein the reference contour is a simulated version of the contour of the given feature.
48. The apparatus of clause 45, wherein correlating each measurement value includes:
correlating each of the first plurality of delta CD values in the first signal to a first linear mixture of the image acquisition tool, mask and resist error contributions,
correlating each of the second plurality of delta CD values in the second signal to a second linear mixture of the image acquisition tool, mask and resist error contributions, and
correlating each of the third plurality of delta CD values in the third signal to a third linear mixture of the image acquisition tool, mask and resist error contributions.
49. The apparatus of clause 48, wherein deriving each of the error contributions includes: deriving, using the first, second, and third linear mixtures, (a) a first output signal having a plurality of the image acquisition tool error contributions, (b) a second output signal having a plurality of the mask error contributions, and (c) a third output signal having a plurality of the resist error contributions from each of the first plurality, second plurality, and third plurality of delta CD values.
50. The apparatus of clause 49, wherein deriving each of the error contributions includes: deriving each of the error contributions using independent component analysis (ICA) method.
51. The apparatus of clause 50, wherein deriving each of the error contributions using the ICA method includes:
determining a mixing matrix having a set of coefficients that generates the first, second and third linear mixtures of the error contributions corresponding to each delta CD value from the first plurality, second plurality and third plurality of delta CD values, respectively,
determining an inverse of the mixing matrix, and
using the inverse of the mixing matrix, determining (a) the first output signal having the plurality of the image acquisition tool error contributions, (b) the second output signal having the plurality of the mask error contributions, and (c) the third output signal having the plurality of the resist error contributions, from the first plurality, second plurality and third plurality of delta CD values, respectively.
52. The apparatus of clause 49, wherein deriving each of the error contributions includes: deriving each of the error contributions using reconstruction ICA method, or orthonormal ICA method.
53. The apparatus of any of clauses 40-41, wherein obtaining the measurement values includes: obtaining a first contour of the feature corresponding to a first threshold value of the different threshold values,
obtaining a first CD value of the first contour,
obtaining a second contour of the feature corresponding to a second threshold value of the different threshold values, and
obtaining a second CD value of the second contour.
54. The apparatus of clause 53 further comprising:
obtaining a first delta CD value of the first CD value, wherein the first delta CD indicates a deviation of the first CD value from a mean of a plurality of first CD values measured at a plurality of measurement points at the first threshold value.
55. The apparatus of clause 54, wherein obtaining the first delta CD value includes:
obtaining the plurality of first CD values corresponding to the first threshold value at the plurality of measurement points,
obtaining a mean value of the plurality of first CD values,
shifting the mean value to a zero value, and
obtaining the first delta CD value as a difference between the first CD value and the mean value.
56. The apparatus of clause 55, wherein the plurality of measurement points are located on at least one of (a) the feature or (b) a plurality of features of the pattern.
57. The apparatus of any of clauses 53-55, wherein correlating each measurement value includes: correlating the first delta CD value corresponding to the first threshold value to a first linear mixture of a first error contribution and a second error contribution of the error contributions, and
correlating a second delta CD value corresponding to the second threshold value to a second linear mixture of the first and second error contributions.
58. The apparatus of clause 57, wherein deriving each of the error contributions includes:
deriving, using the decomposition method, the first and second error contributions from the first and second delta CD values, and the first and second linear mixtures.
59. The apparatus of clause 39, wherein the measurement values correspond to a local critical dimension uniformity (LCDU) value of the feature for the different sensor values.
60. The apparatus of any of clauses 39 and 59, wherein the different sensor values correspond to different dose levels associated with a source of a lithographic apparatus used to print the pattern.
61. The apparatus of any of clauses 39 and 59, wherein the different sensor values correspond to different focus levels associated with a source of a lithographic apparatus used to print the pattern.
62. The apparatus of any of clauses 59-60 further comprising:
obtaining, based on a specified focus level, a first LCDU value corresponding to a first dose level, and
obtaining, based on the specified focus level, a second LCDU value corresponding to a second dose level.
63. The apparatus of any of clauses 59 or 61 further comprising:
obtaining, based on a specified dose level, a first LCDU value corresponding to a first threshold value of a focus level, and
obtaining, based on the specified dose level, a second LCDU value corresponding to a second threshold value of the focus level.
64. The apparatus of any of clauses 62 or 63, wherein correlating each measurement value includes: correlating the first LCDU value to a first linear mixture of a first error contribution of the error contributions and a second error contribution of the error contributions, and correlating the second LCDU value to a second linear mixture of the first and second error contributions.
65. The apparatus of clause 64, wherein deriving each of the error contributions includes: deriving, using the decomposition method, the first and second error contributions from the first and second LCDU values, and the first and second linear mixtures.
66. The apparatus of clause 39, wherein the measurement values correspond to a line width roughness (LWR) value of the feature for the different sensor values.
67. A computer program product comprising a non-transitory computer readable medium having instructions recorded thereon, the instructions when executed by a computer implementing the method of any of the above clauses.
68. A non-transitory computer-readable medium having instructions that, when executed by a computer, cause the computer to execute a method for training a machine learning model to determine a source of error contribution to multiple features of a pattern printed on a substrate, the method comprising:
obtaining training data having multiple datasets, wherein each dataset has error contribution values representative of an error contribution from one of multiple sources to the features, and wherein each dataset is associated with an actual classification that identifies a source of the error contribution of the corresponding dataset; and
training, based on the training data, a machine learning model to predict a classification of a reference dataset of the datasets such that a cost function that determines a difference between the predicted classification and the actual classification of the reference dataset is reduced.
69. The computer-readable medium of clause 68, wherein obtaining the training data includes: obtaining local critical dimension uniformity (LCDU) data associated with the features using different focus and dose level values of an apparatus used for printing the pattern.
70. The computer-readable medium of clause 69, wherein obtaining the training data includes: decomposing LCDU data associated with the features to derive the error contribution values from each of the multiple sources.
71. The computer-readable medium of clause 68, wherein obtaining the training data includes:
generating (a) a first dataset of the training data having error contribution values representative of an error contribution from a first source of the multiple sources, (b) a second dataset of the training data having error contribution values representative of an error contribution from a second source of the multiple sources, and (c) a third dataset of the training data having error contribution values representative of an error contribution from a third source of the multiple sources, and
associating (d) the first dataset with a first classification that identifies the first source as the source of error contribution, (e) the second dataset with a second classification that identifies the second source as the source of error contribution, and (f) the third dataset with a third classification that identifies the third source as the source of error contribution.
72. The computer-readable medium of clause 71, wherein the first source is an image acquisition tool used to acquire an image of the pattern, wherein the second source is a mask used to print the pattern on the substrate, and wherein the third source is a resist used to print the pattern together with a photon shot noise of an apparatus used to print the patter on the substrate.
73. The computer-readable medium of clause 71, wherein generating the first dataset includes:
generating multiple groups of the first, second and third datasets, wherein each group includes error contribution values representative of an error contribution from the first, second, and third sources, respectively, for a different subset of the features.
74. The computer-readable medium of clause 68, wherein training the machine learning model is an iterative process in which each iteration includes:
(a) executing the machine learning model, using the training data, to output the predicted classification of the reference dataset,
(b) determining the cost function as the difference between the predicted classification and the actual classification,
(c) adjusting the machine learning model,
(d) determining whether the cost function is reduced as a result of the adjusting, and
(e) responsive to the cost function not being reduced, repeating steps (a), (b), (c) and (d).
75. The computer-readable medium of any of clauses 68-74, wherein the machine learning model is a convolutional neural network.
76. The computer-readable medium of clause 68 further comprising:
receiving a specified dataset having error contribution values representative of an error contribution from one of the multiple sources to a set of features of a specified pattern printed on a specified substrate; and
executing the machine learning model to determine a classification associated with the specified dataset, wherein the classification identifies a specified source of the multiple sources as the source of error contribution for the error contribution values in the specified dataset.
77. The computer-readable medium of clause 76, wherein receiving the specified dataset includes: decomposing, using a decomposition method, multiple measurement values associated with the set of features to derive a collection of datasets representative of error contributions from each of the multiple sources, wherein the specified dataset is one of the collection of datasets and corresponds to error contribution from one of the multiple sources.
78. The computer-readable medium of clause 77, wherein decomposing the measurement values includes:
obtaining an image of the specified pattern;
obtaining, using the image, the measurement values, wherein the measurement values are obtained for different sensor values;
correlating, using the decomposition method, each measurement value of the measurement values to a linear mixture of the error contributions to generate a plurality of linear mixtures of the error contributions; and
deriving, from the linear mixtures and using the decomposition method, each of the error contributions.
79. The computer-readable medium of clause 78, wherein the different sensor values correspond to different threshold levels associated with the image, wherein each measurement value corresponds to a delta critical dimension (CD) value of a feature of the set of features at one of the different threshold values, wherein the delta CD value indicates a deviation of a CD value of the feature from a mean value of a plurality of CD values of the set of features.
80. The computer-readable medium of any of clauses 79, wherein each threshold value of the different threshold values corresponds to a threshold of a pixel value in the image.
81. The computer-readable medium of clause 78, wherein the measurement values correspond to LCDU value of the feature at the different sensor values.
82. The computer-readable medium of clause 81, wherein the different sensor values correspond to different dose levels associated with a source of a lithographic apparatus used to print the pattern.
83. The computer-readable medium of clause 81, wherein the different sensor values correspond to different focus levels associated with a source of a lithographic apparatus used to print the pattern.
84. The computer-readable medium of any of clauses 78-83, wherein deriving each of the error contributions includes:
deriving each of the error contributions using independent component analysis (ICA) method as the decomposition method.
85. A non-transitory computer-readable medium having instructions that, when executed by a computer, cause the computer to execute a method for determining a source of error contribution to multiple features of a pattern printed on a substrate, the method comprising:
processing one or more images of the pattern to obtain a collection of datasets, wherein each dataset in the collection of datasets has error contribution values representative of an error contribution from one of multiple sources to the features;
inputting a specified dataset of the multiple datasets to a machine learning model; and
executing the machine learning model to determine a classification associated with the specified dataset, wherein the classification identifies a specified source of the multiple sources as the source of error contribution for the error contribution values in the specified dataset.
86. The computer-readable medium of any of clauses 85, wherein executing the machine learning model to determine the classification includes:
training the machine learning model using multiple datasets to determine the classification of the specified dataset, wherein each dataset of the multiple datasets includes error contribution values representative of an error contribution from one of the multiple sources to the features, and wherein each dataset is associated with an actual classification that identifies a source of the error contribution for error contribution values of the corresponding dataset.
87. The computer-readable medium of any of clauses 86, wherein training the machine learning model includes:
training the machine learning model to determine a predicted classification of a reference dataset of the datasets such that a cost function that determines a difference between the predicted classification and an actual classification of the reference dataset is reduced.
88. The computer-readable medium of clause 87, wherein training the machine learning model is an iterative process in which each iteration includes:
(a) executing the machine learning model, using the multiple datasets, to output the predicted classification of the reference dataset,
(b) determining the cost function as the difference between the predicted classification and the actual classification,
(c) adjusting the machine learning model,
(d) determining whether the cost function is reduced as a result of the adjusting, and
(e) responsive to the cost function not being reduced, repeating steps (a), (b), (c) and (d).
89. The computer-readable medium of clause 86, wherein training the machine learning model includes:
generating (a) a first dataset of the multiple datasets having error contribution values representative of an error contribution from a first source of the multiple sources, (b) a second dataset of the multiple datasets having error contribution values representative of an error contribution from a second source of the multiple sources, and (c) a third dataset of the multiple datasets having error contribution values representative of an error contribution from a third source of the multiple sources, and
associating (d) the first dataset with a first classification that identifies the first source as the source of error contribution, (e) the second dataset with a second classification that identifies the second source as the source of error contribution, and (f) the third dataset with a third classification that identifies the third source as the source of error contribution.
90. The computer-readable medium of clause 89, wherein generating the first dataset includes: generating multiple groups of the first, second and third datasets, wherein each group includes error contribution values representative of an error contributions from the first, second, and third sources, respectively, for a different subset of the features.
91. The computer-readable medium of clause 90 further comprising:
training the machine learning model by inputting one group of the first, second and third datasets, after another group of the first, second and third datasets.
92. The computer-readable medium of clause 85, wherein processing the one or more images to obtain the collection of datasets includes:
obtaining a plurality of delta critical dimension (CD) values at different heights of contours of the features, wherein the plurality of delta CD values includes (a) a first set of delta CD values of the features corresponding to a first contour height, (b) a second set of delta CD values of the features corresponding to a second contour height, and (c) a third set of delta CD values of the features corresponding a third contour height;
correlating, using a decomposition method, (a) the first set of delta CD values to a first linear mixture of the error contribution from the multiple sources, (b) the second set of delta CD values to a second linear mixture of the error contribution from the multiple sources, (c) the third set of delta CD values to a third linear mixture of the error contribution from the multiple sources; and
deriving, from the linear mixtures and using the decomposition method, the error contribution from each of the sources,
wherein a first dataset of the collection of datasets includes error contribution values representative of an error contribution from a first one of the multiple sources,
wherein a second dataset of the collection of datasets includes error contribution values representative of an error contribution from a second one of the multiple sources, and
wherein a third dataset of the collection of datasets includes error contribution values representative of an error contribution from a third one of the multiple sources.
93. The computer-readable medium of clause 92, wherein each contour height is determined by thresholding pixel values of the one or more images to a specified value.
94. The computer-readable medium of clause 85, wherein processing the one or more images to obtain the collection of datasets includes:
obtaining local critical dimension uniformity (LCDU) data associated with the pattern, wherein the LCDU data includes, for a specified focus level of a source of a lithographic apparatus used to print the pattern, (a) a first set of LCDU values of the features corresponding to a first dose level of the source, (b) a second set of LCDU values of the features corresponding to a second dose level, and (c) a third set of LCDU values of the features corresponding a third dose level;
correlating, using a decomposition method, (a) the first set of LCDU values to a first linear mixture of the error contribution from the multiple sources, (b) the second set of LCDU values to a second linear mixture of the error contribution from the multiple sources, (c) the third set of LCDU values to a third linear mixture of the error contribution from the multiple sources; and
deriving, from the linear mixtures and using the decomposition method, the error contribution from each of the sources,
wherein a first dataset of the collection of datasets includes error contribution values representative of an error contribution from a first one of the multiple sources,
wherein a second dataset of the collection of datasets includes error contribution values representative of an error contribution from a second one of the multiple sources, and
wherein a third dataset of the collection of datasets includes error contribution values representative of an error contribution from a third one of the multiple sources.
95. A non-transitory computer-readable medium having instructions that, when executed by a computer, cause the computer to execute a method for determining a source of error contribution to multiple features of a pattern printed on a substrate, the method comprising:
inputting, to a machine learning model, a specified dataset having error contribution values representative of an error contribution from one of multiple sources to the features; and
executing the machine learning model to determine a classification associated with the specified dataset, wherein the classification identifies a specified source of the multiple sources as the source of error contribution for the error contribution values in the specified dataset.
96. The computer-readable medium of any of clauses 95, wherein inputting the specified dataset includes:
processing an image of the pattern to obtain a collection of datasets, wherein each dataset in the collection of datasets has error contribution values representative of an error contribution from one of the multiple sources to the features, wherein the specified dataset is one dataset in the collection of datasets.
97. A method for training a machine learning model to determine a source of error contribution to multiple features of a pattern printed on a substrate, the method comprising:
obtaining training data having multiple datasets, wherein each dataset has error contribution values representative of an error contribution from one of multiple sources to the features, and wherein each dataset is associated with an actual classification that identifies a source of the error contribution of the corresponding dataset; and
training, based on the training data, a machine learning model to predict a classification of a reference dataset of the datasets such that a cost function that determines a difference between the predicted classification and the actual classification of the reference dataset is reduced.
98. The method of clause 97, wherein obtaining the training data includes:
obtaining local critical dimension uniformity (LCDU) data or LWR data associated with the features using different focus and dose level values of an apparatus used for printing the pattern.
99. The method of clause 98, wherein obtaining the training data includes:
decomposing LCDU data or LWR data associated with the features to derive the error contribution from each of the multiple sources.
100. The method of clause 97, wherein obtaining the training data includes:
generating (a) a first dataset of the training data having error contribution values representative of an error contribution from a first source of the multiple sources, (b) a second dataset of the training data having error contribution values representative of an error contribution from a second source of the multiple sources, and (c) a third dataset of the training data having error contribution values representative of an error contribution from a third source of the multiple sources, and
associating (d) the first dataset with a first classification that identifies a source of the error contribution as the first source, (e) the second dataset with a second classification that identifies a source of the error contribution as the second source, and (f) the third dataset with a third classification that identifies a source of the error contribution as the third source.
101. The method of clause 100, wherein the first source is an image acquisition tool used to acquire an image of the pattern, wherein the second source is a mask used to print the pattern on the substrate, and wherein the third source is a resist used to print the pattern together with a photon shot noise of an apparatus used to print the patter on the substrate. 102. The method of clause 100, wherein generating the first dataset includes:
generating multiple groups of the first, second and third datasets, wherein each group includes error contribution values representative of an error contribution from the first, second, and third sources, respectively, for a different subset of the features.
103. The method of clause 97, wherein training the machine learning model is an iterative process in which each iteration includes:
(a) executing the machine learning model, using the training data, to output the predicted classification of the reference dataset,
(b) determining the cost function as the difference between the predicted classification and the actual classification,
(c) adjusting the machine learning model,
(d) determining whether the cost function is reduced as a result of the adjusting, and
(e) responsive to the cost function not being reduced, repeating steps (a), (b), (c) and (d).
104. A method for determining a source of error contribution to multiple features of a pattern printed on a substrate, the method comprising:
processing an image of the pattern to obtain a collection of datasets, wherein each dataset in the collection of datasets has error contribution values representative of an error contribution from one of multiple sources to the features;
inputting a specified dataset of the multiple datasets to a machine learning model; and
executing the machine learning model to determine a classification associated with the specified dataset, wherein the classification identifies a specified source of the multiple sources as the source of error contribution for the error contribution values in the specified dataset.
105. An apparatus for training a machine learning model to determine a source of error contribution to multiple features of a pattern printed on a substrate, the apparatus comprising:
a memory storing a set of instructions; and at least one processor configured to execute the set of instructions to cause the apparatus to perform a method of:
obtaining training data having multiple datasets, wherein each dataset has error contribution values representative of an error contribution from one of multiple sources to the features, and wherein each dataset is associated with an actual classification that identifies a source of the error contribution of the corresponding dataset; and
training, based on the training data, a machine learning model to predict a classification of a reference dataset of the datasets such that a cost function that determines a difference between the predicted classification and the actual classification of the reference dataset is reduced.
106. The apparatus of clause 105, wherein obtaining the training data includes:
obtaining local critical dimension uniformity (LCDU) data or line width roughness (LWR) data associated with the features for different threshold levels on an image having the features, or using different focus and dose level values of an apparatus used for printing the pattern.
107. The apparatus of clause 106, wherein obtaining the training data includes:
decomposing LCDU data or LWR data associated with the features to derive the error contribution values from each of the multiple sources.
108. The apparatus of clause 105, wherein obtaining the training data includes:
generating (a) a first dataset of the training data having error contribution values representative of an error contribution from a first source of the multiple sources, (b) a second dataset of the training data having error contribution values representative of an error contribution from a second source of the multiple sources, and (c) a third dataset of the training data having error contribution values representative of an error contribution from a third source of the multiple sources, and
associating (d) the first dataset with a first classification that identifies the first source as the source of error contribution, (e) the second dataset with a second classification that identifies the second source as the source of error contribution, and (f) the third dataset with a third classification that identifies the third source as the source of error contribution.
109. The apparatus of clause 108, wherein the first source is an image acquisition tool used to acquire an image of the pattern, wherein the second source is a mask used to print the pattern on the substrate, and wherein the third source is a resist used to print the pattern together with a photon shot noise of an apparatus used to print the patter on the substrate.
110. The apparatus of clause 108, wherein generating the first dataset includes:
generating multiple groups of the first, second and third datasets, wherein each group includes error contribution values representative of an error contribution from the first, second, and third sources, respectively, for a different subset of the features.
111. The apparatus of clause 105, wherein training the machine learning model is an iterative process in which each iteration includes:
(a) executing the machine learning model, using the training data, to output the predicted classification of the reference dataset,
(b) determining the cost function as the difference between the predicted classification and the actual classification,
(c) adjusting the machine learning model,
(d) determining whether the cost function is reduced as a result of the adjusting, and
(e) responsive to the cost function not being reduced, repeating steps (a), (b), (c) and (d).
112. The apparatus of any of clauses 105-111, wherein the machine learning model is a recurrent neural network.
113. The apparatus of clause 105 further comprising:
receiving a specified dataset having error contribution values representative of an error contribution from one of the multiple sources to a set of features of a specified pattern printed on a specified substrate; and
executing the machine learning model to determine a classification associated with the specified dataset, wherein the classification identifies a specified source of the multiple sources as the source of error contribution for the error contribution values in the specified dataset.
114. The apparatus of clause 113, wherein receiving the specified dataset includes:
decomposing, using a decomposition method, multiple measurement values associated with the set of features to derive a collection of datasets representative of error contributions from each of the multiple sources, wherein the specified dataset is one of the collection of datasets and corresponds to error contribution from one of the multiple sources.
115. The apparatus of clause 114, wherein decomposing the measurement values includes: obtaining an image of the specified pattern;
obtaining, using the image, the measurement values, wherein the measurement values are obtained for different sensor values;
correlating, using the decomposition method, each measurement value of the measurement values to a linear mixture of the error contributions to generate a plurality of linear mixtures of the error contributions; and
deriving, from the linear mixtures and using the decomposition method, each of the error contributions.
116. The apparatus of clause 115, wherein the different sensor values correspond to different threshold levels associated with the image, wherein each measurement value corresponds to a delta critical dimension (CD) value of a feature of the set of features at one of the different threshold values, wherein the delta CD value indicates a deviation of a CD value of the feature from a mean value of a plurality of CD values of the set of features.
117. The apparatus of clause 116, wherein the CD value is a difference between a measured contour of the feature and a simulated contour of the feature.
118. The apparatus of clause 116, wherein each threshold value of the different threshold values corresponds to a threshold of a pixel value in the image.
119. The apparatus of clause 115, wherein the measurement values correspond to LCDU value or LWR value of the feature at the different sensor values.
120. The apparatus of clause 119, wherein the different sensor values correspond to different dose levels associated with a source of a lithographic apparatus used to print the pattern.
121. The apparatus of clause 119, wherein the different sensor values correspond to different focus levels associated with a source of a lithographic apparatus used to print the pattern.
122. The apparatus of any of clauses 115-121, wherein deriving each of the error contributions includes:
deriving each of the error contributions using ICA method as the decomposition method.
123. A computer program product comprising a non-transitory computer readable medium having instructions recorded thereon, the instructions when executed by a computer implementing the method of any of the above clauses.
124. A non-transitory computer-readable medium having instructions that, when executed by a computer, cause the computer to execute a method for training a machine learning model to determine error contributions to a feature of a pattern printed on a substrate, the method comprising:
obtaining training data having multiple datasets, wherein the datasets include a first dataset having (a) a first image data of one or more features of a pattern to be printed on a substrate and (b) a first error contribution data comprising error contributions from multiple sources to the one or more features; and
training, based on the training data, a machine learning model to predict error contribution data for the first dataset such that a cost function that is indicative of a difference between the predicted error contribution data and the first error contribution data is reduced.
125. The computer-readable medium of clause 124, wherein the first image data includes a first image of a feature of the one or more features, and wherein the first error contribution data includes a first set of error contribution values corresponding to delta critical dimension (CD) values of the first feature.
126. The computer-readable medium of clause 125, wherein each delta CD value is indicative of a deviation of a CD value of the first feature from a mean of a plurality of CD values of the one or more features.
127. The computer-readable medium of clause 124, wherein the first image data includes a first set of images of multiple features of the one or more features, and wherein the first error contribution data includes a first set of error contribution values corresponding to local CD uniformity (LCDU) values of the features.
128. The computer-readable medium of clause 124, wherein the first error contribution data includes multiple sets of error contribution values corresponding to multiple measurement points on a feature of the one or more features, wherein the sets of error contribution values include a first set of error contribution values representative of error contributions from the multiple sources at a first measurement point of the measurement points.
129. The computer-readable medium of clause 124, wherein the first error contribution data is determined based on measurement data of the one or more features.
130. The computer-readable medium of clause 129, wherein the measurement data comprises a CD value of a feature of the one or more features or a LCDU value of multiple features of the one or more features.
131. The computer-readable medium of clause 124, wherein the error contributions include:
an image acquisition tool error contribution that is associated with an image acquisition tool used to acquire the first image data,
a mask error contribution that is associated with a mask used to print the pattern on the substrate, and
a resist error contribution that is associated with a resist used to print the pattern, wherein the resist error contribution includes photoresist chemical noise and a shot noise associated with a source of a lithographic apparatus used to print the pattern.
132. The computer-readable medium of clause 124, wherein training the machine learning model is an iterative process in which each iteration includes:
(a) executing the machine learning model, using the multiple datasets, to output the predicted error contribution data,
(b) determining the cost function as the difference between the predicted error contribution data and the first error contribution data,
(c) adjusting the machine learning model,
(d) determining whether the cost function is reduced as a result of the adjusting, and
(e) responsive to the cost function not being reduced, repeating steps (a), (b), (c) and (d).
133. The computer-readable medium of clause 124 further comprising:
receiving image data of a set of features of a specified pattern to be printed on a specified substrate; and executing the machine learning model to determine error contribution data comprising error contributions from the multiple sources to the set of features.
134. The computer-readable medium of clause 133, wherein the image data includes an image of a feature of the set of features, and wherein the error contribution data includes error contribution values corresponding to delta CD values associated with feature.
135. The computer-readable medium of clause 133, wherein the image data includes a set of images of the set of features, and wherein the error contribution data includes error contribution values corresponding to LCDU values associated with the set of features.
136. The computer-readable medium of clause 133, wherein the error contribution data includes multiple sets of error contribution values corresponding to multiple measurement points on a feature of the set of features, wherein the sets of error contribution values include a first set of error contribution values representative of error contributions from the multiple sources at a first measurement point of the measurement points.
137. The computer-readable medium of clause 133 further comprising:
adjusting, based on a mask error contribution of the error contributions, one or more parameters of at least one of a mask or a source of a lithographic apparatus used to print the specified pattern.
138. The computer-readable medium of clause 133 further comprising:
adjusting, based on a resist error contribution of the error contributions, one or more parameters of at least one of a mask or a source of a lithographic apparatus used to print the specified pattern.
139. A non-transitory computer-readable medium having instructions that, when executed by a computer, cause the computer to execute a method for determining error contribution data comprising error contributions from multiple sources to a feature of a pattern to printed on a substrate, the method comprising:
receiving image data of a set of features of a specified pattern to be printed on a first substrate; inputting the image data to a machine learning model; and
executing the machine learning model to determine error contribution data comprising error contributions from multiple sources to the set of features.
140. The computer-readable medium of clause 139, wherein the image data includes an image of a feature of the set of features, and wherein the error contribution data includes error contribution values corresponding to delta CD values associated with feature.
141. The computer-readable medium of clause 139, wherein the image data includes a set of images of the set of features, and wherein the error contribution data includes error contribution values corresponding to LCDU values associated with the set of features.
142. The computer-readable medium of clause 139, wherein the error contribution data includes multiple sets of error contribution values corresponding to multiple measurement points on a feature of the set of features, wherein the sets of error contribution values include a first set of error contribution values representative of error contributions from the multiple sources at a first measurement point of the measurement points.
143. The computer-readable medium of clause 139, wherein executing the machine learning model to determine the error contribution data includes:
training the machine learning model using multiple datasets, wherein the datasets include a first dataset having (a) a first image data of one or more features of a pattern to be printed on a substrate and (b) a first error contribution data comprising error contributions from multiple sources to the one or more features.
144. The computer-readable medium of clause 143, wherein the first image data includes a first image of a feature of the one or more features, and wherein the first error contribution data includes a first set of error contribution values corresponding to delta CD values of the first feature.
145. The computer-readable medium of clause 143, wherein the first image data includes a first set of images of multiple features of the one or more features, and wherein the first error contribution data includes a first set of error contribution values corresponding to LCDU values of the features.
146. The computer-readable medium of clause 143, wherein the error contributions include:
an image acquisition tool error contribution that is associated with an image acquisition tool used to acquire the first image data,
a mask error contribution that is associated with a mask used to print the pattern on the substrate, and
a resist error contribution that is associated with a resist used to print the pattern, wherein the resist error contribution includes photoresist chemical noise and a shot noise associated with a source of a lithographic apparatus used to print the pattern.
147. A method for training a machine learning model to determine error contributions to a feature of a pattern printed on a substrate, the method comprising:
obtaining training data having multiple datasets, wherein the datasets include a first dataset having (a) a first image data of one or more features of a pattern to be printed on a substrate and (b) a first error contribution data comprising error contributions from multiple sources to the one or more features; and
training, based on the training data, a machine learning model to predict error contribution data for the first dataset such that a cost function that is indicative of a difference between the predicted error contribution data and the first error contribution data is reduced.
148. The method of clause 147, wherein the first image data includes a first image of a feature of the one or more features, and wherein the first error contribution data includes a first set of error contribution values corresponding to delta critical dimension (CD) values of the first feature.
149. The method of clause 148, wherein each delta CD value is indicative of a deviation of a CD value of the first feature from a mean of a plurality of CD values of the one or more features.
150. The method of clause 147, wherein the first image data includes a first set of images of multiple features of the one or more features, and wherein the first error contribution data includes a first set of error contribution values corresponding to local CD uniformity (LCDU) values of the features.
151. The method of clause 147, wherein the first error contribution data includes multiple sets of error contribution values corresponding to multiple measurement points on a feature of the one or more features, wherein the sets of error contribution values include a first set of error contribution values representative of error contributions from the multiple sources at a first measurement point of the measurement points.
152. The method of clause 147, wherein the first error contribution data is determined based on measurement data of the one or more features.
153. The method of clause 152, wherein the measurement data comprises a CD value of a feature of the one or more features or a LCDU value of multiple features of the one or more features.
154. The method of clause 147, wherein the error contributions include:
an image acquisition tool error contribution that is associated with an image acquisition tool used to acquire the first image data,
a mask error contribution that is associated with a mask used to print the pattern on the substrate, and
a resist error contribution that is associated with a resist used to print the pattern, wherein the resist error contribution includes photoresist chemical noise and a shot noise associated with a source of a lithographic apparatus used to print the pattern.
155. The method of clause 147, wherein training the machine learning model is an iterative process in which each iteration includes:
(a) executing the machine learning model, using the multiple datasets, to output the predicted error contribution data,
(b) determining the cost function as the difference between the predicted error contribution data and the first error contribution data,
(c) adjusting the machine learning model,
(d) determining whether the cost function is reduced as a result of the adjusting, and
(e) responsive to the cost function not being reduced, repeating steps (a), (b), (c) and (d).
156. The method of clause 147 further comprising:
receiving image data of a set of features of a specified pattern to be printed on a specified substrate; and executing the machine learning model to determine error contribution data comprising error contributions from the multiple sources to the set of features.
157. The method of clause 156, wherein the image data includes an image of a feature of the set of features, and wherein the error contribution data includes error contribution values corresponding to delta CD values associated with feature.
158. The method of clause 156, wherein the image data includes a set of images of the set of features, and wherein the error contribution data includes error contribution values corresponding to LCDU values associated with the set of features.
159. The method of clause 156, wherein the error contribution data includes multiple sets of error contribution values corresponding to multiple measurement points on a feature of the set of features, wherein the sets of error contribution values include a first set of error contribution values representative of error contributions from the multiple sources at a first measurement point of the measurement points.
160. The method of clause 156 further comprising:
adjusting, based on a mask error contribution of the error contributions, one or more parameters of at least one of a mask or a source of a lithographic apparatus used to print the specified pattern.
161. The method of clause 156 further comprising:
adjusting, based on a resist error contribution of the error contributions, one or more parameters of at least one of a mask or a source of a lithographic apparatus used to print the specified pattern.
162. A method for determining error contribution data comprising error contributions from multiple sources to a feature of a pattern to printed on a substrate, the method comprising:
receiving image data of a set of features of a specified pattern to be printed on a first substrate; inputting the image data to a machine learning model; and
executing the machine learning model to determine error contribution data comprising error contributions from multiple sources to the set of features.
163. The method of clause 162, wherein the image data includes an image of a feature of the set of features, and wherein the error contribution data includes error contribution values corresponding to delta CD values associated with feature.
164. The method of clause 162, wherein the image data includes a set of images of the set of features, and wherein the error contribution data includes error contribution values corresponding to LCDU values associated with the set of features.
165. The method of clause 162, wherein the error contribution data includes multiple sets of error contribution values corresponding to multiple measurement points on a feature of the set of features, wherein the sets of error contribution values include a first set of error contribution values representative of error contributions from the multiple sources at a first measurement point of the measurement points.
166. The method of clause 162, wherein executing the machine learning model to determine the error contribution data includes:
training the machine learning model using multiple datasets, wherein the datasets include a first dataset having (a) a first image data of one or more features of a pattern to be printed on a substrate and (b) a first error contribution data comprising error contributions from multiple sources to the one or more features.
167. The method of clause 162, wherein the error contributions include:
an image acquisition tool error contribution that is associated with an image acquisition tool used to acquire the first image data,
a mask error contribution that is associated with a mask used to print the pattern on the substrate, and
a resist error contribution that is associated with a resist used to print the pattern, wherein the resist error contribution includes photoresist chemical noise and a shot noise associated with a source of a lithographic apparatus used to print the pattern.
168. An apparatus for training a machine learning model to determine error contributions to a feature of a pattern printed on a substrate, the apparatus comprising:
a memory storing a set of instructions; and
at least one processor configured to execute the set of instructions to cause the apparatus to perform a method of:
obtaining training data having multiple datasets, wherein the datasets include a first dataset having (a) a first image data of one or more features of a pattern to be printed on a substrate and (b) a first error contribution data comprising error contributions from multiple sources to the one or more features; and
training, based on the training data, a machine learning model to predict error contribution data for the first dataset such that a cost function that is indicative of a difference between the predicted error contribution data and the first error contribution data is reduced.
169. The apparatus of clause 168, wherein the first image data includes a first image of a feature of the one or more features, and wherein the first error contribution data includes a first set of error contribution values corresponding to delta critical dimension (CD) values of the first feature.
170. The apparatus of clause 169, wherein each delta CD value is indicative of a deviation of a CD value of the first feature from a mean of a plurality of CD values of the one or more features.
171. The apparatus of clause 168, wherein the first image data includes a first set of images of multiple features of the one or more features, and wherein the first error contribution data includes a first set of error contribution values corresponding to local CD uniformity (LCDU) values of the features.
172. The apparatus of clause 168, wherein the first error contribution data includes multiple sets of error contribution values corresponding to multiple measurement points on a feature of the one or more features, wherein the sets of error contribution values include a first set of error contribution values representative of error contributions from the multiple sources at a first measurement point of the measurement points.
173. The apparatus of clause 168, wherein the first error contribution data is determined based on measurement data of the one or more features.
174. The apparatus of clause 173, wherein the measurement data comprises a CD value of a feature of the one or more features or a LCDU value of multiple features of the one or more features.
175. The apparatus of clause 168, wherein the error contributions include:
an image acquisition tool error contribution that is associated with an image acquisition tool used to acquire the first image data,
a mask error contribution that is associated with a mask used to print the pattern on the substrate, and
a resist error contribution that is associated with a resist used to print the pattern, wherein the resist error contribution includes photoresist chemical noise and a shot noise associated with a source of a lithographic apparatus used to print the pattern.
176. The apparatus of clause 168, wherein training the machine learning model is an iterative process in which each iteration includes:
(a) executing the machine learning model, using the multiple datasets, to output the predicted error contribution data,
(b) determining the cost function as the difference between the predicted error contribution data and the first error contribution data,
(c) adjusting the machine learning model,
(d) determining whether the cost function is reduced as a result of the adjusting, and
(e) responsive to the cost function not being reduced, repeating steps (a), (b), (c) and (d).
177. The apparatus of clause 168 further comprising:
receiving image data of a set of features of a specified pattern to be printed on a specified substrate; and executing the machine learning model to determine error contribution data comprising error contributions from the multiple sources to the set of features.
178. The apparatus of clause 177, wherein the image data includes an image of a feature of the set of features, and wherein the error contribution data includes error contribution values corresponding to delta CD values associated with feature.
179. The apparatus of clause 177, wherein the image data includes a set of images of the set of features, and wherein the error contribution data includes error contribution values corresponding to LCDU values associated with the set of features.
180. The apparatus of clause 177, wherein the error contribution data includes multiple sets of error contribution values corresponding to multiple measurement points on a feature of the set of features, wherein the sets of error contribution values include a first set of error contribution values representative of error contributions from the multiple sources at a first measurement point of the measurement points.
181. The apparatus of clause 177 further comprising:
adjusting, based on a mask error contribution of the error contributions, one or more parameters of at least one of a mask or a source of a lithographic apparatus used to print the specified pattern.
182. The apparatus of clause 177 further comprising:
adjusting, based on a resist error contribution of the error contributions, one or more parameters of at least one of a mask or a source of a lithographic apparatus used to print the specified pattern.
183. An apparatus for determining error contribution data comprising error contributions from multiple sources to a feature of a pattern to printed on a substrate, the apparatus comprising:
a memory storing a set of instructions; and
at least one processor configured to execute the set of instructions to cause the apparatus to perform a method of:
receiving image data of a set of features of a specified pattern to be printed on a first substrate; inputting the image data to a machine learning model; and
executing the machine learning model to determine error contribution data comprising error contributions from multiple sources to the set of features.
184. The apparatus of clause 183, wherein the image data includes an image of a feature of the set of features, and wherein the error contribution data includes error contribution values corresponding to delta CD values associated with feature.
185. The apparatus of clause 183, wherein the image data includes a set of images of the set of features, and wherein the error contribution data includes error contribution values corresponding to LCDU values associated with the set of features.
186. The apparatus of clause 183, wherein the error contribution data includes multiple sets of error contribution values corresponding to multiple measurement points on a feature of the set of features, wherein the sets of error contribution values include a first set of error contribution values representative of error contributions from the multiple sources at a first measurement point of the measurement points.
187. The apparatus of clause 183, wherein executing the machine learning model to determine the error contribution data includes:
training the machine learning model using multiple datasets, wherein the datasets include a first dataset having (a) a first image data of one or more features of a pattern to be printed on a substrate and (b) a first error contribution data comprising error contributions from multiple sources to the one or more features.
188. The apparatus of clause 183, wherein the error contributions include:
an image acquisition tool error contribution that is associated with an image acquisition tool used to acquire the first image data,
a mask error contribution that is associated with a mask used to print the pattern on the substrate, and
a resist error contribution that is associated with a resist used to print the pattern, wherein the resist error contribution includes photoresist chemical noise and a shot noise associated with a source of a lithographic apparatus used to print the pattern.
As used herein, unless specifically stated otherwise, the term “or” encompasses all possible combinations, except where infeasible. For example, if it is stated that a component includes A or B, then, unless specifically stated otherwise or infeasible, the component may include A, or B, or A and B. As a second example, if it is stated that a component includes A, B, or C, then, unless specifically stated otherwise or infeasible, the component may include A, or B, or C, or A and B, or A and C, or B and C, or A and B and C. Expressions such as “at least one of” do not necessarily modify an entirety of a following list and do not necessarily modify each member of the list, such that “at least one of A, B, and C” should be understood as including only one of A, only one of B, only one of C, or any combination of A, B, and C. The phrase “one of A and B” or “any one of A and B” shall be interpreted in the broadest sense to include one of A, or one of B.
The descriptions above are intended to be illustrative, not limiting. Thus, it will be apparent to one skilled in the art that modifications may be made as described without departing from the scope of the claims set out below.
Number | Date | Country | Kind |
---|---|---|---|
20174556.9 | May 2020 | EP | regional |
20177933.7 | Jun 2020 | EP | regional |
21171063.7 | Apr 2021 | EP | regional |
This application claims priority of International Application PCT/EP2021/062772, which was filed on May 21, 2021, which claims priority of EP application 20174556.9, which was filed on May 14, 2020, of EP application 20177933.7, which was filed on Jun. 3, 2020, and of EP application 21171063.7 which was filed on Apr. 28, 2021, all of which are incorporated herein by reference in their entireties.
Number | Date | Country | |
---|---|---|---|
Parent | PCT/EP2021/062772 | May 2021 | US |
Child | 17986829 | US |