The performance of semiconductor device fabrication operations such as plasma-assisted etch processes is often essential to the success of a semiconductor device processing workflow. However, optimization or tuning of the fabrication processes and/or the tools associated with them (e.g., etch reactors, lithography masks, etc.) may prove technically difficult and time-consuming, often involving skilled personnel manually adjusting etch process parameters or tool component designs to generate the desired target feature profile. Metrology tools should accurately gauge the performance of etch or deposition processes by measuring features on fabricated or partially fabricated devices. Currently, various metrology tools are used for this purpose.
Background and contextual descriptions contained herein are provided solely for the purpose of generally presenting the context of the disclosure. Much of this disclosure presents work of the inventors, and simply because such work is described in the background section or presented as context elsewhere herein does not mean that it is admitted to be prior art.
Certain aspects of this disclosure pertain to method of determining a profile, critical dimension, and/or contour of one or more features on a processed substrate. Such methods may be characterized by the following features: (a) performing optical metrology on the one or more features on the processed substrate to produce an optical metrology output; (b) providing the optical metrology output to a metrology machine learning model that has been trained using a training set of (i) profiles, critical dimensions, and/or contours for a plurality of features, and (ii) optical metrology outputs for the plurality of features; and (c) receiving, from the metrology machine learning model, the profile, critical dimension, and/or contour of the one or more features on the processed substrate. In certain embodiments, the methods additionally include training the metrology machine learning model by producing multiple sets of optical metrology output, each generated for a different orientation and/or location of test features with respect an optical metrology apparatus, for the test features. In certain embodiments, the metrology machine learning model was produced using a supervised machine learning technique.
In certain embodiments, the optical metrology is a scatterometry technique. For example, the optical metrology output may include a reflectance spectra and/or ellipsometric output data.
In certain embodiments, the methods additionally include performing pattern recognition on the one or more features on the processed substrate; and determining that a pattern of the one or more features conforms with an expected pattern. In certain embodiments, the methods additionally include performing optical metrology on the one or more features on the processed substrate includes raster scanning over the processed substrate.
In some cases, the profiles, critical dimensions, and/or contours for a plurality of features in the training set were obtained using an electron microscopy technique and/or using CD-SAXS.
In certain embodiments, the methods additionally include decomposing the optical metrology output to produce a reduced set of optical metrology values and providing the reduced set of optical metrology values to the metrology machine learning model. In some cases, decomposing the optical metrology output includes identifying principal components of the optical metrology output or applying the optical metrology output to an autoencoder.
Other aspects of the disclosure pertain to metrology systems that may be characterized by the following elements: (a) an optical metrology tool including an optical probe source and an optical detector and a processor configured to produce an optical metrology output from data produced by the optical detector when an optical probe is directed onto one or more features on a processed substrate; and (b) a metrology machine learning model that has been trained using a training set of (i) profiles, critical dimensions, and/or contours for a plurality of features, and (ii) optical metrology outputs for the plurality of features. In certain embodiments, the metrology machine learning model is configured to: receive the optical metrology output from the optical metrology tool; and output the profile, critical dimension, and/or contour of the one or more features on the processed substrate.
In some systems, the metrology machine learning model was trained using multiple sets of optical metrology output, each generated for a different orientation and/or location of test features with respect a test optical metrology apparatus. In some embodiments, the metrology machine learning model was generated using a supervised machine learning technique.
In certain embodiments, the optical metrology tool is a scatterometer. In some metrology systems, the optical metrology output inlcudes a reflectance spectra and/or ellipsometric output data.
Some metrology systems additionally include a pattern recognition tool comprising a camera and processor configured to: perform pattern recognition on the one or more features on the processed substrate; and determine that a pattern of the one or more features conforms with an expected pattern. In some metrology systems, the optical metrology tool is configured to raster scan over the processed substrate.
In some metrology systems, the profiles, critical dimensions, and/or contours for a plurality of features in the training set were obtained using an electron microscopy technique and/or by using CD-SAXS.
In certain embodiments, the metrology machine learning model is further configured to: decompose the optical metrology output to produce a reduced set of optical metrology values and provide the reduced set of optical metrology values to the metrology machine learning model. In some such embodiments, the metrology machine learning model is configured to decompose the optical metrology output by identifying principal components of the optical metrology output or by applying the optical metrology output to an autoencoder.
These and other features will be described below with reference to the associated drawings.
Current methods of evaluating semiconductor wafers and other samples used in semiconductor process development employ techniques such as x-SEM (x-ray Scanning Electron Microscopy), STEM (Scanning Transmission Electron Microscopy), and CD-SAXS (Critical Dimension Small Angle X-ray Scattering) for cross-sectional image capture. Current methods then use image analysis software to extract the feature of interest. Coupons are often used instead of full wafers to study the effect of etch or other fabrication processes on patterned substrates. Coupons are test samples or substrates that typically have materials and layers that are identical to or resemble those materials and layers used in a semiconductor device fabrication operation. Frequently, coupons are smaller than semiconductor wafers. They may be circular or polygonal (e.g., rectangular) and occupy only a few square centimeters.
Among the challenges sometimes encountered with current electron microscopy-based approach are the following.
Disclosed herein are methods and systems for supervised machine learning using spectra collected off coupon samples to generate models that can easily and accurately predict feature characteristics normally obtained by a destructive metrology technique such as electron microscopy (e.g., x-SEM and STEM) or by another time consuming process such as CD-SAXS. During training, many samples may be processed according to a defined process (e.g., etch or deposition) and then cross sectioned to extract the profile information useful for process development and/or process monitoring. Models trained using this information can receive optical signals from optical metrology tools—which tend operate quickly, inexpensively, and non-destructively—and output the ground truth of features established by another metrology tool—which tend to operate slowly, expensively, and often destructively.
In some cases, the profile information may suffer from inaccuracy due to feature extraction off of blurry boundaries in micrographs or other images. If a sufficiently large training set is used (e.g., at least about 20 samples, or at least about 100 samples), and each of these samples is evaluated using both electron microscopy (and/or CD-SAXS) and optical metrology (e.g., x-SEM and STEM is conducted along with measuring spectra) on the feature of interest, the method can enable reliable supervised training.
In some cases, at least some of the samples are evaluated at locations where a single type of feature occupies at least a significant fraction of an optical spot created by incident light on the sample. Often the features of interest have dimensions on the order of a few hundred nanometers or less, while optical spots are on the order of 5-40 micrometers. For a single feature to make a strong contribution to the optical signals obtained from the metrology tool, when spot sizes are significantly larger than feature sizes, the feature may need to be repeated over the area occupied by the incident spot. However, this is not a rigid requirement of the samples/locations used in the training set. Many machine learning techniques are sufficiently robust that at least some training set members (or at least their optical metrology signals) can have contributions of multiple types of features in an optical signal. Regardless of the composition of the training set, the pairing of spectra and CD, profile, and/or contour information extracted using a reliable profiling technique such as x-SEM, STEM, CD-SAXS or other destructive or slow metrology can enable a supervised training.
Note that the training is supervised in the sense that the destructive metrology technique (e.g., x-SEM/STEM) or other slow profiling technique provides known feature characteristics such as a feature's profile, contour, or critical dimension (CD) that are used to identify optical signal characteristics that correspond to the known feature characteristics. The resulting model can accurately predict feature profile information when supplied with optical data collected under conditions used to train the model. In the methods and systems described here, there may be two implementation steps—training and testing. Ultimately, a generated and validated metrology machine learning model may be employed in an optical metrology system to directly provide information about feature profiles, CDs, and/or contours.
Training
In certain embodiments, the training employs scatterometry to capture the reflected spectra using a range of optical information, which may span ranges of angle of incidence (normal incidence and oblique incidence), azimuthal direction including a full 360 degrees, wavelength (e.g., including visible, infrared, and/or ultraviolet), and polarization on the feature of interest. After collecting this optical metrology information, the coupons are analyzed by x-SEM, TEM, CD-SAXS, or the like. An accumulation of several samples (e.g., greater than about 20 samples or greater than about 100 samples) may be employed. Each such sample may generate one or more training set member. For example, a single sample may be analyzed using a plurality of optical metrology settings (e.g., angles of incidence), each setting producing a separate optical metrology reading and associated training set member.
It can be challenging to position and/or orient a coupon in an error-free manner. While metrology tools are designed to fine tune the position and orientation of wafers that have standard sizes, and thereby provide relatively low error capture of optical metrology results, these tools are not designed to so carefully control the position and orientation of coupons. Coupons frequently have sizes and shapes that are much different from the sizes and shapes of production wafers. Therefore steps may need to be taken to address the possible errors in optical metrology signals produced from coupons.
In certain embodiments, positional and/or rotational error is reduced by translating, rotating, and/or otherwise reorienting the sample different ways under the beam of an optical metrology tool and capturing optical signals at all these different positions or orientations. Each of the different positions/orientations may present the same location for the spot of a beam of the optical metrology tool. Collectively, these different positions average out error that might occur in a small fraction of the different translational positions and/or orientations. For training, each of the different optical signals captured at these different positions or orientations is associated with the same x-SEM, STEM, CD-SAXS, or other direct measurement. Thus, a given position on a sample (coupon) may produce a large number of training set members, each using the same x-SEM, STEM, or CD-SAXS measurement but different optical signals obtained at different translational positions or orientations. Collectively, the number of such training set members is large enough to cover realistic error of typical experimental conditions.
As should be apparent, each sample subject to optical metrology is subsequently sent to an x-SEM apparatus, a TEM apparatus, a CDSAXS apparatus, or other apparatus for conducting measurement on features having been produced by a process under consideration. The destructive process provides information for extracting the CD, profile, contour, or other characteristics of features.
Z-direction resolved information can be obtained, as noted, by non-destructive (or less destructive) techniques such as CD-SAXS (critical dimension small angle X-ray scattering). CD-SAXS is an X-ray scattering technique that can be used to reconstruct the in-plane and out-of-plane structure of nanostructured thin-films. The technique involves collecting a series of transmission small angle X-ray scattering images, at a variety of sample rotation angles. These (qx,qy) images can be combined to reconstruct the 3D reciprocal space (e.g., Fourier transformed), in particular probing the (qy,qz) slice that contains both in-plane and out-of-plane (film normal direction) information.
Training may start using, e.g., about 100 pairs of spectra and CD values, profiles, and/or contours. Different supervised learning techniques can be employed (regression, decision tree, support vector machine, neural networks, etc.) to map the relationship between spectra and CD/ profile/contour features. Hyper-parameters may be optimized to ensure best validation result before testing. If other information pertaining to coupon-to-coupon variations, e.g., layer thicknesses and optical constants, is available, it can also be used as the input to supervised learning.
In one implementation, the supervised training method utilizes a dimensionality reduction technique/tool such as principal component analysis (PCA) or an autoencoder. The dimensionality reduction technique is optionally used in conjunction with producing a neural network or other supervised machine learning model. For example, PCA may be employed to identify optical signal characteristics or combinations of optical signal characteristics from optical metrology signals that are affected relatively strongly, compared to other optical signal characteristics, by a feature characteristic of interest such as feature critical dimension. The resulting principal components (vectors of optical signal characteristics that are strongly affected by features of interest) are used in training the model and as inputs to resulting model. More generally, any technique that decomposes or identifies aspects of optical signals for use in training a model (and as inputs to subsequently generated models) may be employed. Any such technique may receive n spectral intensity values (from optical metrology) that result from m sources of variation in substrate features and identify approximately m unique spectral signatures. The value of n will typically be much greater than the value of m, e.g., but at least an order of magnitude.
In another implementation, a deep learning neural network is employed in a manner that does not require a principal components extraction step. As with the PCA-neural network implementation, once training is done to produce a deep learning model with sufficient accuracy to predict the CD, contour, and/or profile information, it can circumvent x-SEM, STEM, CD-SAXS and the like to directly report the above information (e.g., feature profiles or geometric characteristics of the features).
With the averaging power of optical technology on large targets, this method provides average CD/contour/profile information more effectively than multiple x-SEM, STEM, CD-SAXS, etc. measurements. Periodic structures are most useful due to their translational invariance, but the methods are not limited to periodic structures.
In certain embodiments, the training can be supplemented with pattern data representing an expected pattern at the locations on a sample where optical metrology obtains signal and the SEM cross-section is taken. Standard pattern recognition software is available to take camera images of such locations and determine whether the location has an expected pattern. If the location does not have the expected pattern, optical and microscopy data may not be acquired, or if it is acquired, may be eliminated or given low weight in the model training.
The non-invasive metrology data used to train the model can come from any of various conventional and unconventional optical techniques. Such techniques may capture reflectance and/or ellipsometric spectra using, e.g., scatterometry and/or ellipsometry.
In certain embodiments, the training can be supplemented with optical signal obtained by raster scanning the beam from the metrology tool over the coupon or other sample. While the resulting data is collected over a range of locations on the sample, it may be associated with feature profiles (or other geometric characteristics) at one or more specific locations on the sample.
Example Process Flow for Training
Next, at an operation 105, optical metrology is performed on the features of the test substrates produced in operation 103. In each case, the resulting optical signals may be associated with the substrate or features that produced those signals, although the profiles or other geometric details of the features typically are not yet known. As should be understood, the optical signals may not directly indicate the profile, critical dimension, contour, or other geometric characteristic of the features that produced those signals.
Optionally, at a process operation 107, the process repeats performing optical metrology on the features under different conditions (e.g., different substrate orientations, different probe beam spot sizes, etc.). As explained, many forms of optical metrology may be employed, and, as well, many modes of executing a single metrology may be employed. Such modes include specific settings of the optical metrology tool such as azimuthal positions and angles of incidence of beams, spot size, spot position on the substrate, rasterized scanning, etc. Further, a pattern recognition pre-step may be employed to remove certain measurements that do not appear to correspond to an expected feature pattern.
Next, in an operation 109, the method performs a form of metrology that directly generates profiles or other geometric characteristics of the features on the test substrates. As indicated, such metrologies are often slow and/or destructive. Examples include various forms of electron microscopy including various forms of TEM (e.g., STEM) and SEM (e.g., x-SEM and CD-SEM), as well as CD-SAXS. The resulting directly measured geometric characteristics of the features are associated with the corresponding optical signals produced by the features in operation 105, and optionally operation 107.
At this point, information for a full training set is available, and, as illustrated at operation 111, the process trains a metrology model using optical metrology signals produced in 105 and, optionally, 107 together with the associated feature profiles, contours, CDs, etc. produced in 109. The resulting model is tested and/or validated as indicated at operation 113. The resulting model may be installed in a metrology tool or system.
The following sequence presents a further example of a process flow.
In testing, an inference is conducted by using the same optical metrology technique (e.g., scatterometry) used to capture the raw data during training, and then feed into the previously established supervised learning model, to predict the CD/profile feature of interest. If training is conducted with metrology conducted at multiple translational positions and/or orientations, then some or all of those positions/orientations may be employed to capture the data fed into the model. Further, if the training is conducted using pattern recognition by a camera, that same pattern recognition may be employed to filter data provided to the model. Still further, if the training is conducted using a raster scan of an optical metrology probe beam, that same raster scan may be conducted during testing and the resulting scan data provided to the model. Of course, test data should be generated under different conditions than the training data. So typically, the test data is generated on substrate features produced using fabrication processes that are in some ways different from those employed to generate the training data. For example, the test data may be generated using combinations of etch conditions that are different from any combination used to generate any of the training data.
Applications
The metrology machine learning models generated as described herein have various applications. The scope of the application of any given model depends, at least in part, on the range of applicability for the model; some models are applicable more generally—in terms of number of processes for which they can validly predict feature characteristics—than other models. For example, a model may be produced for a particular process; e.g., an anisotropic etch process employing a particular chemistry to produce the pattern of a particular design layout.
After effectively training a model, this method can drastically reduce the costs associated with performing destructive testing such as STEM and x-SEM. This is useful if the same type of samples is to be repeatedly used to tune the process on, e.g., >100 coupon samples.
This method also brings in better coupon-to-coupon consistency as random errors & LER/LWR (line edge roughness and line width roughness) on CD/contours/profiles are spatially averaged out by the optical signals. Machine learning is utilized here to map the complex relationship between spectra and CD/contour/profile of feature of interest. When the errors are well controlled, this work flow may be operator free. The nondestructive nature of the measurement also allows multi-etch-step consistency of coupon result. Multiple inference results and multiple features can come out of each etch step of a single coupon, maximizing its value.
A trained metrology model may be employed as a part of a process or associated system for determining the profile, CD, contours, or other geometric characteristic of one or more features on a substrate subject to a device fabrication process. The model and associated metrology system may be used during process development, process monitoring, and/or process modification. Depending on the application, the substrate subjected to the metrology and analysis may be a test coupon or wafer, a production wafer having a partially fabricated device, a production wafer having a fully fabricated device, etc.
An example process flow for using a metrology machine learning model is depicted in
After the substrate has been processed to produce or modify features on the substrate, optical metrology is performed on one or more of the features. See operation 205. The optical metrology technique is the same as the used in the training set for the model. In certain embodiments, operation 205 is performed by moving a beam or spot over multiple positions on the substrate. For example, a light beam may be raster scanned over the processed substrate
Per normal operation, the optical metrology technique generates optical signals having characteristics dictated by the feature(s) on the substrate. The optical signals may be reflectance values provided, e.g., as a function of wavelength, polarization state, azimuthal position, etc. Regardless of their information content, the resulting optical signals are provided to the metrology machine learning model, which may have been trained as described above. See operation 207. With using the optical signals, the machine learning model outputs information identifying one or more geometric characteristics of the one or more features. See operation 209.
In certain embodiments, operation 205 is performed only after first performing pattern recognition on the one or more features on the processed substrate. The pattern recognition process may determine that a pattern of the one or more features conforms with an expected pattern. If the pattern does not conform, then the optical metrology operation may be dispensed with.
Terminology
The terms “semiconductor wafer,” “wafer,” “substrate,” “wafer substrate” and “partially fabricated integrated circuit” may be used interchangeably. Those of ordinary skill in the art understand that the term “partially fabricated integrated circuit” can refer to a semiconductor wafer during any of many stages of integrated circuit fabrication thereon. A wafer or substrate used in the semiconductor device industry typically has a diameter of 200 mm, or 300 mm, or 450 mm. This detailed description assumes the embodiments are implemented on a wafer. However, the disclosure is not so limited. The work piece may be of various shapes, sizes, and materials. Besides semiconductor wafers, other work pieces that may take advantage of the disclosed embodiments include various articles such as printed circuit boards, magnetic recording media, magnetic recording sensors, mirrors, optical elements, micro-mechanical devices and the like. Further, the term “coupon” is sometimes used generically herein to describe any substrate, typically including those used for testing a process and/or for generating model training set data.
A “semiconductor device fabrication operation” as used herein is a unit operation performed during fabrication of semiconductor devices. Typically, the overall fabrication process includes multiple semiconductor device fabrication operations, each performed in its own semiconductor fabrication tool such as an etch and/or deposition plasma reactor, an electroplating cell, a chemical mechanical planarization tool, a wet etch tool, and the like. Categories of semiconductor device fabrication operations include subtractive processes, such as etch processes and planarization processes, and material additive processes, such as deposition processes. In the context of etch processes, a substrate etch process includes processes that etch a mask layer or, more generally, processes that etch any layer of material previously deposited on and/or otherwise residing on a substrate surface. Such etch process may etch a stack of layers in the substrate. The models described herein may be used to replace destructive metrology techniques with non-destructive optical metrology techniques for any type of semiconductor device fabrication operation.
A “metrology result” as used herein refers to a result produced, at least in part, by measuring features of the processed substrate. The measurement may be made while or after performing the semiconductor device fabrication operation in a reaction chamber operating under the set of process parameter values. In certain embodiments, measuring features of the processed substrate produces profile coordinates. In such embodiments, measuring features of the processed substrate may include performing microscopy (e.g., SEM, x-SEM, TEM, STEM, CD-SEM, CD-SAXS, REM, AFM). Such techniques may be employed to generate profile coordinates directly or indirectly such as by way of a set of geometric profile parameters characterizing a geometry of the feature in the processed substrate (e.g., critical dimension, side wall angles, depth, etc.).
Optical metrology results are produced using optical metrology on features of a processed substrate. In certain embodiments, the optical metrology result is produced by performing reflectometry, dome scatterometry, angle-resolved scatterometry, small-angle X-ray scatterometry, and/or ellipsometry on a processed substrate. An optical metrology output may include any of various values obtained from a detector in an optical metrology tool. examples include reflectance values, optionally over a range of wavelengths (reflectance spectra), polarization information, angular/positional information about the detected optical signals (e.g., angle with respect to the plane of a substrate or a detector or with respect to angle of incidence), ocd parameters), etc. When using optical metrology, the system may obtain profile coordinates by calculating them from measured optical metrology signals using a metrology machine learning model as described herein.
In certain embodiments, the metrology result is provided as a time sequence of measured geometric profiles, reflectance or ellipsometric data, or profile parameters of a substrate feature. These measured metrology results are produced at the different durations of the semiconductor device fabrication operation.
The geometric information about a feature may take various forms. It can be provided as profile coordinates in two or three dimensions and/or a more global parameter such as trench depth, side wall angle, and the like. For repeating structures, one-dimensional information may be provided such as pitch, critical dimension, and the like. In the case of two-dimensional representations, the geometric information of a feature may be a profile (viewed as a cross-section of the substrate in an x-z plane) or as a contour (viewed from above the plane of the substrate in an x-y plane).
Machine learning model—A machine learning model is a trained computational model that takes optical metrology outputs or data (e.g., scatterometry data) and predicts a geometric characteristic of one or more features on a sample. Examples of machine learning models include a random forests models, including deep random forests, neural networks, including recurrent neural networks and convolutional neural networks, restricted Boltzmann machines, recurrent tensor networks, and gradient boosted trees. Each of these has a structure and associated logical sequence of operations (algorithm) well known to those of skill in the art. For example, a neural network includes multiple computational nodes organized in layers configured to successively pass computed results from one layer to the next, on a node-by-node basis. A given node may have connections to one or more nodes in an upstream layer and connections to one or more modes in a downstream layer. The layers may be organized as an input layer, an output layer, and one or more hidden layers. The input side of a neural network may be structured in a way to implement a convolutional neural network. Nodes in a hidden layer or output layer receive inputs from one or other nodes in an upstream layer. Each node operates on its inputs via a function (sometimes an activation function) to produce an output for transmitting to one or more downstream nodes in the next successive layer. Training applies weights to the individual connections between upstream and downstream nodes. The term “classifier” (or classification model) is sometimes used to describe all forms of classification model including deep learning models (e.g., neural networks having many layer) as well as random forests models. More details may be found in “Hands-On Machine Learning with Scikit-Learn and TensorFlow: Concepts, Tools, and Techniques to Build Intelligent Systems,” 1st Edition, by A. Geron, O'Reilly Media, Inc. (2017); and Deep Learning by Goodfellow et al., MIT Press (2016) which are incorporated herein by reference in their entireties.
Deep learning model—A deep learning model as used herein is a form of classification model. It is also a form of machine learning model. It may be implemented in various forms such as by a neural network (e.g., a convolutional neural network), etc. In general, though not necessarily, it includes multiple layers. Each such layer includes multiple processing nodes, and the layers process in sequence, with nodes of layers closer to the model input layer processing before nodes of layers closer to the model output. In various embodiments, one layer feeds to the next, etc. The output layer may include nodes that represent various classifications. In some implementations, a deep learning model is a model that takes data with very little preprocessing.
In various embodiments, a deep learning model has significant depth (e.g., five or more layers or ten or more layers) and can classify a large or heterogeneous sets of optical metrology data. In some contexts, the term “deep” means that model has more than two (or more than three or more than four or more than five) layers of processing nodes that receive values from preceding layers (or as direct inputs) and that output values to succeeding layers (or the final output). Interior nodes are often “hidden” in the sense that their input and output values are not visible outside the model. In various embodiments, the operation of the hidden nodes is not monitored or recorded during operation.
The nodes and connections of a deep learning model can be trained and retrained without redesigning their number, arrangement, interface with image inputs, etc.
As indicated, in various implementations, the node layers may collectively form a neural network, although many deep learning models have other structures and formats. Some embodiments of deep learning models do not have a layered structure, in which case the above characterization of “deep” as having many layers is not relevant.
Context for Disclosed Computational Embodiments
Certain embodiments disclosed herein relate to systems for generating and/or using metrology models as disclosed herein. Certain embodiments disclosed herein relate to methods for generating and/or using a model implemented on such systems. A system for generating a model may be configured to analyze data for calibrating or optimizing the expressions or relationships used to represent the effects of a semiconductor device fabrication operation on a substrate. A programmed system for using a model may be configured to (i) receive input such as optical metrology data, and (ii) execute instructions that determine characteristics of features on the substrate. A programmed system for using a model may be configured to (i) receive optical metrology data obtain for a plurality of features, (ii) receive SEM, TEM, and/or CD-SAXS data for the same plurality of features, and (iii) execute instructions for training a metrology machine learning model using the data in (i) and (ii).
Many types of computing systems having any of various computer architectures may be employed as the disclosed systems for implementing models, and algorithms for generating, using, and/or optimizing such models. For example, the systems may include software components executing on one or more general purpose processors or specially designed processors such as programmable logic devices (e.g., Field Programmable Gate Arrays (FPGAs)). Further, the systems may be implemented on a single device or distributed across multiple devices. The functions of the computational elements may be merged into one another or further split into multiple sub-modules.
In some embodiments, code executed during generation or execution of a model on an appropriately programmed system can be embodied in the form of software elements which can be stored in a nonvolatile storage medium (such as optical disk, flash storage device, mobile hard disk, etc.), including a number of instructions for making a computer device (such as personal computers, servers, network equipment, etc.).
At one level a software element is implemented as a set of commands prepared by the programmer/developer. However, the module software that can be executed by the computer hardware is executable code committed to memory using “machine codes” selected from the specific machine language instruction set, or “native instructions,” designed into the hardware processor. The machine language instruction set, or native instruction set, is known to, and essentially built into, the hardware processor(s). This is the “language” by which the system and application software communicates with the hardware processors. Each native instruction is a discrete code that is recognized by the processing architecture and that can specify particular registers for arithmetic, addressing, or control functions; particular memory locations or offsets; and particular addressing modes used to interpret operands. More complex operations are built up by combining these simple native instructions, which are executed sequentially, or as otherwise directed by control flow instructions.
The inter-relationship between the executable software instructions and the hardware processor is structural. In other words, the instructions per se are a series of symbols or numeric values. They do not intrinsically convey any information. It is the processor, which by design was preconfigured to interpret the symbols/numeric values, which imparts meaning to the instructions.
The models used herein may be configured to execute on a single machine at a single location, on multiple machines at a single location, or on multiple machines at multiple locations. When multiple machines are employed, the individual machines may be tailored for their particular tasks. For example, operations requiring large blocks of code and/or significant processing capacity may be implemented on large and/or stationary machines.
In addition, certain embodiments relate to tangible and/or non-transitory computer readable media or computer program products that include program instructions and/or data (including data structures) for performing various computer-implemented operations. Examples of computer-readable media include, but are not limited to, semiconductor memory devices, phase-change devices, magnetic media such as disk drives, magnetic tape, optical media such as CDs, magneto-optical media, and hardware devices that are specially configured to store and perform program instructions, such as read-only memory devices (ROM) and random access memory (RAM). The computer readable media may be directly controlled by an end user or the media may be indirectly controlled by the end user. Examples of directly controlled media include the media located at a user facility and/or media that are not shared with other entities. Examples of indirectly controlled media include media that is indirectly accessible to the user via an external network and/or via a service providing shared resources such as the “cloud.” Examples of program instructions include both machine code, such as produced by a compiler, and files containing higher level code that may be executed by the computer using an interpreter.
In various embodiments, the data or information employed in the disclosed methods and apparatus is provided in an electronic format. Such data or information may include design layouts, fixed parameter values, floated parameter values, feature profiles, metrology results, and the like. In some cases, a collection of instructions provided in electronic format defines an algorithm for executing a metrology machine learning model such as a neural network. As used herein, data or other information provided in electronic format is available for storage on a machine and transmission between machines. Conventionally, data in electronic format is provided digitally and may be stored as bits and/or bytes in various data structures, lists, databases, etc. The data may be embodied electronically, optically, etc.
In certain embodiments, a model can each be viewed as a form of application software that interfaces with a user and with system software. System software typically interfaces with computer hardware and associated memory. In certain embodiments, the system software includes operating system software and/or firmware, as well as any middleware and drivers installed in the system. The system software provides basic non-task-specific functions of the computer. In contrast, the modules and other application software are used to accomplish specific tasks. Each native instruction for a module is stored in a memory device and is represented by a numeric value.
An example computer system 800 is depicted in
Program code may be stored in non-transitory media such as persistent storage 810 or memory 808 or both. One or more processors 804 reads program code from one or more non-transitory media and executes the code to enable the computer system to accomplish the methods performed by the embodiments herein, such as those involved with generating or using a process simulation model as described herein. Those skilled in the art will understand that the processor may accept source code, such as statements for executing training and/or interpretation operations, and interpret or compile the source code into machine code that is understandable at the hardware gate level of the processor. A bus couples the I/O subsystem 802, the processor 804, peripheral devices 806, memory 808, and persistent storage 810. In some embodiments, at least some of the computer system features, e.g., processor(s) 804, memory 808, and/or storage 810, are implemented in a cloud or edge network that is physically distinct from an I/O subsystem, if any. In some implementations, the some of the computer system features, e.g., e.g., processor(s) 804, memory 808, and/or storage 810, are distributed over multiple physical devices.
Context for Metrology Tool Embodiments
In various embodiments, a metrology system includes an optical metrology tool and a metrology machine learning model such as one produced using a training process as described herein. The optical metrology tool may include various configurations, but typically includes a light source and a detector for detecting light reflected from, transmitted through, and/or scattered off of features on a substrate being probed. The light source may be monochromatic or polychromatic. The wavelength(s) produced may be in the infrared, visible, ultraviolet, or x-ray range. It may be unpolarized, linearly polarized, circularly polarized, etc. The optical metrology tool may additionally include one or more optical elements for directing light onto substrate features and/or collecting and/or directly light that has interacted with the substrate features. The optical metrology tool may additionally include one or more processors or other processing elements to capture and/or interpret signals from the detector.
In certain embodiments, an optical metrology tool includes a reflectance/ellipsometric spectra generator is part of a tool such as the YieldStar™ scatterometer products available from ASML Netherlands B.V., Veldhoven, The Netherlands. See e.g., Cramer et al., “High-NA optical CD metrology on small in-cell targets enabling improved higher order dose control and process control for logic,” Proceedings of SPIE, 10145, Metrology, Inspection, and Process Control for Microlithography XXXI, 101451B (28 Mar. 2017), incorporated herein by reference in its entirety.
A system such as metrology system 401 may determine a feature's etch profile using optical information taken directly from a test sample, such as in the case of OCD measurements, followed by treatment of that data by a model or relationship as described herein.
Conclusion
In the description, numerous specific details were set forth in order to provide a thorough understanding of the presented embodiments. The disclosed embodiments may be practiced without some or all of these specific details. In other instances, well-known process operations were not described in detail to not unnecessarily obscure the disclosed embodiments. While the disclosed embodiments were described in conjunction with the specific embodiments, it will be understood that the specific embodiments are not intended to limit the disclosed embodiments.
A PCT Request Form is filed concurrently with this specification as part of the present application. Each application that the present application claims benefit of or priority to as identified in the concurrently filed PCT Request Form is incorporated by reference herein in its entirety and for all purposes.
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/US2019/026851 | 4/10/2019 | WO |
Publishing Document | Publishing Date | Country | Kind |
---|---|---|---|
WO2019/200015 | 10/17/2019 | WO | A |
Number | Name | Date | Kind |
---|---|---|---|
5114233 | Clark et al. | May 1992 | A |
5421934 | Misaka et al. | Jun 1995 | A |
6268226 | Angell et al. | Jul 2001 | B1 |
6410351 | Bode et al. | Jun 2002 | B1 |
6650423 | Markle et al. | Nov 2003 | B1 |
6684382 | Liu | Jan 2004 | B2 |
6753115 | Zhang et al. | Jun 2004 | B2 |
6765651 | Fiekowsky et al. | Jul 2004 | B1 |
6804572 | Cooperberg et al. | Oct 2004 | B1 |
6869739 | Ausschnitt et al. | Mar 2005 | B1 |
6903826 | Usui et al. | Jun 2005 | B2 |
7139632 | Cooperberg et al. | Nov 2006 | B2 |
7402257 | Sonderman et al. | Jul 2008 | B1 |
7504182 | Stewart et al. | Mar 2009 | B2 |
7588946 | Tso et al. | Sep 2009 | B2 |
7600212 | Zach et al. | Oct 2009 | B2 |
7739651 | Melvin, III et al. | Jun 2010 | B2 |
7812966 | Hoffmann et al. | Oct 2010 | B2 |
7849423 | Yenikaya et al. | Dec 2010 | B1 |
7921383 | Wei | Apr 2011 | B1 |
7962867 | White et al. | Jun 2011 | B2 |
7967995 | Funk et al. | Jun 2011 | B2 |
8001512 | White | Aug 2011 | B1 |
8102408 | Verma et al. | Jan 2012 | B2 |
8279409 | Sezginer et al. | Oct 2012 | B1 |
8832610 | Ye et al. | Sep 2014 | B2 |
9015016 | Lorenz et al. | Apr 2015 | B2 |
9448470 | Liu et al. | Sep 2016 | B2 |
9471746 | Rieger et al. | Oct 2016 | B2 |
9547740 | Moroz et al. | Jan 2017 | B2 |
9659126 | Greiner et al. | May 2017 | B2 |
9792393 | Tetiker et al. | Oct 2017 | B2 |
9996647 | Tetiker et al. | Jun 2018 | B2 |
10032681 | Bailey, III et al. | Jul 2018 | B2 |
10197908 | Sriraman et al. | Feb 2019 | B2 |
10254641 | Mailfert et al. | Apr 2019 | B2 |
10303830 | Tetiker | May 2019 | B2 |
10386828 | Tetiker et al. | Aug 2019 | B2 |
10409946 | Ünal et al. | Sep 2019 | B2 |
10495967 | Huang et al. | Dec 2019 | B2 |
10534257 | Tetiker et al. | Jan 2020 | B2 |
10572697 | Feng et al. | Feb 2020 | B2 |
10585347 | Sriraman et al. | Mar 2020 | B2 |
10712672 | Jochemsen et al. | Jul 2020 | B2 |
10977405 | Bowes et al. | Apr 2021 | B2 |
10997345 | Feng et al. | May 2021 | B2 |
11624981 | Sriraman et al. | Apr 2023 | B2 |
11704463 | Feng et al. | Jul 2023 | B2 |
20020012876 | Angelopoulos et al. | Jan 2002 | A1 |
20030008507 | Bell et al. | Jan 2003 | A1 |
20030113766 | Pepper et al. | Jun 2003 | A1 |
20030208728 | Pierrat | Nov 2003 | A1 |
20040019872 | Lippincott et al. | Jan 2004 | A1 |
20050074907 | Kriz et al. | Apr 2005 | A1 |
20050192914 | Drege et al. | Sep 2005 | A1 |
20060064280 | Vi Vuong et al. | Mar 2006 | A1 |
20060141484 | Rucker et al. | Jun 2006 | A1 |
20060208205 | Chen et al. | Sep 2006 | A1 |
20070031745 | Ye et al. | Feb 2007 | A1 |
20070050749 | Ye et al. | Mar 2007 | A1 |
20070061772 | Ye et al. | Mar 2007 | A1 |
20070061773 | Ye et al. | Mar 2007 | A1 |
20070087273 | Look et al. | Apr 2007 | A1 |
20070249071 | Lian et al. | Oct 2007 | A1 |
20070281478 | Ikegami et al. | Dec 2007 | A1 |
20080007739 | Vi Vuong et al. | Jan 2008 | A1 |
20080035608 | Thomas et al. | Feb 2008 | A1 |
20080072207 | Verma et al. | Mar 2008 | A1 |
20080078948 | Saito | Apr 2008 | A1 |
20080243730 | Bischoff et al. | Oct 2008 | A1 |
20090048813 | Ichikawa et al. | Feb 2009 | A1 |
20090087143 | Jeon et al. | Apr 2009 | A1 |
20090126589 | Maxwell et al. | May 2009 | A1 |
20090242513 | Funk et al. | Oct 2009 | A1 |
20090253222 | Morisawa et al. | Oct 2009 | A1 |
20110022215 | Keil et al. | Jan 2011 | A1 |
20110091815 | Dunn et al. | Apr 2011 | A1 |
20110138343 | Granik | Jun 2011 | A1 |
20110159690 | Chandrashekar et al. | Jun 2011 | A1 |
20110188732 | Stroessner | Aug 2011 | A1 |
20110292375 | Marx | Dec 2011 | A1 |
20120002912 | Studenkov et al. | Jan 2012 | A1 |
20120009785 | Chandrashekar et al. | Jan 2012 | A1 |
20120022836 | Ferns et al. | Jan 2012 | A1 |
20120179282 | Sarma et al. | Jul 2012 | A1 |
20120228125 | Wu | Sep 2012 | A1 |
20130171822 | Chandrashekar et al. | Jul 2013 | A1 |
20140032463 | Jin et al. | Jan 2014 | A1 |
20140123084 | Tang et al. | May 2014 | A1 |
20150056803 | Chandrashekar et al. | Feb 2015 | A1 |
20150079500 | Shih et al. | Mar 2015 | A1 |
20150142395 | Cao et al. | May 2015 | A1 |
20150154145 | Watanabe et al. | Jun 2015 | A1 |
20150356233 | Fouquet et al. | Dec 2015 | A1 |
20150366523 | Ben-Haim | Dec 2015 | A1 |
20150371134 | Chien et al. | Dec 2015 | A1 |
20160078166 | Ünal et al. | Mar 2016 | A1 |
20160091769 | Rozbicki | Mar 2016 | A1 |
20160189971 | Yi et al. | Jun 2016 | A1 |
20160217233 | Kamon et al. | Jul 2016 | A1 |
20160284077 | Brill | Sep 2016 | A1 |
20160313651 | Middlebrooks et al. | Oct 2016 | A1 |
20160322267 | Kim et al. | Nov 2016 | A1 |
20160335384 | Song et al. | Nov 2016 | A1 |
20170004233 | Han et al. | Jan 2017 | A1 |
20170039308 | Moroz et al. | Feb 2017 | A1 |
20170115556 | Shim et al. | Apr 2017 | A1 |
20170147724 | Regli et al. | May 2017 | A1 |
20170176983 | Tetiker et al. | Jun 2017 | A1 |
20170220723 | Song et al. | Aug 2017 | A1 |
20170228482 | Tetiker et al. | Aug 2017 | A1 |
20170256463 | Bailey, III et al. | Sep 2017 | A1 |
20170351952 | Zhang et al. | Dec 2017 | A1 |
20170363950 | Sriraman et al. | Dec 2017 | A1 |
20170365513 | Yang et al. | Dec 2017 | A1 |
20170371991 | Tetiker et al. | Dec 2017 | A1 |
20180157161 | Mailfert et al. | Jun 2018 | A1 |
20180182632 | Ye et al. | Jun 2018 | A1 |
20180239851 | Ypma et al. | Aug 2018 | A1 |
20180253015 | Ypma et al. | Sep 2018 | A1 |
20180260509 | Tetiker et al. | Sep 2018 | A1 |
20180284597 | Weisbuch et al. | Oct 2018 | A1 |
20180314148 | Tetiker et al. | Nov 2018 | A1 |
20180364589 | Chen et al. | Dec 2018 | A1 |
20190049937 | Tetiker et al. | Feb 2019 | A1 |
20190204749 | Tel et al. | Jul 2019 | A1 |
20190250501 | Sriraman et al. | Aug 2019 | A1 |
20190311083 | Feng et al. | Oct 2019 | A1 |
20200218844 | Feng et al. | Jul 2020 | A1 |
20200242209 | Bowes et al. | Jul 2020 | A1 |
20210080838 | Tel et al. | Mar 2021 | A1 |
20210150116 | Fan et al. | May 2021 | A1 |
20210157228 | Sriraman et al. | May 2021 | A1 |
20210165332 | Van Adrichem et al. | Jun 2021 | A1 |
20210216695 | Feng et al. | Jul 2021 | A1 |
20210374936 | Koopman et al. | Dec 2021 | A1 |
20230205076 | Sriraman et al. | Jun 2023 | A1 |
Number | Date | Country |
---|---|---|
1501178 | Jun 2004 | CN |
1739185 | Feb 2006 | CN |
1799045 | Jul 2006 | CN |
1868043 | Nov 2006 | CN |
1940715 | Apr 2007 | CN |
101151719 | Mar 2008 | CN |
101313308 | Nov 2008 | CN |
101359611 | Feb 2009 | CN |
101976045 | Feb 2011 | CN |
103109314 | May 2013 | CN |
104103510 | Oct 2014 | CN |
104518753 | Apr 2015 | CN |
104736744 | Jun 2015 | CN |
107526864 | Dec 2017 | CN |
1720062 | Nov 2006 | EP |
2549523 | Jan 2013 | EP |
2980646 | Feb 2016 | EP |
2002541663 | Dec 2002 | JP |
2003092237 | Mar 2003 | JP |
2004253516 | Sep 2004 | JP |
2005277361 | Oct 2005 | JP |
2005536074 | Nov 2005 | JP |
2007219208 | Aug 2007 | JP |
2008091673 | Apr 2008 | JP |
2009152269 | Jul 2009 | JP |
2009294308 | Dec 2009 | JP |
2010034393 | Feb 2010 | JP |
2012186394 | Sep 2012 | JP |
WO2011115023 | Jun 2013 | JP |
2015103769 | Jun 2015 | JP |
2015528569 | Sep 2015 | JP |
2016033656 | Mar 2016 | JP |
2017135365 | Aug 2017 | JP |
2017-227892 | Dec 2017 | JP |
20010041543 | May 2001 | KR |
20060050931 | May 2006 | KR |
20080005548 | Jan 2008 | KR |
20130082110 | Jul 2013 | KR |
101460375 | Nov 2014 | KR |
20150060524 | Jun 2015 | KR |
20170017789 | Feb 2017 | KR |
200540573 | Dec 2005 | TW |
200710986 | Mar 2007 | TW |
201011474 | Mar 2010 | TW |
201327659 | Jul 2013 | TW |
201415003 | Apr 2014 | TW |
201433878 | Sep 2014 | TW |
201435621 | Sep 2014 | TW |
201606465 | Feb 2016 | TW |
I528201 | Apr 2016 | TW |
201620009 | Jun 2016 | TW |
201633410 | Sep 2016 | TW |
201732459 | Sep 2017 | TW |
201742250 | Dec 2017 | TW |
201802574 | Jan 2018 | TW |
201809856 | Mar 2018 | TW |
WO-2004015364 | Feb 2004 | WO |
WO-2004015727 | Feb 2004 | WO |
WO 2006104655 | Oct 2006 | WO |
WO-2014027354 | Feb 2014 | WO |
WO-2015027088 | Feb 2015 | WO |
WO-2016162157 | Oct 2016 | WO |
WO 2018204193 | Nov 2018 | WO |
Entry |
---|
U.S. Office Action, dated Jan. 25, 2018, issued in U.S. Appl. No. 14/972,969. |
U.S. Final Office Action, dated Aug. 27, 2018, issued in U.S. Appl. No. 14/972,969. |
U.S. Office Action, dated Jan. 10, 2019, issued in U.S. Appl. No. 14/972,969. |
U.S. Notice of Allowance, dated Apr. 5, 2019, issued in U.S. Appl. No. 14/972,969. |
U.S. Office Action, dated Mar. 22, 2017, issued in U.S. Appl. No. 15/018,708. |
U.S. Notice of Allowance, dated Jun. 7, 2017, issued in U.S. Appl. No. 15/018,708. |
U.S. Notice of Allowance, dated Feb. 6, 2018 issued in U.S. Appl. No. 15/698,458. |
U.S. Office Action, dated Oct. 2, 2017, issued in U.S. Appl. No. 15/698,458. |
Office Action dated Jun. 14, 2018 issued in U.S. Appl. No. 15/972,063. |
Final Office Action dated Nov. 7, 2018 issued in U.S. Appl. No. 15/972,063. |
Notice of Allowance dated Feb. 1, 2019 issued in U.S. Appl. No. 15/972,063. |
U.S. Office Action, dated Aug. 10, 2017, issued in U.S. Appl. No. 15/059,073. |
Notice of Allowance dated Jan. 11, 2018 issued in U.S. Appl. No. 15/059,073. |
U.S. Office Action, dated Dec. 7, 2017, issued in U.S. Appl. No. 15/188,910. |
U.S. Final Office Action, dated May 23, 2018 issued in U.S. Appl. No. 15/188,910. |
U.S. Notice of Allowance, dated Sep. 27, 2018 issued in U.S. Appl. No. 15/188,910. |
U.S. Office Action, dated Jul. 10, 2019 issued in U.S. Appl. No. 16/224,651. |
U.S. Notice of Allowance, dated Oct. 30, 2019 issued in U.S. Appl. No. 16/224,651. |
U.S. Office Action, dated Jul. 11, 2018 issued in U.S. Appl. No. 15/367,060. |
U.S. Notice of Allowance, dated Nov. 26, 2018 issued in U.S. Appl. No. 15/367,060. |
U.S. Notice of Allowance, dated Jul. 23, 2019 issued in U.S. Appl. No. 15/583,610. |
U.S. Office Action dated Jul. 10, 2019 issued in U.S. Appl. No. 15/946,940. |
U.S. Notice of Allowance dated Oct. 16, 2019 issued in U.S. Appl. No. 15/946,940. |
U.S. Office Action dated Oct. 8, 2020 issued in U.S. Appl. No. 16/741,735. |
U.S. Office Action, dated Mar. 4, 2019 issued in U.S. Appl. No. 15/673,321. |
U.S. Office Action dated Mar. 17, 2020 issued in U.S. Appl. No. 16/260,870. |
U.S. Final Office Action dated Aug. 27, 2020 issued in U.S. Appl. No. 16/260,870. |
Chinese First Office Action dated Dec. 2, 2019 issued in Application No. CN 201611166040.9. |
Chinese First Office Action dated Mar. 11, 2020 issued in Application No. CN 201710475240.0. |
Chinese Second Office Action dated Aug. 10, 2020 issued in Application No. CN 201710475240.0. |
Singapore Notice of Eligibility for Grant and Supplemental Examination Report dated Jun. 22, 2020 issued in Application No. SG 10201705049V. |
Chinese First Office Action dated Mar. 6, 2020 issued in Application No. CN 201710068974.7. |
Taiwanese Notice of Allowance dated Jul. 17, 2020 issued in Application No. TW 106103883. |
International Search Report and Written Opinion dated Aug. 10, 2018 issued in Application No. PCT/US2018/029874. |
International Preliminary Report on Patentability dated Nov. 14, 2019 issued in Application No. PCT/US2018/029874. |
Chinese First Office Action dated Feb. 25, 2019 issued in Application No. CN 201710121052.8. |
International Search Report and Written Opinion dated Jul. 22, 2019 issued in Application No. PCT/US2019/026396. |
International Preliminary Report on Patentability dated Oct. 22, 2020 issued in Application No. PCT/US2019/026396. |
International Search Report and Written Opinion dated Jul. 31, 2019 issued in Application No. PCT/US2019/026851. |
International Preliminary Report on Patentability dated Oct. 22, 2020 issued in Application No. PCT/US2019/026851. |
International Search Report and Written Opinion dated Jul. 5, 2019 issued in Application No. PCT/US2019/025668. |
International Preliminary Report on Patentability dated Oct. 15, 2020 issued in Application No. PCT/US2019/025668. |
International Search Report and Written Opinion dated Jun. 5, 2020 issued in Application No. PCT/US2020/015464. |
Abdo (2010) “The feasibility of using image parameters for test pattern selection during OPC model calibration,” Proc. of SPIE, 7640:76401E-1-76401E-11. |
Byrd, R.H., RB. Schnabel, and G.A. Shultz, “Approximate Solution of the Trust Region Problem by Minimization over Two-Dimensional Subspaces,” Mathematical 5 Programming, vol. 40, pp. 247-263 (1988). |
Cooperberg, D.J., et al. (Sep./Oct. 2002) “Semiempirical profile simulation of aluminum etching in a Cl2 / BCl3 plasma,” Journal of Vacuum Science & Technology A., 20(5):1536-1556 [doi: http://dx.doi.org/10.1116/1.1494818]. |
Cramer et al., “High-NA optical CD metrology on small in-cell targets enabling improved higher order 10 dose control and process control for logic,” Proceedings of SPIE,10145, Metrology, Inspection, and Process Control for Microlithography XXXI, 101451B (Mar. 28, 2017). |
“SEMulator3D,” Product Brochure, COVENTOR, A Lam Research Company, 3 pp. (known as of Mar. 2, 2018) <URL:https://www.coventor.com/semiconductor-solutions/semulator3d/>. |
“SEMulator3D Advanced Modeling,” Web page, COVENTOR, A Lam Research Company, 5 pp. <URL: https://www.coventor.com/semiconductor-solutions/semulator3d/semulator3d-advanced-modeling/> (known as of Mar. 2, 2018). |
“SEMulator3D,” Web page, COVENTOR, A Lam Research Company, 5 pp. <URL:https://www.coventor.com/semiconductor-solutions/semulator3d/> (known as of Mar. 2, 2018). |
Goodlin et al. (May 2002) “Quantitative Analysis and Comparison of Endpoint Detection Based on Multiple Wavelength Analysis,” 201st Meeting of the Electrochemical Society, International Symposium on Plasma Processing XIV, Abs. 415, Philadelphia, PA, 30 pages. |
Hoekstra, R. et al. (Jul./Aug. 1997) “Integrated plasma equipment model for polysilicon etch profiles in an inductively coupled plasma reactor with subwafer and superwafer topography,” Journal of Vacuum Science & Technology A, 15(4):1913-1921. [doi:http://dx.doi.org/10.1116/1.580659]. |
Hoekstra, R. et al. (Jul./Aug. 1998) “Microtrenching resulting from specular reflection during chlorine etching of silicon,” Journal of Vacuum Science & Technology B, Nanotechnology and Microelectronics, 16(4):2102-2104 [doi: http://dx.doi.org/10.1116/1.590135]. |
Hoekstra, R. et al. (Nov. 1, 1998) “Comparison of two-dimensional and three-dimensional models for profile simulation of poly-Si etching of finite length trenches,” Journal of Vacuum Science & Technology A, 16(6):3274-3280 [doi:http://dx.doi.org/10.1116/1.581533]. |
Huard, C.M., et al. (Jan. 17, 2017) “Role of neutral transport in aspect ratio dependent plasma etching of three-dimensional features,” Journal of Vacuum Science & Technology A., 35(5):05C301-1-05C301-18. [doi: http://dx.doi.org/10.1116/1.4973953]. |
Kushner, M.J., (Sep. 18, 2009) “Hybrid modelling of low temperature plasmas for fundamental investigations and equipment design,” Journal of Physics D., 42(1904013):1-20 [doi: 10.1088/0022-3727/42/19/194013]. |
Lynn et al. (2009) “Virtual Metrology for Plasma Etch using Tool Variables,” IEEE, pp. 143-148. |
Lynn, Shane (Apr. 2011) “Virtual Metrology for Plasma Etch Processes,” A thesis submitted in partial fulfillment for the degree of Doctor of Philosophy in the Faculty of Science and Engineering, Electronic Engineering Department, National University of Ireland, Maynooth, 361 pages. |
Moharam et al. (Jul. 1981) “Rigorous coupled-wave analysis of planar-grating diffraction,” J. Opt. Soc. Am., 71(7):811-818. |
More, J.J. and D.C. Sorensen, “Computing a Trust Region Step,” SIAM Journal on Scientific and Statistical Computing, vol. 3, pp. 553-572, (1983). |
More, J.J., “The Levenberg-Marquardt Algorithm: Implementation and Theory,” Numerical Analysis, ed. G. A Watson, Lecture Notes in Mathematics 630, Springer Verlag, pp. 105-116 (1977). |
Radjenović et al. (2007) “3D Etching profile evolution simulations: Time dependence analysis of the profile charging during SiO2 etching in plasma,” 5th EU-Japan Joint Symposium on Plasma Process, Journal of Physics: Conference Series, 86:13 pages. |
Ringwood et al. (Feb. 2010) “Estimation and Control in Semiconductor Etch: Practice and Possibilities,” IEEE Transactions on Semiconductor Manufacturing, 23(1):87-98. |
Sankaran, A. et al. (Jan. 15, 2005) “Etching of porous and solid SiO2 in Ar/c-C4F8, O2/c-C4F8 and Ar/O2/c-C4F8 plasmas,” Journal of Applied Physics, 97(2):023307-1-023307-10. [doi: http://dx.doi.org/10.1063/1.1834979] [retrieved on Jan. 29, 2005]. |
Sankaran, A. et al. (Jul./Aug. 2004) “Integrated feature scale modeling of plasma processing of porous and solid SiO2. I. Fluorocarbon etching,” Journal of Vacuum Science & Technology A., 22(4):1242-1259 [doi: http://dx.doi.org/10.1116/1.1764821]. |
Yue et al. (Jan./Feb. 2001) “Plasma etching endpoint detection using multiple wavelengths for small open-area wafers,” J. Vac. Sci. Technol. A, 19(1):66-75. |
Zeng (2012) “Statistical Methods for Enhanced Metrology in Semiconductor/Photovoltaic Manufacturing,” A dissertation submitted in partial satisfaction of the requirements for the degree of Doctor of Philosophy in Engineering—Electrical Engineering and Computer Sciences in the Graduate Division of the University of California, Berkeley, 171 pages. |
Zhang, D. et al. (Mar./Apr. 2001) “Investigations of surface reactions during C2F6 plasma etching of SiO2 with equipment and feature scale models,” Journal of Vacuum Science & Technology A, 19(2):524-538 [doi: http://dx.doi.org/10.1116/1.1349728]. |
Zhang, Y., (Sep. 30, 2015) Doctoral Dissertation of “Low Temperature Plasma Etching Control through Ion Energy Angular Distribution and 3-Dimensional Profile Simulation,” Dept. of Electrical Engineering at the University of Michigan, pp. 49-71, Appendix B. pp. 243-248 [doi:http://hdl.handle.net/2027.42/113432]. |
U.S. Appl. No. 15/733,728, filed Oct. 6, 2020, Feng et al. |
Chinese First Office Action dated Aug. 16, 2021 issued in Application No. CN 201880044397.0. |
Chinese Third Office Action dated Nov. 18, 2020 issued in Application No. CN 201710475240.0. |
CN Office Action dated Apr. 26, 2022, in Application No. CN201880044397 with English translation. |
CN Office action dated Aug. 16, 2021 in CN Application No. CN201880044397 with English translation. |
CN Office action dated Jan. 13, 2022 in CN Application No. CN201880044397 with English translation. |
European Extended Search Report dated Feb. 10, 2021 issued in EP Application No. EP 18794270. |
Japanese First Office Action dated Feb. 16, 2021 issued in Application No. JP 2017-020050. |
Japanese First Office Action dated Mar. 24, 2021 issued in Application No. JP 2017-037874. |
JP Office Action dated Jun. 14, 2022 in Application No. JP20190559822. |
JP Office Action dated Oct. 18, 2022, in Application No. 2021-152896 with English translation. |
KR Office Action dated Apr. 15, 2022, in application No. KR1020170026906 with English Translation. |
KR Office Action dated Dec. 16, 2022, in Application No. KR10-2019-7035372 with English translation. |
KR Office Action dated Dec. 29, 2022 in Application No. KR10-2022-0086361 with English translation. |
KR Office Action dated May 16, 2022, in Application No. KR1020170077116 with English translation. |
KR Office Action dated Oct. 19, 2021, in application No. KR10-2017-0077116 with English Translation. |
KR Office Action dated Oct. 22, 2021, in application No. KR10-2017-0026906 with English Translation. |
Taiwanese First Office Action dated Dec. 10, 2020 issued in Application No. TW 106120279. |
Taiwanese First Office Action dated Dec. 7, 2020 issued in Application No. TW 106106609. |
TW Office Action dated Aug. 17, 2022, in Application No. TW108112113 With English Translation. |
TW Office Action dated Aug. 10, 2023, in application No. TW108112271 with English Translation. |
TW Office Action dated Jan. 10, 2023 in Application No. TW108112271 with English translation. |
TW Office Action dated Jan. 19, 2023, in Application No. TW108112113 with English translation. |
TW Office Action dated Mar. 22, 2022, in Application No. TW107114630 with English translation. |
TW Office Action dated May 23, 2023, in Application No. TW109102831 with English translation. |
TW Office Action dated May 30, 2023 in Application No. TW108112113 with English translation. |
TW Office Action dated Sep. 7, 2022, in Application No. TW108112271 with English translation. |
U.S. Corrected Notice of Allowance dated Jun. 12, 2023, in U.S. Appl. No. 17/301,345. |
U.S. Final office Action dated Jun. 17, 2022 in U.S. Appl. No. 17/045,168. |
U.S. Non-Final office Action dated Sep. 9, 2022 in U.S. Appl. No. 17/301,345. |
U.S. Notice of Allowance dated Dec. 9, 2020 issued in U.S. Appl. No. 16/260,870. |
U.S. Notice of Allowance dated Feb. 15, 2023 in U.S. Appl. No. 17/301,345. |
U.S. Notice of Allowance dated Jan. 13, 2021 issued in U.S. Appl. No. 16/741,735. |
U.S. Notice of Allowance dated Nov. 30, 2022 in U.S. Appl. No. 17/045,168. |
U.S. Office Action dated Feb. 22, 2022 issued in U.S. Appl. No. 17/045,168. |
CN Office Action dated Dec. 8, 2023 in CN Application No. 201980025007.X with English Translation. |
KR Office Action dated Dec. 15, 2023 in KR Application No. 10-2020-7032501 with English Translation. |
Number | Date | Country | |
---|---|---|---|
20210035833 A1 | Feb 2021 | US |
Number | Date | Country | |
---|---|---|---|
62655728 | Apr 2018 | US |