[go: up one dir, main page]
More Web Proxy on the site http://driver.im/

US11847776B2 - System using film thickness estimation from machine learning based processing of substrate images - Google Patents

System using film thickness estimation from machine learning based processing of substrate images Download PDF

Info

Publication number
US11847776B2
US11847776B2 US17/359,345 US202117359345A US11847776B2 US 11847776 B2 US11847776 B2 US 11847776B2 US 202117359345 A US202117359345 A US 202117359345A US 11847776 B2 US11847776 B2 US 11847776B2
Authority
US
United States
Prior art keywords
substrate
polishing
color image
die
region
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active, expires
Application number
US17/359,345
Other versions
US20210407066A1 (en
Inventor
Sivakumar Dhandapani
Arash Alahgholipouromrani
Dominic J. Benvegnu
Jun Qian
Kiran Lall Shrestha
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Applied Materials Inc
Original Assignee
Applied Materials Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Applied Materials Inc filed Critical Applied Materials Inc
Priority to US17/359,345 priority Critical patent/US11847776B2/en
Assigned to APPLIED MATERIALS, INC. reassignment APPLIED MATERIALS, INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: ALAHGHOLIPOUROMRANI, Arash, SHRESTHA, KIRAN LALL, DHANDAPANI, SIVAKUMAR, BENVEGNU, DOMINIC J., QIAN, JUN
Publication of US20210407066A1 publication Critical patent/US20210407066A1/en
Priority to US18/500,811 priority patent/US20240062364A1/en
Application granted granted Critical
Publication of US11847776B2 publication Critical patent/US11847776B2/en
Active legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • G06T7/0004Industrial image inspection
    • G06T7/001Industrial image inspection using an image reference approach
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B24GRINDING; POLISHING
    • B24BMACHINES, DEVICES, OR PROCESSES FOR GRINDING OR POLISHING; DRESSING OR CONDITIONING OF ABRADING SURFACES; FEEDING OF GRINDING, POLISHING, OR LAPPING AGENTS
    • B24B37/00Lapping machines or devices; Accessories
    • B24B37/005Control means for lapping machines or devices
    • B24B37/013Devices or means for detecting lapping completion
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N20/00Machine learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • G06T7/0004Industrial image inspection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/60Analysis of geometric attributes
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/60Analysis of geometric attributes
    • G06T7/62Analysis of geometric attributes of area, perimeter, diameter or volume
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10024Color image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20021Dividing image into blocks, subimages or windows
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30108Industrial image inspection
    • G06T2207/30148Semiconductor; IC; Wafer

Definitions

  • This disclosure relates to optical metrology, e.g., to detect the thickness of a layer on a substrate using a machine learning approach.
  • An integrated circuit is typically formed on a substrate by the sequential deposition of conductive, semiconductive, or insulative layers on a silicon wafer. Planarization of a substrate surface may be required for the removal of a filler layer or to improve planarity for photolithography during fabrication of the integrated circuit.
  • CMP Chemical mechanical polishing
  • Various optical metrology systems e.g., spectrographic or ellipsometric, can be used to measure the thickness of the substrate layer pre-polishing and post-polishing, e.g., at an in-line or stand-alone metrology station.
  • GPU Graphical Processing Unit
  • TPU Tensor Processing Unit
  • a method of training a neural network for use in a substrate thickness measurement system includes obtaining ground truth thickness measurements of a top layer of a calibration substrate at a plurality of locations, each location at a defined position for a die being fabricated on the substrate. A plurality of color images of the calibration substrate are obtained, each color image corresponding to a region for a die being fabricated on the substrate. A neural network is trained to convert color images of die regions from an in-line substrate imager to thickness measurements for the top layer in the die region. The training is performed using training data that includes the plurality of color images and ground truth thickness measurements with each respective color image paired with a ground truth thickness measurement for the die region associated with the respective color image.
  • a method of controlling polishing includes obtaining a first color image of a first substrate at an in-line monitoring station of a polishing system, dividing the first color image into a plurality of second color images using a die mask such that each second color image corresponds to a region for a die being fabricated on the first substrate, generating thickness measurements for one or more locations, and determining a polishing parameter for the first substrate or a subsequent second substrate based on the thickness measurements. Each respective location of the one more locations corresponding to a respective region for a die being fabricated on the first substrate.
  • a second color image corresponding to the region is processed through a neural network that was trained using training data that included a plurality of third color images of dies of a calibration substrate and ground truth thickness measurements of the calibration substrate, with each respective third color image paired with a ground truth thickness measurement for a die region associated with the respective third color image.
  • Thicknesses for multiple dies on a substrate may be measured quickly.
  • an in-line metrology system may determine thicknesses for a substrate based on a color images of the substrate without impacting throughput.
  • the estimated thickness may be directly used in a multivariable run-to-run control scheme.
  • the described approach may be used to train a model to generate thickness measurements with an error of less than 5% of the actual film thickness.
  • thickness measurements can be extracted from a color image having three color channels, a hyperspectral camera may be added to the substrate imager system to provide higher dimensional feature input to the model. This may facilitate the training of a more complicated model to understand more physical properties of the film stack.
  • the deep learning in the metrology system may have high inference speed and still be able to achieve a high-resolution measurement of the thickness profile on the substrate. It enables the metrology system to be a fast and low-cost pre- and post-metrology measurement tool for memory applications with great thickness accuracy.
  • FIG. 1 illustrates a view of an example of an in-line optical measurement system.
  • FIG. 2 A illustrates an example of an exemplary image of a substrate used for model training.
  • FIG. 2 B is a schematic illustration of a computer data storage system.
  • FIG. 3 illustrates a neural network used as a part of the controller for the polishing apparatus.
  • FIG. 4 illustrates a flow chart for a method of detecting a thickness of a layer on a substrate using a deep learning approach.
  • Thin-film thickness measurements from dry metrology systems are used in CMP processing because of the variability in the polishing rate that occurs in CMP processes.
  • Such dry metrology measurement techniques often use a spectrographic or ellipsometric approach in which variables in an optical model of a film stack are fit to the collected measurements.
  • Such measurement techniques typically require precise alignment of a sensor to a measurement spot of the substrate to ensure that the model is applicable to the collected measurements. Therefore, measuring a large number of points on the substrate can be time-consuming, and collecting a high-resolution thickness profile is not feasible.
  • the usage of machine learning can enable measurement of a thickness of a film on a substrate with reduced time.
  • film thicknesses of dies can be measured by applying an input image to the neural network.
  • This system can be used as a high throughput and economical solution, e.g., for low-cost memory applications.
  • this technique can be used to classify levels of residue on the substrate using the image segmentation.
  • a polishing apparatus 100 includes one or more carrier heads 126 , each of which is configured to carry a substrate 10 , one or more polishing stations 106 , and a transfer station to load substrate to and unload substrates from a carrier head.
  • Each polishing station 106 includes a polishing pad 130 supported on a platen 120 .
  • the polishing pad 130 can be a two-layer polishing pad with an outer polishing layer and a softer backing layer.
  • the carrier heads 126 can be suspended from a support 128 , and movable between the polishing stations.
  • the support 128 is an overhead track and each carrier head 126 is coupled to a carriage 108 that is mounted to the track so that each carriage 108 can be selectively moved between the polishing stations 124 and the transfer station.
  • the support 128 is a rotatable carousel, and rotation of the carousel moves the carrier heads 126 simultaneously along a circular path.
  • Each polishing station 106 of the polishing apparatus 100 can include a port, e.g., at the end of an arm 134 , to dispense polishing liquid 136 , such as abrasive slurry, onto the polishing pad 130 .
  • Each polishing station 106 of the polishing apparatus 100 can also include pad conditioning apparatus to abrade the polishing pad 130 to maintain the polishing pad 130 in a consistent abrasive state.
  • Each carrier head 126 is operable to hold a substrate 10 against the polishing pad 130 .
  • Each carrier head 126 can have independent control of the polishing parameters, for example, pressure associated with each respective substrate.
  • each carrier head 126 can include a retaining ring 142 to retain the substrate 10 below a flexible membrane 144 .
  • Each carrier head 126 can also include a plurality of independently controllable pressurizable chambers defined by the membrane, e.g., three chambers 146 a - 146 c , which can apply independently controllable pressures to associated zones on the flexible membrane 144 and thus on the substrate 10 . Although only three chambers are illustrated in FIG. 1 for ease of illustration, there could be one or two chambers, or four or more chambers, e.g., five chambers.
  • Each carrier head 126 is suspended from the support 128 and is connected by a drive shaft 154 to a carrier head rotation motor 156 so that the carrier head can rotate about an axis 127 .
  • each carrier head 126 can oscillate laterally, e.g., by driving the carriage 108 on a track, or by the rotational oscillation of the carousel itself.
  • the platen is rotated about its central axis, and each carrier head is rotated about its central axis 127 and translated laterally across the top surface of the polishing pad.
  • a controller 190 such as a programmable computer, is connected to each motor to independently control the rotation rate of the platen 120 and the carrier heads 126 .
  • the controller 190 can include a central processing unit (CPU) 192 , a memory 194 , and support circuits 196 , e.g., input/output circuitry, power supplies, clock circuits, cache, and the like.
  • the memory is connected to the CPU 192 .
  • the memory is a non-transitory computable readable medium, and can be one or more readily available memory such as random access memory (RAM), read-only memory (ROM), floppy disk, hard disk, or another form of digital storage.
  • RAM random access memory
  • ROM read-only memory
  • floppy disk hard disk, or another form of digital storage.
  • the controller 190 could be a distributed system, e.g., including multiple independently operating processors and memories.
  • the polishing apparatus 100 also includes an in-line (also referred to as in-sequence) optical metrology system 160 .
  • a color imaging system of the in-line optical metrology system 160 is positioned within the polishing apparatus 100 , but does not perform measurements during the polishing operation; rather measurements are collected between polishing operations, e.g., while the substrate is being moved from one polishing station to another, or pre- or post-polishing, e.g., while the substrate is being moved from the transfer station to a polishing station or vice versa.
  • the in-line optical metrology system 160 can be positioned in a fab interface unit or a module accessible from the fab interface unit to measure a substrate after the substrate is extracted from a cassette but before the substrate is moved to the polishing unit, or after the substrate has been cleaned but before the substrate is returned to the cassette.
  • the in-line optical metrology system 160 includes a sensor assembly 161 that provides the color imaging of the substrate 10 .
  • the sensor assembly 161 can include a light source 162 , a light detector 164 , and circuitry 166 for sending and receiving signals between the controller 190 and the light source 162 and light detector 164 .
  • the light source 162 can be operable to emit white light.
  • the white light emitted includes light having wavelengths of 200-800 nanometers.
  • a suitable light source is an array of white-light light-emitting diodes (LEDs), or a xenon lamp or a xenon mercury lamp.
  • the light source 162 is oriented to direct light 168 onto the exposed surface of the substrate 10 at a non-zero angle of incidence ⁇ .
  • the angle of incidence ⁇ can be, for example, about 30° to 75°, e.g., 50°.
  • the light source can illuminate a substantially linear elongated region that spans the width of the substrate 10 .
  • the light source 162 can include optics, e.g., a beam expander, to spread the light from the light source into an elongated region.
  • the light source 162 can include a linear array of light sources.
  • the light source 162 itself, and the region illuminated on the substrate, can elongate and have a longitudinal axis parallel to the surface of the substrate.
  • a diffuser 170 can be placed in the path of the light 168 , or the light source 162 can include a diffuser, to diffuse the light before it reaches the substrate 10 .
  • the detector 164 is a color camera that is sensitive to light from the light source 162 .
  • the camera includes an array of detector elements.
  • the camera can include a CCD array.
  • the array is a single row of detector elements.
  • the camera can be a line-scan camera.
  • the row of detector elements can extend parallel to the longitudinal axis of the elongated region illuminated by the light source 162 .
  • the light source 162 includes a row of light-emitting elements
  • the row of detector elements can extend along a first axis parallel to the longitudinal axis of the light source 162 .
  • a row of detector elements can include 1024 or more elements.
  • the camera 164 is configured with appropriate focusing optics 172 to project a field of view of the substrate onto the array of detector elements.
  • the field of view can be long enough to view the entire width of the substrate 10 , e.g., 150 to 300 mm long.
  • the camera 164 including associated optics 172 , can be configured such that individual pixels correspond to a region having a length equal to or less than about 0.5 mm. For example, assuming that the field of view is about 200 mm long and the detector 164 includes 1024 elements, then an image generated by the line-scan camera can have pixels with a length of about 0.5 mm. To determine the length resolution of the image, the length of the field of view (FOV) can be divided by the number of pixels onto which the FOV is imaged to arrive at a length resolution.
  • FOV field of view
  • the camera 164 can be also be configured such that the pixel width is comparable to the pixel length.
  • an advantage of a line-scan camera is its very fast frame rate.
  • the frame rate can be at least 5 kHz.
  • the frame rate can be set at a frequency such that as the imaged area scans across the substrate 10 , the pixel width is comparable to the pixel length, e.g., equal to or less than about 0.3 mm.
  • the light source 162 and the light detector 164 can be supported on a stage 180 .
  • the light detector 164 is a line-scan camera
  • the light source 162 and camera 164 are movable relative to the substrate 10 such that the imaged area can scan across the length of the substrate.
  • the relative motion can be in a direction parallel to the surface of the substrate 10 and perpendicular to the row of detector elements of the line-scan camera 164 .
  • the stage 182 is stationary, and the support for the substrate moves.
  • the carrier head 126 can move, e.g., either by motion of the carriage 108 or by rotational oscillation of the carousel, or the robot arm holding the substrate in a factory interface unit can move the substrate 10 past the line-scan camera 182 .
  • the stage 180 is movable while the carrier head or robot arm remains stationary for the image acquisition.
  • the stage 180 can be movable along a rail 184 by a linear actuator 182 . In either case, this permits the light source 162 and camera 164 to stay in a fixed position relative to each other as the area being scanned moves across the substrate 10 .
  • a possible advantage of having a line-scan camera and light source that move together across the substrate is that, e.g., as compared to a conventional 2D camera, the relative angle between the light source and the camera remains constant for different positions across the wafer. Consequently, artifacts caused by variation in the viewing angle can be reduced or eliminated.
  • a line scan camera can eliminate perspective distortion, whereas a conventional 2D camera exhibits inherent perspective distortion, which then needs to be corrected by an image transformation.
  • the sensor assembly 161 can include a mechanism to adjust vertical distance between the substrate 10 and the light source 162 and detector 164 .
  • the sensor assembly 161 can include an actuator to adjust the vertical position of the stage 180 .
  • a polarizing filter 174 can be positioned in the path of the light, e.g., between the substrate 10 and the detector 164 .
  • the polarizing filter 174 can be a circular polarizer (CPL).
  • CPL circular polarizer
  • a typical CPL is a combination of a linear polarizer and quarter-wave plate. Proper orientation of the polarizing axis of the polarizing filter 174 can reduce haze in the image and sharpen or enhance desirable visual features.
  • the color of light detected at detector 164 depends on, e.g., the composition of the substrate surface, substrate surface smoothness, and/or the amount of interference between light reflected from different interfaces of one or more layers (e.g., dielectric layers) on the substrate.
  • the light source 162 and light detector 164 can be connected to a computing device, e.g., the controller 190 , operable to control their operation and receive their signals.
  • the computing device that performs the various functions to convert the color image to a thickness measurement can be considered part of the metrology system 160 .
  • an example of an image 202 of a substrate 10 collected with in-line optical metrology system 160 is shown.
  • the in-line optical metrology system 160 produces a high-resolution color image 202 , e.g., an image of at least 720 ⁇ 1080 pixels with at least three color channels, e.g., RGB channels), e.g., an image of at least 2048 ⁇ 2048 pixels.
  • the color at any particular pixel depends on the thickness of one or more layers, including the top layer, in the area of the substrate corresponding to the pixel.
  • the image 202 is divided into one or more regions 208 , each region corresponding to a die 206 being fabricated on the substrate.
  • the portion of the image that provides a region 208 can be a predetermined area in the image, or the portion that provides a region 208 can be determined automatically by an algorithm based on the image.
  • the controller 190 can store a die mask that identifies a location and area in the image for each region 208 .
  • the area can be defined by upper right and lower left coordinates in the image.
  • the mask could be a data file that includes a pair of an upper right and a lower left coordinate for each rectangular region. In other cases, where regions are non-rectangular, more complex functions can be used.
  • the orientation and position of the substrate can be determined, and the die mask can be aligned with respect to the image.
  • the substrate orientation can be determined by a notch finder, or by image processing of the color image 202 , e.g., to determine the angle of scribe lines in the image.
  • the substrate position can also be determined by image processing of the color image 202 , e.g., by detecting the circular substrate edge and then determining a center of the circle.
  • an image processing algorithm can analyze the image 202 and detects scribe lines. The image 202 can then be divided into regions between the identified scribe lines.
  • each color image 204 corresponds to a die 206 being fabricated on the substrate.
  • the color images collected can be exported as PNG images, although many other formats, e.g., JPEG, etc., are possible.
  • a color images 204 can be fed to an image processing algorithm to generate a thickness measurement for the die shown in the color image 204 .
  • the image is used as input data to an image processing algorithm that has been trained, e.g., by a supervised deep learning approach, to estimate a layer thickness based on a color image.
  • the supervised deep learning-based algorithm establishes a model between color images and thickness measurements.
  • the image processing algorithm can include a neural network as the deep learning-based algorithm.
  • the intensity value for each color channel of each pixel of the color image 204 is entered into the image processing algorithm, e.g., into the input neurons of the neural network. Based on this input data, a layer thickness measurement are calculated for the color image.
  • input of the color image 204 to the image processing algorithm result in output of an estimated thickness.
  • This system can be used as high-throughput and economical solution, e.g., for the low-cost memory applications. Aside from the thickness inferences, this technique can be used to classify levels of residue on the substrate using the image segmentation.
  • calibration images of dies from one or more calibration substrates can be obtained as discussed above. That is, each calibration substrate can be scanned by the line-scan camera of the in-line optical metrology system 160 to generate an initial calibration image, and the initial calibration image can be divided into a plurality of color images of the individual regions on the calibration substrate.
  • ground truth thickness measurements are collected at multiple locations on the calibration substrate using a high-accuracy metrology system, e.g., an in-line or stand-alone metrology system.
  • the high-accuracy metrology system can be a dry optical metrology system.
  • the ground truth measurement can come from offline reflectometry, ellipsometry, scatterometry or more advanced TEM measurements, although other techniques may be suitable. Such systems are available from Nova Measuring Instruments Inc. or Nanometrics, Inc. Each location corresponds to one of the dies being fabricated, i.e., to one of the individual regions.
  • a color calibration image 212 is collected with the in-line sensor of the optical metrology system 160 .
  • Each color calibration image is associated with the ground truth thickness measurement 214 for the corresponding die on the calibration substrate.
  • the images 212 and associated ground truth thickness measurements 214 can be stored in a database 220 .
  • the data can be stored as records 210 with each record including a calibration image 212 and a ground truth thickness measurement 214 .
  • the deep learning-based algorithm e.g., the neural network
  • the deep learning-based algorithm is then train using the combined data set 218 .
  • the thickness measurements corresponding to the center of die measured from dry metrology tool is used as a label for the input image while training the model.
  • the model may be trained on about 50,000 images collected from five dies on a substrate that have a wide range of post thicknesses.
  • FIG. 3 illustrates a neural network 320 used as a part of the controller 190 for the polishing apparatus 100 .
  • the neural network 320 can be a deep neural network developed for regression analysis of RGB intensity values of the input images from the calibration substrate and the ground truth thickness measurements to generate a model to predict the layer thickness of a region of a substrate based on a color image of that region.
  • the neutral network 320 includes a plurality of input nodes 322 .
  • the neural network 320 can include an input node for each color channel associated with each pixel of the input color image, a plurality of hidden nodes 324 (also called “intermediate nodes” below), and an output node 326 that will generate the layer thickness measurement value.
  • each hidden node 324 can be coupled to each input node 322
  • the output node 326 can be coupled to each hidden node 320 .
  • the neural network for image processing is likely to have many layers of hidden nodes 324 .
  • a hidden node 324 outputs a value that a non-linear function of a weighted sum of the values from the input nodes 322 or prior layers of hidden nodes to which the hidden node 324 is connected.
  • the output of a hidden node 324 in the first layer can be expressed as: tan h (0.5* a k1 ( I 1 )+ a k2 ( I 2 )+ . . . + a kM ( I M )+ b k )
  • tan h is the hyperbolic tangent
  • a kx is a weight for the connection between the k th intermediate node and the x th input node (out of M input nodes)
  • I M is the value at the M th input node.
  • other non-linear functions can be used instead of tan h, such as a rectified linear unit (ReLU) function and its variants.
  • ReLU rectified linear unit
  • H k tan h (0.5* a k1 ( I 1 )+ a k2 ( I 2 )+ . . . + a kL ( I L )+ b k )
  • H k tan h (0.5* a k1 ( V 1 ⁇ S )+ a k2 ( V 2 ⁇ S )+ . . . + a kL ( V L ⁇ S )+ b k )
  • V is a value (v 1 , v 2 , . . . , v L ) of weights, with V x being the weight for the x th intensity value out of L intensity values from the color image).
  • CV e.g., thickness
  • neural network 320 may optionally include one or more other input nodes, e.g., node 322 a , to receive other data.
  • This other data could be from a prior measurement of the substrate by the in-situ monitoring system, e.g., pixel intensity values collected from earlier in the processing of the substrate, from a measurement of a prior substrate, e.g., pixel intensity values collected during processing of another substrate, from another sensor in the polishing system, e.g., a measurement of a temperature of the pad or substrate by a temperature sensor, from a polishing recipe stored by the controller that is used to control the polishing system, e.g., a polishing parameter such as carrier head pressure or platen rotation rate use for polishing the substrate, from a variable tracked by the controller, e.g., a number of substrates since the pad was changed, or from a sensor that is not part of the polishing system, e.g., a measurement of a thickness of underlying films by a metrology station.
  • the thickness measurement generated at the output node 326 is fed to a process control module 330 .
  • the process control module can adjust, based on the thickness measurements of one or more regions, the process parameters, e.g., carrier head pressure, platen rotation rate, etc. The adjustment can be performed for a polishing process to be performed on the substrate or a subsequent substrate.
  • the neutral network 320 Before being used for, e.g., substrate measurements, the neutral network 320 needs to be configured.
  • the controller 190 can receive a plurality of calibration images.
  • Each calibration image has multiple intensity values, e.g., an intensity value for each color channel, for each pixel of the calibration image.
  • the controller also receives a characterizing value, e.g., thickness, for each calibration image.
  • the color calibration image can be measured at a particular die being fabricated on one or more calibration or test substrates.
  • ground truth measurements of the thickness at the particular die locations can be performed with dry metrology equipment, e.g., a contact profilometer or ellipsometer. A ground truth thickness measurement can thus be associated with the color image from the same die location on a substrate.
  • the plurality of color calibration images can be generated from, for example, five to ten calibration substrates, by dividing images of the calibration substrates as discussed above.
  • the neural network 320 is trained using the color image and the characteristic value for each die fabricated on a calibration substrate.
  • V corresponds to one of the color images and is thus associated with a characteristic value.
  • the neural network 320 While the neural network 320 is operating in a training mode, such as a backpropagation mode, the values (v 1 , v 2 , . . . , v L ) are fed to the respective input nodes N 1 , N 2 . . . N L while the characteristic value CV is fed to the output node 326 . This can be repeated for each row. This process sets the values for a k1 , etc., in Equations 1 or 2 above.
  • the neural network 320 While the neural network 320 is used in an inference mode, these values (S 1 , S 2 , . . . , S L ) are fed as inputs to the respective input nodes N 1 , N 2 , . . . N L . As a result, the neural network 320 generates a characteristic value, e.g., a layer thickness, at the output node 326 .
  • a characteristic value e.g., a layer thickness
  • the architecture of the neural network 320 can vary in depth and width.
  • the neural network 320 is shown with a single column of intermediate nodes 324 , it could include multiple columns.
  • the number of intermediate nodes 324 can be equal to or greater than the number of input nodes 322 .
  • the controller 190 can associate the various color images with different dies (see FIG. 2 ) on the substrate.
  • the output of each neural network 320 can be classified as belonging to one of the dies based on the position of the sensor on the substrate at the time the image is collected. This permits the controller 190 to generate a separate sequence of measurement values for each die.
  • the controller 190 can be configured to have a neural network model structure with composed of multiple different types of building blocks.
  • the neural network be a residual neural network that includes a res-block feature in its architecture.
  • a residual neural network may utilizing skip connections, or shortcuts, to jump over some layers.
  • a residual neural network can be implemented. e.g., with a ResNet model.
  • a non-residual network may be described as a plain network.
  • a neural network may be trained to take the underlying layer thickness from the stack into consideration during calculations, which can improve errors due to underlying variation in thickness measurements.
  • the effect of the underlying thickness variation in the film stack can be alleviated by feeding the intensity values of the color images of the thicknesses of the underlying layers as extra inputs to the model to improve the performance of the model.
  • the reliability of the computed thickness measurements can be evaluated by comparing them to the measured values and then determining a difference between the computed values and the original measured values.
  • This deep learning model may then be used for predicting the thickness in the inference mode, immediately after a new test substrate is scanned. This new approach enhances the throughput of the entire system and enables the thickness measurement to be performed on all substrates in the lot.
  • a method of an image processing algorithm generated by machine learning techniques for use in a substrate thickness measurement system can receive RGB images collected from an integrated line-scan camera inspection system and enables film thickness estimates with much faster speed.
  • the inference time for about 2000 measurement points is about a few seconds, as compared to 2 hours with dry metrology.
  • the method includes the controller assembling individual image lines from the light detector 164 into a two-dimensional color image ( 500 ).
  • the controller can apply an offset and/or a gain adjustment to the intensity values of the image in each color channel ( 510 ).
  • Each color channel can have a different offset and/or gain.
  • the image can be normalized ( 515 ). For example, the difference between the measured image and a standard predefined image can be calculated.
  • the controller can store a background image for each of the red, green, and blue color channels, and the background image can be subtracted from the measured image for each color channel.
  • the measured image can be divided by the standard predefined image.
  • the image can be filtered to remove low-frequency spatial variations ( 530 ).
  • a luminance channel is used to generate the filter, which is then applied to the red, green, and blue images.
  • the image is transformed, e.g., scaled and/or rotated and/or translated, into a standard image coordinate frame ( 540 ).
  • the image can be translated so that the die center is at the center point of the image and/or the image can be scaled so that the edge of the substrate is at the edge of the image, and/or the image can be rotated so that there is a 0° angle between the x-axis of the image and the radial segment connecting the substrate center and the substrate orientation feature.
  • One or more regions on the substrate are selected and an image is generated for each selected region ( 550 ). This can be performed using the techniques described above, e.g., the regions can be predetermined areas, or the portion that provides a region 208 can be determined automatically by an algorithm
  • the intensity values provided by each color channel for each pixel of the image is considered used as input to the supervised deep learning-trained image processing algorithm.
  • the image processing algorithm outputs a layer thickness measurement for the particular region ( 560 ).
  • data can be used to control one or more operation parameters of the CMP apparatus.
  • Operational parameters include, for example, platen rotational velocity, substrate rotational velocity, the polishing path of the substrate, the substrate speed across the plate, the pressure exerted on the substrate, slurry composition, slurry flow rate, and temperature at the substrate surface. Operational parameters can be controlled real-time and can be automatically adjusted without the need for further human intervention.
  • the term substrate can include, for example, a product substrate (e.g., which includes multiple memory or processor dies), a test substrate, a bare substrate, and a gating substrate.
  • the substrate can be at various stages of integrated circuit fabrication, e.g., the substrate can be a bare wafer, or it can include one or more deposited and/or patterned layers.
  • the term substrate can include circular disks and rectangular sheets.
  • the color image processing technique described above can be particularly useful in the context of 3D vertical NAND (VNAND) flash memory.
  • the layer stack used in fabrication of VNAND is so complicated that current metrology methods (e.g., Nova spectrum analysis) may be unable to perform with sufficient reliability in detecting regions of improper thickness.
  • the color image processing technique can have superior reliability in this application.
  • Embodiments of the invention and all of the functional operations described in this specification can be implemented in digital electronic circuitry, or in computer software, firmware, or hardware, including the structural means disclosed in this specification and structural equivalents thereof, or in combinations of them.
  • Embodiments of the invention can be implemented as one or more computer program products, i.e., one or more computer programs tangibly embodied in a non-transitory machine readable storage media, for execution by, or to control the operation of, data processing apparatus, e.g., a programmable processor, a computer, or multiple processors or computers.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • General Physics & Mathematics (AREA)
  • Geometry (AREA)
  • Quality & Reliability (AREA)
  • Mechanical Engineering (AREA)
  • Software Systems (AREA)
  • Evolutionary Computation (AREA)
  • Data Mining & Analysis (AREA)
  • Medical Informatics (AREA)
  • Artificial Intelligence (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Mathematical Physics (AREA)
  • Length Measuring Devices By Optical Means (AREA)
  • Mechanical Treatment Of Semiconductor (AREA)
  • Testing Or Measuring Of Semiconductors Or The Like (AREA)
  • Image Analysis (AREA)

Abstract

A neural network is trained for use in a substrate thickness measurement system by obtaining ground truth thickness measurements of a top layer of a calibration substrate at a plurality of locations, each location at a defined position for a die being fabricated on the substrate. A plurality of color images of the calibration substrate are obtained, each color image corresponding to a region for a die being fabricated on the substrate. A neural network is trained to convert color images of die regions from an in-line substrate imager to thickness measurements for the top layer in the die region. The training is performed using training data that includes the plurality of color images and ground truth thickness measurements with each respective color image paired with a ground truth thickness measurement for the die region associated with the respective color image.

Description

CROSS-REFERENCE TO RELATED APPLICATIONS
This application claims priority to U.S. Application Ser. No. 63/045,782, filed on Jun. 29, 2020, the disclosure of which is incorporated by reference.
TECHNICAL FIELD
This disclosure relates to optical metrology, e.g., to detect the thickness of a layer on a substrate using a machine learning approach.
BACKGROUND
An integrated circuit is typically formed on a substrate by the sequential deposition of conductive, semiconductive, or insulative layers on a silicon wafer. Planarization of a substrate surface may be required for the removal of a filler layer or to improve planarity for photolithography during fabrication of the integrated circuit.
Chemical mechanical polishing (CMP) is one accepted method of planarization. This planarization method typically requires that the substrate be mounted on a carrier or polishing head. The exposed surface of the substrate is typically placed against a rotating polishing pad. The carrier head provides a controllable load on the substrate to push it against the polishing pad. An abrasive polishing slurry is typically supplied to the surface of the polishing pad.
Various optical metrology systems, e.g., spectrographic or ellipsometric, can be used to measure the thickness of the substrate layer pre-polishing and post-polishing, e.g., at an in-line or stand-alone metrology station.
As a parallel issue, advancements in hardware resources such as Graphical Processing Units (GPU) and Tensor Processing Units (TPU) have resulted in a vast improvement in the deep learning algorithms and their applications. One of the evolving fields of deep learning is computer vision and image recognition. Such computer vision algorithms are mostly designed for image classification or segmentation.
SUMMARY
In one aspect, a method of training a neural network for use in a substrate thickness measurement system includes obtaining ground truth thickness measurements of a top layer of a calibration substrate at a plurality of locations, each location at a defined position for a die being fabricated on the substrate. A plurality of color images of the calibration substrate are obtained, each color image corresponding to a region for a die being fabricated on the substrate. A neural network is trained to convert color images of die regions from an in-line substrate imager to thickness measurements for the top layer in the die region. The training is performed using training data that includes the plurality of color images and ground truth thickness measurements with each respective color image paired with a ground truth thickness measurement for the die region associated with the respective color image.
In another aspect, a method of controlling polishing includes obtaining a first color image of a first substrate at an in-line monitoring station of a polishing system, dividing the first color image into a plurality of second color images using a die mask such that each second color image corresponds to a region for a die being fabricated on the first substrate, generating thickness measurements for one or more locations, and determining a polishing parameter for the first substrate or a subsequent second substrate based on the thickness measurements. Each respective location of the one more locations corresponding to a respective region for a die being fabricated on the first substrate. To generate the thickness measurements for a region, a second color image corresponding to the region is processed through a neural network that was trained using training data that included a plurality of third color images of dies of a calibration substrate and ground truth thickness measurements of the calibration substrate, with each respective third color image paired with a ground truth thickness measurement for a die region associated with the respective third color image.
Implementations can include one or more of the following potential advantages. Thicknesses for multiple dies on a substrate may be measured quickly. For example, an in-line metrology system may determine thicknesses for a substrate based on a color images of the substrate without impacting throughput. The estimated thickness may be directly used in a multivariable run-to-run control scheme.
The described approach may be used to train a model to generate thickness measurements with an error of less than 5% of the actual film thickness. Although thickness measurements can be extracted from a color image having three color channels, a hyperspectral camera may be added to the substrate imager system to provide higher dimensional feature input to the model. This may facilitate the training of a more complicated model to understand more physical properties of the film stack.
The deep learning in the metrology system may have high inference speed and still be able to achieve a high-resolution measurement of the thickness profile on the substrate. It enables the metrology system to be a fast and low-cost pre- and post-metrology measurement tool for memory applications with great thickness accuracy.
The details of one or more implementations are set forth in the accompanying drawings and the description below. Other aspects, features and advantages will be apparent from the description and drawings, and from the claims.
DESCRIPTION OF DRAWINGS
FIG. 1 illustrates a view of an example of an in-line optical measurement system.
FIG. 2A illustrates an example of an exemplary image of a substrate used for model training.
FIG. 2B is a schematic illustration of a computer data storage system.
FIG. 3 illustrates a neural network used as a part of the controller for the polishing apparatus.
FIG. 4 illustrates a flow chart for a method of detecting a thickness of a layer on a substrate using a deep learning approach.
Like reference symbols in the various drawings indicate like elements.
DETAILED DESCRIPTION
Thin-film thickness measurements from dry metrology systems are used in CMP processing because of the variability in the polishing rate that occurs in CMP processes. Such dry metrology measurement techniques often use a spectrographic or ellipsometric approach in which variables in an optical model of a film stack are fit to the collected measurements. Such measurement techniques typically require precise alignment of a sensor to a measurement spot of the substrate to ensure that the model is applicable to the collected measurements. Therefore, measuring a large number of points on the substrate can be time-consuming, and collecting a high-resolution thickness profile is not feasible.
However, the usage of machine learning can enable measurement of a thickness of a film on a substrate with reduced time. By training a deep neural network using color images of dies from a substrate and associated thickness measurements by other reliable metrology systems, film thicknesses of dies can be measured by applying an input image to the neural network. This system can be used as a high throughput and economical solution, e.g., for low-cost memory applications. Aside from the thickness inferences, this technique can be used to classify levels of residue on the substrate using the image segmentation.
Referring to FIG. 1 , a polishing apparatus 100 includes one or more carrier heads 126, each of which is configured to carry a substrate 10, one or more polishing stations 106, and a transfer station to load substrate to and unload substrates from a carrier head. Each polishing station 106 includes a polishing pad 130 supported on a platen 120. The polishing pad 130 can be a two-layer polishing pad with an outer polishing layer and a softer backing layer.
The carrier heads 126 can be suspended from a support 128, and movable between the polishing stations. In some implementations, the support 128 is an overhead track and each carrier head 126 is coupled to a carriage 108 that is mounted to the track so that each carriage 108 can be selectively moved between the polishing stations 124 and the transfer station. Alternatively, in some implementations, the support 128 is a rotatable carousel, and rotation of the carousel moves the carrier heads 126 simultaneously along a circular path.
Each polishing station 106 of the polishing apparatus 100 can include a port, e.g., at the end of an arm 134, to dispense polishing liquid 136, such as abrasive slurry, onto the polishing pad 130. Each polishing station 106 of the polishing apparatus 100 can also include pad conditioning apparatus to abrade the polishing pad 130 to maintain the polishing pad 130 in a consistent abrasive state.
Each carrier head 126 is operable to hold a substrate 10 against the polishing pad 130. Each carrier head 126 can have independent control of the polishing parameters, for example, pressure associated with each respective substrate. In particular, each carrier head 126 can include a retaining ring 142 to retain the substrate 10 below a flexible membrane 144. Each carrier head 126 can also include a plurality of independently controllable pressurizable chambers defined by the membrane, e.g., three chambers 146 a-146 c, which can apply independently controllable pressures to associated zones on the flexible membrane 144 and thus on the substrate 10. Although only three chambers are illustrated in FIG. 1 for ease of illustration, there could be one or two chambers, or four or more chambers, e.g., five chambers.
Each carrier head 126 is suspended from the support 128 and is connected by a drive shaft 154 to a carrier head rotation motor 156 so that the carrier head can rotate about an axis 127. Optionally each carrier head 126 can oscillate laterally, e.g., by driving the carriage 108 on a track, or by the rotational oscillation of the carousel itself. In operation, the platen is rotated about its central axis, and each carrier head is rotated about its central axis 127 and translated laterally across the top surface of the polishing pad.
A controller 190, such as a programmable computer, is connected to each motor to independently control the rotation rate of the platen 120 and the carrier heads 126. The controller 190 can include a central processing unit (CPU) 192, a memory 194, and support circuits 196, e.g., input/output circuitry, power supplies, clock circuits, cache, and the like. The memory is connected to the CPU 192. The memory is a non-transitory computable readable medium, and can be one or more readily available memory such as random access memory (RAM), read-only memory (ROM), floppy disk, hard disk, or another form of digital storage. In addition, although illustrated as a single computer, the controller 190 could be a distributed system, e.g., including multiple independently operating processors and memories.
The polishing apparatus 100 also includes an in-line (also referred to as in-sequence) optical metrology system 160. A color imaging system of the in-line optical metrology system 160 is positioned within the polishing apparatus 100, but does not perform measurements during the polishing operation; rather measurements are collected between polishing operations, e.g., while the substrate is being moved from one polishing station to another, or pre- or post-polishing, e.g., while the substrate is being moved from the transfer station to a polishing station or vice versa. In addition, the in-line optical metrology system 160 can be positioned in a fab interface unit or a module accessible from the fab interface unit to measure a substrate after the substrate is extracted from a cassette but before the substrate is moved to the polishing unit, or after the substrate has been cleaned but before the substrate is returned to the cassette.
The in-line optical metrology system 160 includes a sensor assembly 161 that provides the color imaging of the substrate 10. The sensor assembly 161 can include a light source 162, a light detector 164, and circuitry 166 for sending and receiving signals between the controller 190 and the light source 162 and light detector 164.
The light source 162 can be operable to emit white light. In one implementation, the white light emitted includes light having wavelengths of 200-800 nanometers. A suitable light source is an array of white-light light-emitting diodes (LEDs), or a xenon lamp or a xenon mercury lamp. The light source 162 is oriented to direct light 168 onto the exposed surface of the substrate 10 at a non-zero angle of incidence α. The angle of incidence α can be, for example, about 30° to 75°, e.g., 50°.
The light source can illuminate a substantially linear elongated region that spans the width of the substrate 10. For example, the light source 162 can include optics, e.g., a beam expander, to spread the light from the light source into an elongated region. Alternatively or in addition, the light source 162 can include a linear array of light sources. The light source 162 itself, and the region illuminated on the substrate, can elongate and have a longitudinal axis parallel to the surface of the substrate.
A diffuser 170 can be placed in the path of the light 168, or the light source 162 can include a diffuser, to diffuse the light before it reaches the substrate 10.
The detector 164 is a color camera that is sensitive to light from the light source 162. The camera includes an array of detector elements. For example, the camera can include a CCD array. In some implementations, the array is a single row of detector elements. For example, the camera can be a line-scan camera. The row of detector elements can extend parallel to the longitudinal axis of the elongated region illuminated by the light source 162. Where the light source 162 includes a row of light-emitting elements, the row of detector elements can extend along a first axis parallel to the longitudinal axis of the light source 162. A row of detector elements can include 1024 or more elements.
The camera 164 is configured with appropriate focusing optics 172 to project a field of view of the substrate onto the array of detector elements. The field of view can be long enough to view the entire width of the substrate 10, e.g., 150 to 300 mm long. The camera 164, including associated optics 172, can be configured such that individual pixels correspond to a region having a length equal to or less than about 0.5 mm. For example, assuming that the field of view is about 200 mm long and the detector 164 includes 1024 elements, then an image generated by the line-scan camera can have pixels with a length of about 0.5 mm. To determine the length resolution of the image, the length of the field of view (FOV) can be divided by the number of pixels onto which the FOV is imaged to arrive at a length resolution.
The camera 164 can be also be configured such that the pixel width is comparable to the pixel length. For example, an advantage of a line-scan camera is its very fast frame rate. The frame rate can be at least 5 kHz. The frame rate can be set at a frequency such that as the imaged area scans across the substrate 10, the pixel width is comparable to the pixel length, e.g., equal to or less than about 0.3 mm.
The light source 162 and the light detector 164 can be supported on a stage 180. Where the light detector 164 is a line-scan camera, the light source 162 and camera 164 are movable relative to the substrate 10 such that the imaged area can scan across the length of the substrate. In particular, the relative motion can be in a direction parallel to the surface of the substrate 10 and perpendicular to the row of detector elements of the line-scan camera 164.
In some implementations, the stage 182 is stationary, and the support for the substrate moves. For example, the carrier head 126 can move, e.g., either by motion of the carriage 108 or by rotational oscillation of the carousel, or the robot arm holding the substrate in a factory interface unit can move the substrate 10 past the line-scan camera 182. In some implementations, the stage 180 is movable while the carrier head or robot arm remains stationary for the image acquisition. For example, the stage 180 can be movable along a rail 184 by a linear actuator 182. In either case, this permits the light source 162 and camera 164 to stay in a fixed position relative to each other as the area being scanned moves across the substrate 10.
A possible advantage of having a line-scan camera and light source that move together across the substrate is that, e.g., as compared to a conventional 2D camera, the relative angle between the light source and the camera remains constant for different positions across the wafer. Consequently, artifacts caused by variation in the viewing angle can be reduced or eliminated. In addition, a line scan camera can eliminate perspective distortion, whereas a conventional 2D camera exhibits inherent perspective distortion, which then needs to be corrected by an image transformation.
The sensor assembly 161 can include a mechanism to adjust vertical distance between the substrate 10 and the light source 162 and detector 164. For example, the sensor assembly 161 can include an actuator to adjust the vertical position of the stage 180.
Optionally a polarizing filter 174 can be positioned in the path of the light, e.g., between the substrate 10 and the detector 164. The polarizing filter 174 can be a circular polarizer (CPL). A typical CPL is a combination of a linear polarizer and quarter-wave plate. Proper orientation of the polarizing axis of the polarizing filter 174 can reduce haze in the image and sharpen or enhance desirable visual features.
Assuming that the outermost layer on the substrate is a semitransparent layer, e.g., a dielectric layer, the color of light detected at detector 164 depends on, e.g., the composition of the substrate surface, substrate surface smoothness, and/or the amount of interference between light reflected from different interfaces of one or more layers (e.g., dielectric layers) on the substrate. As noted above, the light source 162 and light detector 164 can be connected to a computing device, e.g., the controller 190, operable to control their operation and receive their signals. The computing device that performs the various functions to convert the color image to a thickness measurement, can be considered part of the metrology system 160.
Referring to FIG. 2A, an example of an image 202 of a substrate 10 collected with in-line optical metrology system 160 is shown. The in-line optical metrology system 160 produces a high-resolution color image 202, e.g., an image of at least 720×1080 pixels with at least three color channels, e.g., RGB channels), e.g., an image of at least 2048×2048 pixels. The color at any particular pixel depends on the thickness of one or more layers, including the top layer, in the area of the substrate corresponding to the pixel.
The image 202 is divided into one or more regions 208, each region corresponding to a die 206 being fabricated on the substrate. The portion of the image that provides a region 208 can be a predetermined area in the image, or the portion that provides a region 208 can be determined automatically by an algorithm based on the image.
As an example of a predetermined area in the image, the controller 190 can store a die mask that identifies a location and area in the image for each region 208. For example, for rectangular regions, the area can be defined by upper right and lower left coordinates in the image. Thus, the mask could be a data file that includes a pair of an upper right and a lower left coordinate for each rectangular region. In other cases, where regions are non-rectangular, more complex functions can be used.
In some implementations, the orientation and position of the substrate can be determined, and the die mask can be aligned with respect to the image. The substrate orientation can be determined by a notch finder, or by image processing of the color image 202, e.g., to determine the angle of scribe lines in the image. The substrate position can also be determined by image processing of the color image 202, e.g., by detecting the circular substrate edge and then determining a center of the circle.
As an example determining the region 208 automatically, an image processing algorithm can analyze the image 202 and detects scribe lines. The image 202 can then be divided into regions between the identified scribe lines.
By dividing the initial color image, a plurality of color images 204 of individual regions 208 can be collected from the substrate 10. As noted above, each color image 204 corresponds to a die 206 being fabricated on the substrate. The color images collected can be exported as PNG images, although many other formats, e.g., JPEG, etc., are possible.
A color images 204 can be fed to an image processing algorithm to generate a thickness measurement for the die shown in the color image 204. The image is used as input data to an image processing algorithm that has been trained, e.g., by a supervised deep learning approach, to estimate a layer thickness based on a color image. The supervised deep learning-based algorithm establishes a model between color images and thickness measurements. The image processing algorithm can include a neural network as the deep learning-based algorithm.
The intensity value for each color channel of each pixel of the color image 204 is entered into the image processing algorithm, e.g., into the input neurons of the neural network. Based on this input data, a layer thickness measurement are calculated for the color image. Thus, input of the color image 204 to the image processing algorithm result in output of an estimated thickness. This system can be used as high-throughput and economical solution, e.g., for the low-cost memory applications. Aside from the thickness inferences, this technique can be used to classify levels of residue on the substrate using the image segmentation.
In order to train the image processing algorithm, e.g., the neural network, using the supervised deep learning approach, calibration images of dies from one or more calibration substrates can be obtained as discussed above. That is, each calibration substrate can be scanned by the line-scan camera of the in-line optical metrology system 160 to generate an initial calibration image, and the initial calibration image can be divided into a plurality of color images of the individual regions on the calibration substrate.
Either before or after the initial color calibration image is collected, ground truth thickness measurements are collected at multiple locations on the calibration substrate using a high-accuracy metrology system, e.g., an in-line or stand-alone metrology system. The high-accuracy metrology system can be a dry optical metrology system. The ground truth measurement can come from offline reflectometry, ellipsometry, scatterometry or more advanced TEM measurements, although other techniques may be suitable. Such systems are available from Nova Measuring Instruments Inc. or Nanometrics, Inc. Each location corresponds to one of the dies being fabricated, i.e., to one of the individual regions.
For example, referring to FIG. 2B, for each individual region on each calibration substrate, a color calibration image 212 is collected with the in-line sensor of the optical metrology system 160. Each color calibration image is associated with the ground truth thickness measurement 214 for the corresponding die on the calibration substrate. The images 212 and associated ground truth thickness measurements 214 can be stored in a database 220. For example, the data can be stored as records 210 with each record including a calibration image 212 and a ground truth thickness measurement 214.
The deep learning-based algorithm, e.g., the neural network, is then train using the combined data set 218. The thickness measurements corresponding to the center of die measured from dry metrology tool is used as a label for the input image while training the model. For example, the model may be trained on about 50,000 images collected from five dies on a substrate that have a wide range of post thicknesses.
FIG. 3 illustrates a neural network 320 used as a part of the controller 190 for the polishing apparatus 100. The neural network 320 can be a deep neural network developed for regression analysis of RGB intensity values of the input images from the calibration substrate and the ground truth thickness measurements to generate a model to predict the layer thickness of a region of a substrate based on a color image of that region.
The neutral network 320 includes a plurality of input nodes 322. The neural network 320 can include an input node for each color channel associated with each pixel of the input color image, a plurality of hidden nodes 324 (also called “intermediate nodes” below), and an output node 326 that will generate the layer thickness measurement value. In a neural network having a single layer of hidden nodes, each hidden node 324 can be coupled to each input node 322, and the output node 326 can be coupled to each hidden node 320. However, as a practical matter, the neural network for image processing is likely to have many layers of hidden nodes 324.
In general, a hidden node 324 outputs a value that a non-linear function of a weighted sum of the values from the input nodes 322 or prior layers of hidden nodes to which the hidden node 324 is connected.
For example, the output of a hidden node 324 in the first layer, designated node k, can be expressed as:
tan h(0.5*a k1(I 1)+a k2(I 2)+ . . . +a kM(I M)+b k)
where tan h is the hyperbolic tangent, akx is a weight for the connection between the kth intermediate node and the xth input node (out of M input nodes), and IM is the value at the Mth input node. However, other non-linear functions can be used instead of tan h, such as a rectified linear unit (ReLU) function and its variants.
The neural network 320 thus includes an input node 322 for each color channel associated with each pixel of the input color image, e.g., where there are J pixels and K color channels, then L=J*K is number of intensity values in an input color image, and the neural network 320 will include at least input nodes N1, N2 . . . NL.
Thus, where the number of input nodes corresponds to the number of intensity values in the color image, the output Hk of a hidden node 324, designated node k, can be expressed as:
H k=tan h(0.5*a k1(I 1)+a k2(I 2)+ . . . +a kL(I L)+b k)
Assuming that the measured color image S is represented by a column matrix (i1, i2, . . . , iL), the output of an intermediate node 324, designated node k, can be expressed as:
H k=tan h(0.5*a k1(V 1 ·S)+a k2(V 2 ·S)+ . . . +a kL(V L ·S)+b k)
where V is a value (v1, v2, . . . , vL) of weights, with Vx being the weight for the xth intensity value out of L intensity values from the color image).
The output node 326 can generate a characteristic value CV, e.g., thickness, that is a weighted sum of the outputs of the hidden nodes. For example, this can be expressed as
CV=C 1 *H 1 +C 2 *H 2 ++C L *H L
where Ck is the weight for the output of the kth hidden node.
However, neural network 320 may optionally include one or more other input nodes, e.g., node 322 a, to receive other data. This other data could be from a prior measurement of the substrate by the in-situ monitoring system, e.g., pixel intensity values collected from earlier in the processing of the substrate, from a measurement of a prior substrate, e.g., pixel intensity values collected during processing of another substrate, from another sensor in the polishing system, e.g., a measurement of a temperature of the pad or substrate by a temperature sensor, from a polishing recipe stored by the controller that is used to control the polishing system, e.g., a polishing parameter such as carrier head pressure or platen rotation rate use for polishing the substrate, from a variable tracked by the controller, e.g., a number of substrates since the pad was changed, or from a sensor that is not part of the polishing system, e.g., a measurement of a thickness of underlying films by a metrology station. This permits the neural network 320 to take into account other processing or environmental variables in the calculation of the layer thickness measurement value.
The thickness measurement generated at the output node 326 is fed to a process control module 330. The process control module can adjust, based on the thickness measurements of one or more regions, the process parameters, e.g., carrier head pressure, platen rotation rate, etc. The adjustment can be performed for a polishing process to be performed on the substrate or a subsequent substrate.
Before being used for, e.g., substrate measurements, the neutral network 320 needs to be configured.
As part of a configuration procedure, the controller 190 can receive a plurality of calibration images. Each calibration image has multiple intensity values, e.g., an intensity value for each color channel, for each pixel of the calibration image. The controller also receives a characterizing value, e.g., thickness, for each calibration image. For example, the color calibration image can be measured at a particular die being fabricated on one or more calibration or test substrates. In addition, ground truth measurements of the thickness at the particular die locations can be performed with dry metrology equipment, e.g., a contact profilometer or ellipsometer. A ground truth thickness measurement can thus be associated with the color image from the same die location on a substrate. The plurality of color calibration images can be generated from, for example, five to ten calibration substrates, by dividing images of the calibration substrates as discussed above. As of the configuration procedure for the neural network 320, the neural network 320 is trained using the color image and the characteristic value for each die fabricated on a calibration substrate.
V corresponds to one of the color images and is thus associated with a characteristic value. While the neural network 320 is operating in a training mode, such as a backpropagation mode, the values (v1, v2, . . . , vL) are fed to the respective input nodes N1, N2 . . . NL while the characteristic value CV is fed to the output node 326. This can be repeated for each row. This process sets the values for ak1, etc., in Equations 1 or 2 above.
The system is now ready for operation. A color image measured from a substrate using the in-line monitoring system 160. The measured color image can be represented by a column matrix S=(i1, i2, . . . , IL), where ij represents the intensity value at the jth intensity value out of L intensity values, with L=3n when the image includes a total of n pixels and each pixel includes three color channels.
While the neural network 320 is used in an inference mode, these values (S1, S2, . . . , SL) are fed as inputs to the respective input nodes N1, N2, . . . NL. As a result, the neural network 320 generates a characteristic value, e.g., a layer thickness, at the output node 326.
The architecture of the neural network 320 can vary in depth and width. For example, although the neural network 320 is shown with a single column of intermediate nodes 324, it could include multiple columns. The number of intermediate nodes 324 can be equal to or greater than the number of input nodes 322.
As noted above, the controller 190 can associate the various color images with different dies (see FIG. 2 ) on the substrate. The output of each neural network 320 can be classified as belonging to one of the dies based on the position of the sensor on the substrate at the time the image is collected. This permits the controller 190 to generate a separate sequence of measurement values for each die.
In some implementations, the controller 190 can be configured to have a neural network model structure with composed of multiple different types of building blocks. For example, the neural network be a residual neural network that includes a res-block feature in its architecture. A residual neural network may utilizing skip connections, or shortcuts, to jump over some layers. A residual neural network can be implemented. e.g., with a ResNet model. In the context of residual neural networks, a non-residual network may be described as a plain network.
In some implementations, a neural network may be trained to take the underlying layer thickness from the stack into consideration during calculations, which can improve errors due to underlying variation in thickness measurements. The effect of the underlying thickness variation in the film stack can be alleviated by feeding the intensity values of the color images of the thicknesses of the underlying layers as extra inputs to the model to improve the performance of the model.
The reliability of the computed thickness measurements can be evaluated by comparing them to the measured values and then determining a difference between the computed values and the original measured values. This deep learning model may then be used for predicting the thickness in the inference mode, immediately after a new test substrate is scanned. This new approach enhances the throughput of the entire system and enables the thickness measurement to be performed on all substrates in the lot.
Referring to FIG. 4 , a method of an image processing algorithm generated by machine learning techniques for use in a substrate thickness measurement system. Such an image processing algorithm can receive RGB images collected from an integrated line-scan camera inspection system and enables film thickness estimates with much faster speed. The inference time for about 2000 measurement points is about a few seconds, as compared to 2 hours with dry metrology.
The method includes the controller assembling individual image lines from the light detector 164 into a two-dimensional color image (500). The controller can apply an offset and/or a gain adjustment to the intensity values of the image in each color channel (510). Each color channel can have a different offset and/or gain. Optionally, the image can be normalized (515). For example, the difference between the measured image and a standard predefined image can be calculated. For example, the controller can store a background image for each of the red, green, and blue color channels, and the background image can be subtracted from the measured image for each color channel. Alternatively, the measured image can be divided by the standard predefined image. The image can be filtered to remove low-frequency spatial variations (530). In some implementations, a luminance channel is used to generate the filter, which is then applied to the red, green, and blue images.
The image is transformed, e.g., scaled and/or rotated and/or translated, into a standard image coordinate frame (540). For example, the image can be translated so that the die center is at the center point of the image and/or the image can be scaled so that the edge of the substrate is at the edge of the image, and/or the image can be rotated so that there is a 0° angle between the x-axis of the image and the radial segment connecting the substrate center and the substrate orientation feature.
One or more regions on the substrate are selected and an image is generated for each selected region (550). This can be performed using the techniques described above, e.g., the regions can be predetermined areas, or the portion that provides a region 208 can be determined automatically by an algorithm
The intensity values provided by each color channel for each pixel of the image is considered used as input to the supervised deep learning-trained image processing algorithm. The image processing algorithm outputs a layer thickness measurement for the particular region (560).
Various deep model architectures were trained and validated on small die test patterned substrates with a goal of reducing errors in the measurements. The model that took into consideration the characteristics of the underlying layer had a lower error. In addition, preliminary tool-to-tool matching validation was performed by training the model with data collected on one tool and using it for inferences on the data from other tools. Results were comparable to training and inferencing with data from the same tool.
In general, data can be used to control one or more operation parameters of the CMP apparatus. Operational parameters include, for example, platen rotational velocity, substrate rotational velocity, the polishing path of the substrate, the substrate speed across the plate, the pressure exerted on the substrate, slurry composition, slurry flow rate, and temperature at the substrate surface. Operational parameters can be controlled real-time and can be automatically adjusted without the need for further human intervention.
As used in the instant specification, the term substrate can include, for example, a product substrate (e.g., which includes multiple memory or processor dies), a test substrate, a bare substrate, and a gating substrate. The substrate can be at various stages of integrated circuit fabrication, e.g., the substrate can be a bare wafer, or it can include one or more deposited and/or patterned layers. The term substrate can include circular disks and rectangular sheets.
However, the color image processing technique described above can be particularly useful in the context of 3D vertical NAND (VNAND) flash memory. In particular, the layer stack used in fabrication of VNAND is so complicated that current metrology methods (e.g., Nova spectrum analysis) may be unable to perform with sufficient reliability in detecting regions of improper thickness. In contrast, the color image processing technique can have superior reliability in this application.
Embodiments of the invention and all of the functional operations described in this specification can be implemented in digital electronic circuitry, or in computer software, firmware, or hardware, including the structural means disclosed in this specification and structural equivalents thereof, or in combinations of them. Embodiments of the invention can be implemented as one or more computer program products, i.e., one or more computer programs tangibly embodied in a non-transitory machine readable storage media, for execution by, or to control the operation of, data processing apparatus, e.g., a programmable processor, a computer, or multiple processors or computers.
Terms of relative positioning are used to denote positioning of components of the system relative to each other, not necessarily with respect to gravity; it should be understood that the polishing surface and substrate can be held in a vertical orientation or some other orientations.
A number of implementations have been described. Nevertheless, it will be understood that various modifications may be made. For example
    • Rather than a line scan camera, a camera that images the entire substrate could be used. In this case, motion of the camera relative to the substrate is not needed.
    • The camera could cover less than the entire width of the substrate. In this case, the camera would need to undergo motion in two perpendicular directions, e.g., be supported on an X-Y stage, in order to scan the entire substrate.
    • The light source could illuminate the entire substrate. In this case, the light source need not move relative to the substrate.
    • The light detector can be a spectrometer rather than a color camera; the spectra data can then be reduced to the RGB color space.
    • The sensory assembly need not an in-line system positioned between polishing stations or between a polishing station and a transfer station. For example, the sensor assembly could be positioned within the transfer station, positioned in a cassette interface unit, or be a stand-alone system.
    • The uniformity analysis step is optional. For example, the image generated by applying the threshold transformation can be fed into a feed-forward process to adjust a later processing step for the substrate, or into a feed-back process to adjust the processing step for a subsequent substrate.
Accordingly, other implementations are within the scope of the claims.

Claims (18)

What is claimed is:
1. A method of controlling polishing, comprising:
obtaining a first color image of a first substrate at an in-line monitoring station of a polishing system;
dividing the first color image into a plurality of second color images using a die mask such that each second color image corresponds to a region for a die being fabricated on the first substrate;
generating thickness measurements for one or more locations, each respective location of the one more locations corresponding to a respective region for a die being fabricated on the first substrate, wherein generating a thickness measurement for a region includes processing a second color image corresponding to the region through a neural network that was trained using training data that included a plurality of third color images of dies of a calibration substrate and ground truth thickness measurements of the calibration substrate with each respective third color image paired with a ground truth thickness measurement for a die region associated with the respective third color image; and
determining a value for a polishing parameter for the first substrate or a subsequent second substrate based on the thickness measurements.
2. The method of claim 1, wherein determining the value for the polishing parameter comprises calculating a polishing pressure for a chamber in a carrier head.
3. The method of claim 1, comprising obtaining the first color image of the first substrate after polishing of the first substrate at a polishing station.
4. The method of claim 3, comprising determining the value for the polishing parameter for the polishing station for the subsequent second substrate based on the thickness measurements.
5. The method of claim 1, comprising obtaining the first color image of the first substrate before polishing of the first substrate at a polishing station.
6. The method of claim 5, comprising determining the value for the polishing parameter for the polishing station for the first substrate based on the thickness measurements.
7. A computer program product, comprising a non-transitory computer-readable medium encoded with instructions to cause one or more processors to:
receive a first color image of a first substrate from an in-line monitoring station of a polishing system;
divide the first color image into a plurality of second color images using a die mask such that each second color image corresponds to a region for a die being fabricated on the first substrate;
generate thickness measurements for one or more locations, each respective location of the one more locations corresponding to a respective region for a die being fabricated on the first substrate, wherein the instructions to generate the thickness measurement for a region include instructions to process a second color image corresponding to the region through a neural network that was trained using training data that included a plurality of third color images of dies of a calibration substrate and ground truth thickness measurements of the calibration substrate with each respective third color image paired with a ground truth thickness measurement for a die region associated with the respective third color image; and
determine a value for a polishing parameter for the first substrate or a subsequent second substrate based on the thickness measurements.
8. The computer program product of claim 7, comprising instructions to receive the first color image of the first substrate after polishing of the first substrate at a polishing station.
9. The computer program product of claim 8, comprising instructions to determine the polishing parameter for the polishing station for the subsequent second substrate based on the thickness measurements.
10. The computer program product of claim 7, comprising instructions to receive the first color image of the first substrate before polishing of the first substrate at a polishing station.
11. The computer program product of claim 10, comprising instructions to determine the polishing parameter for the polishing station for the first substrate based on the thickness measurements.
12. The computer program product of claim 7, wherein the polishing parameter comprises a pressure of a chamber in a carrier head.
13. A polishing apparatus, comprising:
a polishing station including a platen to support a polishing pad and a carrier head to hold a first substrate against the polishing pad;
an in-line metrology station having a color camera to generate a color image of the first substrate; and
a control system configured to
receive a first color image of the first substrate from an in-line monitoring station of a polishing system,
divide the first color image into a plurality of second color images using a die mask such that each second color image corresponds to a region for a die being fabricated on the first substrate;
generate thickness measurements for one or more locations, each respective location of the one more locations corresponding to a respective region for a die being fabricated on the first substrate, wherein the control system is configure to generate the thickness measurement for a region by processing a second color image corresponding to the region through a neural network that was trained using training data that included a plurality of third color images of dies of a calibration substrate and ground truth thickness measurements of the calibration substrate with each respective third color image paired with a ground truth thickness measurement for a die region associated with the respective third color image, and
determine a value for a polishing parameter for the first substrate or a subsequent second substrate based on the thickness measurements;
cause the polishing station to polish the first substrate or the subsequent second substrate using the determined value for the polishing parameter.
14. The apparatus of claim 13, wherein the polishing parameter comprises a pressure of a chamber in the carrier head.
15. The apparatus of claim 13, wherein the control system is configured to receive the first color image of the first substrate after polishing of the first substrate at a polishing station.
16. The apparatus of claim 15, wherein the control system is configured to determine the polishing parameter for the polishing station for the subsequent second substrate based on the thickness measurements.
17. The apparatus of claim 16, wherein control system is configured to receive the first color image of the first substrate before polishing of the first substrate at a polishing station.
18. The apparatus of claim 17, wherein control system is configured to determine the polishing parameter for the polishing station for the first substrate based on the thickness measurements.
US17/359,345 2020-06-29 2021-06-25 System using film thickness estimation from machine learning based processing of substrate images Active 2042-01-22 US11847776B2 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
US17/359,345 US11847776B2 (en) 2020-06-29 2021-06-25 System using film thickness estimation from machine learning based processing of substrate images
US18/500,811 US20240062364A1 (en) 2020-06-29 2023-11-02 System using film thickness estimation from machine learning based processing of substrate images

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US202063045782P 2020-06-29 2020-06-29
US17/359,345 US11847776B2 (en) 2020-06-29 2021-06-25 System using film thickness estimation from machine learning based processing of substrate images

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US18/500,811 Continuation US20240062364A1 (en) 2020-06-29 2023-11-02 System using film thickness estimation from machine learning based processing of substrate images

Publications (2)

Publication Number Publication Date
US20210407066A1 US20210407066A1 (en) 2021-12-30
US11847776B2 true US11847776B2 (en) 2023-12-19

Family

ID=79032657

Family Applications (4)

Application Number Title Priority Date Filing Date
US17/359,345 Active 2042-01-22 US11847776B2 (en) 2020-06-29 2021-06-25 System using film thickness estimation from machine learning based processing of substrate images
US17/359,307 Active 2042-06-02 US11836913B2 (en) 2020-06-29 2021-06-25 Film thickness estimation from machine learning based processing of substrate images
US18/496,303 Pending US20240054634A1 (en) 2020-06-29 2023-10-27 Residue classification from machine learning based processing of substrate images
US18/500,811 Pending US20240062364A1 (en) 2020-06-29 2023-11-02 System using film thickness estimation from machine learning based processing of substrate images

Family Applications After (3)

Application Number Title Priority Date Filing Date
US17/359,307 Active 2042-06-02 US11836913B2 (en) 2020-06-29 2021-06-25 Film thickness estimation from machine learning based processing of substrate images
US18/496,303 Pending US20240054634A1 (en) 2020-06-29 2023-10-27 Residue classification from machine learning based processing of substrate images
US18/500,811 Pending US20240062364A1 (en) 2020-06-29 2023-11-02 System using film thickness estimation from machine learning based processing of substrate images

Country Status (7)

Country Link
US (4) US11847776B2 (en)
EP (1) EP4172560A4 (en)
JP (2) JP7572436B2 (en)
KR (1) KR20220123069A (en)
CN (1) CN115104001A (en)
TW (2) TWI792370B (en)
WO (1) WO2022005916A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20240054634A1 (en) * 2020-06-29 2024-02-15 Applied Materials, Inc Residue classification from machine learning based processing of substrate images

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20230316486A1 (en) * 2022-03-30 2023-10-05 Applied Materials, Inc. Chemical-dose substrate deposition monitoring
CN114918817B (en) * 2022-05-27 2024-09-13 河南科技学院 Roll-to-Roll chemical mechanical polishing device and method
CN116086330B (en) * 2023-02-17 2024-01-12 无锡星微科技有限公司 Thickness detection platform for large-size wafer

Citations (40)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5897378A (en) 1995-05-17 1999-04-27 Matsushita Electric Industrial Co., Ltd. Method of monitoring deposit in chamber, method of plasma processing, method of dry-cleaning chamber, and semiconductor manufacturing apparatus
JP2000233369A (en) 1999-02-15 2000-08-29 Noritake Co Ltd Grinding condition monitoring device and dressing condition monitoring device
JP2001287159A (en) 2000-04-05 2001-10-16 Nikon Corp Surface condition measuring method and measuring device, polishing machine, and semiconductor device manufacturing method
US20020013058A1 (en) 1992-12-28 2002-01-31 Applied Materials, Inc., A Delaware Corporation In-situ real-time monitoring technique and apparatus for endpoint detection of thin films during chemical/mechanical polishing planarization
US6603874B1 (en) 1996-11-12 2003-08-05 Robotic Vision Systems, Inc. Method and system for imaging an object or pattern
US6712927B1 (en) 1998-06-11 2004-03-30 Applied Materials Inc. Chamber having process monitoring window
US20040259472A1 (en) 2003-04-01 2004-12-23 Chalmers Scott A. Whole-substrate spectral imaging system for CMP
US20050054268A1 (en) 2002-03-28 2005-03-10 Lam Research Corporation Methods for detecting transitions of wafer surface properties in chemical mechanical polishing for process status and control
US6959255B2 (en) 2002-11-04 2005-10-25 Brion Technologies, Inc. Method and apparatus for monitoring integrated circuit fabrication
US20050287929A1 (en) 2001-05-02 2005-12-29 Applied Materials, Inc., A Delwaware Corporation Integrated endpoint detection system with optical and eddy current monitoring
US20060009128A1 (en) 2000-05-19 2006-01-12 Hiroji Hanawa Eddy current sensing of metal removal for chemical mechanical polishing
US20060025052A1 (en) 2002-02-06 2006-02-02 Manoocher Birang Method and apparatus of eddy current monitoring for chemical mechanical polishing
US20060025048A1 (en) 2004-07-28 2006-02-02 3M Innovative Properties Company Abrasive article detection system and method
US7001243B1 (en) 2003-06-27 2006-02-21 Lam Research Corporation Neural network control of chemical mechanical planarization
US20060048697A1 (en) 2002-02-22 2006-03-09 Houge Erik C Monitoring and control of a fabrication process
US20060063469A1 (en) 2002-01-17 2006-03-23 Homayoun Talieh Advanced chemical mechanical polishing system with smart endpoint detection
US7076320B1 (en) 2004-05-04 2006-07-11 Advanced Micro Devices, Inc. Scatterometry monitor in cluster process tool environment for advanced process control (APC)
US20070042675A1 (en) 2005-08-22 2007-02-22 Applied Materials, Inc. Spectrum based endpointing for chemical mechanical polishing
US20070077671A1 (en) 2005-10-03 2007-04-05 Applied Materials In-situ substrate imaging
WO2007024807A3 (en) 2005-08-22 2007-07-12 Applied Materials Inc Apparatus and methods for spectrum based monitoring of chemical mechanical polishing
US20070224915A1 (en) 2005-08-22 2007-09-27 David Jeffrey D Substrate thickness measuring during polishing
JP2007266235A (en) 2006-03-28 2007-10-11 Ebara Corp Polishing device
US20080031510A1 (en) 2006-08-07 2008-02-07 Young-Seok Jung Method of and apparatus for inspecting wafers in chemical mechanical polishing equipment
JP2010067918A (en) 2008-09-12 2010-03-25 Tokyo Seimitsu Co Ltd Method and device for predicting and detecting end of polishing
US20140024291A1 (en) 2012-07-18 2014-01-23 Jimin Zhang Endpoint Detection During Polishing Using Integrated Differential Intensity
US20140206259A1 (en) 2013-01-23 2014-07-24 Applied Materials, Inc. Reflectivity measurements during polishing using a camera
US20170140525A1 (en) * 2015-11-16 2017-05-18 Dominic J. Benvegnu Color imaging for cmp monitoring
US20170177997A1 (en) 2015-12-22 2017-06-22 Applied Materials Israel Ltd. Method of deep learining-based examination of a semiconductor specimen and system thereof
US9816810B2 (en) 2013-12-23 2017-11-14 Kla-Tencor Corporation Measurement of multiple patterning parameters
US20180061032A1 (en) * 2016-08-26 2018-03-01 Applied Materials, Inc. Thickness measurement of substrate using color metrology
US20180107928A1 (en) 2016-10-14 2018-04-19 Kla-Tencor Corporation Diagnostic systems and methods for deep learning models configured for semiconductor applications
US20180150052A1 (en) 2016-11-30 2018-05-31 Applied Materials, Inc. Spectrographic monitoring using a neural network
US10012494B2 (en) 2013-10-25 2018-07-03 Applied Materials, Inc. Grouping spectral data from polishing substrates
US10030971B2 (en) 2015-08-04 2018-07-24 GlobalFoundries, Inc. Measurement system and method for measuring in thin films
US20190072482A1 (en) 2017-09-06 2019-03-07 Lam Research Corporation Systems and methods for combining optical metrology with mass metrology
US20200005140A1 (en) 2018-06-28 2020-01-02 Benjamin Cherian Training Spectrum Generation for Machine Learning System for Spectrographic Monitoring
US20200094370A1 (en) 2018-09-24 2020-03-26 Applied Materials, Inc. Machine vision as input to a cmp process control algorithm
US20200258214A1 (en) * 2019-02-07 2020-08-13 Applied Materials, Inc. Thickness Measurement of Substrate Using Color Metrology
US20210407066A1 (en) * 2020-06-29 2021-12-30 Applied Materials, Inc. System using film thickness estimation from machine learning based processing of substrate images
US20220285227A1 (en) * 2021-03-04 2022-09-08 Applied Materials, Inc. Pixel classification of film non-uniformity based on processing of substrate images

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2003249472A (en) * 2002-02-26 2003-09-05 Hitachi Ltd Method and device for film thickness measurement and manufacturing method of thin film device
KR20240089695A (en) * 2018-03-13 2024-06-20 어플라이드 머티어리얼스, 인코포레이티드 Machine learning systems for monitoring of semiconductor processing
NL2021938B1 (en) * 2018-11-05 2020-05-15 Suss Microtec Lithography Gmbh Method for measuring a thickness of a layer, method for controlling a substrate processing device as well as substrate processing device

Patent Citations (45)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20020013058A1 (en) 1992-12-28 2002-01-31 Applied Materials, Inc., A Delaware Corporation In-situ real-time monitoring technique and apparatus for endpoint detection of thin films during chemical/mechanical polishing planarization
US5897378A (en) 1995-05-17 1999-04-27 Matsushita Electric Industrial Co., Ltd. Method of monitoring deposit in chamber, method of plasma processing, method of dry-cleaning chamber, and semiconductor manufacturing apparatus
US6603874B1 (en) 1996-11-12 2003-08-05 Robotic Vision Systems, Inc. Method and system for imaging an object or pattern
US6712927B1 (en) 1998-06-11 2004-03-30 Applied Materials Inc. Chamber having process monitoring window
JP2000233369A (en) 1999-02-15 2000-08-29 Noritake Co Ltd Grinding condition monitoring device and dressing condition monitoring device
JP2001287159A (en) 2000-04-05 2001-10-16 Nikon Corp Surface condition measuring method and measuring device, polishing machine, and semiconductor device manufacturing method
US20060009128A1 (en) 2000-05-19 2006-01-12 Hiroji Hanawa Eddy current sensing of metal removal for chemical mechanical polishing
US20050287929A1 (en) 2001-05-02 2005-12-29 Applied Materials, Inc., A Delwaware Corporation Integrated endpoint detection system with optical and eddy current monitoring
US20060063469A1 (en) 2002-01-17 2006-03-23 Homayoun Talieh Advanced chemical mechanical polishing system with smart endpoint detection
US20060025052A1 (en) 2002-02-06 2006-02-02 Manoocher Birang Method and apparatus of eddy current monitoring for chemical mechanical polishing
US20060048697A1 (en) 2002-02-22 2006-03-09 Houge Erik C Monitoring and control of a fabrication process
US20050054268A1 (en) 2002-03-28 2005-03-10 Lam Research Corporation Methods for detecting transitions of wafer surface properties in chemical mechanical polishing for process status and control
US6959255B2 (en) 2002-11-04 2005-10-25 Brion Technologies, Inc. Method and apparatus for monitoring integrated circuit fabrication
US20040259472A1 (en) 2003-04-01 2004-12-23 Chalmers Scott A. Whole-substrate spectral imaging system for CMP
US7001243B1 (en) 2003-06-27 2006-02-21 Lam Research Corporation Neural network control of chemical mechanical planarization
US7076320B1 (en) 2004-05-04 2006-07-11 Advanced Micro Devices, Inc. Scatterometry monitor in cluster process tool environment for advanced process control (APC)
US20060025048A1 (en) 2004-07-28 2006-02-02 3M Innovative Properties Company Abrasive article detection system and method
US20070042675A1 (en) 2005-08-22 2007-02-22 Applied Materials, Inc. Spectrum based endpointing for chemical mechanical polishing
WO2007024807A3 (en) 2005-08-22 2007-07-12 Applied Materials Inc Apparatus and methods for spectrum based monitoring of chemical mechanical polishing
US20070224915A1 (en) 2005-08-22 2007-09-27 David Jeffrey D Substrate thickness measuring during polishing
US20070077671A1 (en) 2005-10-03 2007-04-05 Applied Materials In-situ substrate imaging
JP2007266235A (en) 2006-03-28 2007-10-11 Ebara Corp Polishing device
US20080031510A1 (en) 2006-08-07 2008-02-07 Young-Seok Jung Method of and apparatus for inspecting wafers in chemical mechanical polishing equipment
JP2010067918A (en) 2008-09-12 2010-03-25 Tokyo Seimitsu Co Ltd Method and device for predicting and detecting end of polishing
US20140024291A1 (en) 2012-07-18 2014-01-23 Jimin Zhang Endpoint Detection During Polishing Using Integrated Differential Intensity
US20140206259A1 (en) 2013-01-23 2014-07-24 Applied Materials, Inc. Reflectivity measurements during polishing using a camera
US10012494B2 (en) 2013-10-25 2018-07-03 Applied Materials, Inc. Grouping spectral data from polishing substrates
TW201940833A (en) 2013-12-23 2019-10-16 美商克萊譚克公司 Measurement of multiple patterning parameters
US9816810B2 (en) 2013-12-23 2017-11-14 Kla-Tencor Corporation Measurement of multiple patterning parameters
US10030971B2 (en) 2015-08-04 2018-07-24 GlobalFoundries, Inc. Measurement system and method for measuring in thin films
CN108966674B (en) 2015-08-04 2023-06-27 诺威有限公司 Hybrid measurement system and method for measurement in thin films
US20170140525A1 (en) * 2015-11-16 2017-05-18 Dominic J. Benvegnu Color imaging for cmp monitoring
US20170177997A1 (en) 2015-12-22 2017-06-22 Applied Materials Israel Ltd. Method of deep learining-based examination of a semiconductor specimen and system thereof
US10325364B2 (en) * 2016-08-26 2019-06-18 Applied Materials, Inc. Thickness measurement of substrate using color metrology
US20180061032A1 (en) * 2016-08-26 2018-03-01 Applied Materials, Inc. Thickness measurement of substrate using color metrology
US20180107928A1 (en) 2016-10-14 2018-04-19 Kla-Tencor Corporation Diagnostic systems and methods for deep learning models configured for semiconductor applications
US20180150052A1 (en) 2016-11-30 2018-05-31 Applied Materials, Inc. Spectrographic monitoring using a neural network
US20190072482A1 (en) 2017-09-06 2019-03-07 Lam Research Corporation Systems and methods for combining optical metrology with mass metrology
US20200005140A1 (en) 2018-06-28 2020-01-02 Benjamin Cherian Training Spectrum Generation for Machine Learning System for Spectrographic Monitoring
US20200094370A1 (en) 2018-09-24 2020-03-26 Applied Materials, Inc. Machine vision as input to a cmp process control algorithm
US20200258214A1 (en) * 2019-02-07 2020-08-13 Applied Materials, Inc. Thickness Measurement of Substrate Using Color Metrology
US11100628B2 (en) * 2019-02-07 2021-08-24 Applied Materials, Inc. Thickness measurement of substrate using color metrology
US20210407066A1 (en) * 2020-06-29 2021-12-30 Applied Materials, Inc. System using film thickness estimation from machine learning based processing of substrate images
US20210407065A1 (en) * 2020-06-29 2021-12-30 Applied Materials, Inc. Film thickness estimation from machine learning based processing of substrate images
US20220285227A1 (en) * 2021-03-04 2022-09-08 Applied Materials, Inc. Pixel classification of film non-uniformity based on processing of substrate images

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
International Search Report and Written Opinion in International Appln. No. PCT/US2021/0339247, dated Oct. 6, 2021, 11 pages.
McNeil, "Scatterometry Applied to Microelectronics Processing," presented at LEOS Summer Topical Meetings, Aventura, FL, Jul. 24-28, 2000; IEEE, Aug. 2002. 37-38.
Office Action in Taiwanese Appln. No. 110123728, dated May 2, 2022, 11 pages (with English Summary).

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20240054634A1 (en) * 2020-06-29 2024-02-15 Applied Materials, Inc Residue classification from machine learning based processing of substrate images
US20240062364A1 (en) * 2020-06-29 2024-02-22 Applied Materials, Inc. System using film thickness estimation from machine learning based processing of substrate images

Also Published As

Publication number Publication date
US20210407065A1 (en) 2021-12-30
US20240062364A1 (en) 2024-02-22
JP2023516872A (en) 2023-04-21
US20210407066A1 (en) 2021-12-30
EP4172560A4 (en) 2024-05-29
US20240054634A1 (en) 2024-02-15
WO2022005916A1 (en) 2022-01-06
EP4172560A1 (en) 2023-05-03
JP7572436B2 (en) 2024-10-23
TW202211072A (en) 2022-03-16
CN115104001A (en) 2022-09-23
TW202321970A (en) 2023-06-01
TWI792370B (en) 2023-02-11
JP2024117768A (en) 2024-08-29
KR20220123069A (en) 2022-09-05
US11836913B2 (en) 2023-12-05

Similar Documents

Publication Publication Date Title
US11682114B2 (en) Thickness measurement of substrate using color metrology
US11847776B2 (en) System using film thickness estimation from machine learning based processing of substrate images
US11776109B2 (en) Thickness measurement of substrate using color metrology
TWI848669B (en) Non-transitory computer readable medium, method, and polishing system for pixel and region classification of film non-uniformity based on processing of substrate images
US11557048B2 (en) Thickness measurement of substrate using color metrology
KR102672511B1 (en) Board thickness measurement using color metrology
US12148148B2 (en) Thickness measurement of substrate using color metrology

Legal Events

Date Code Title Description
FEPP Fee payment procedure

Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS

STPP Information on status: patent application and granting procedure in general

Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT VERIFIED

STPP Information on status: patent application and granting procedure in general

Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT VERIFIED

STCF Information on status: patent grant

Free format text: PATENTED CASE

CC Certificate of correction