[go: up one dir, main page]
More Web Proxy on the site http://driver.im/

US20230253073A1 - System for acquisition and processing of multiplexed fluorescence in-situ hybridization images - Google Patents

System for acquisition and processing of multiplexed fluorescence in-situ hybridization images Download PDF

Info

Publication number
US20230253073A1
US20230253073A1 US18/301,875 US202318301875A US2023253073A1 US 20230253073 A1 US20230253073 A1 US 20230253073A1 US 202318301875 A US202318301875 A US 202318301875A US 2023253073 A1 US2023253073 A1 US 2023253073A1
Authority
US
United States
Prior art keywords
image
sample
images
filter
excitation light
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
US18/301,875
Inventor
Yun-Ching CHANG
Dan Xie
Chloe Kim
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Applied Materials Inc
Original Assignee
Applied Materials Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Applied Materials Inc filed Critical Applied Materials Inc
Priority to US18/301,875 priority Critical patent/US20230253073A1/en
Assigned to APPLIED MATERIALS, INC. reassignment APPLIED MATERIALS, INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: XIE, Dan, CHANG, YUN-CHING, KIM, CHLOE
Publication of US20230253073A1 publication Critical patent/US20230253073A1/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16BBIOINFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR GENETIC OR PROTEIN-RELATED DATA PROCESSING IN COMPUTATIONAL MOLECULAR BIOLOGY
    • G16B40/00ICT specially adapted for biostatistics; ICT specially adapted for bioinformatics-related machine learning or data mining, e.g. knowledge discovery or pattern finding
    • G16B40/10Signal processing, e.g. from mass spectrometry [MS] or from PCR
    • CCHEMISTRY; METALLURGY
    • C12BIOCHEMISTRY; BEER; SPIRITS; WINE; VINEGAR; MICROBIOLOGY; ENZYMOLOGY; MUTATION OR GENETIC ENGINEERING
    • C12QMEASURING OR TESTING PROCESSES INVOLVING ENZYMES, NUCLEIC ACIDS OR MICROORGANISMS; COMPOSITIONS OR TEST PAPERS THEREFOR; PROCESSES OF PREPARING SUCH COMPOSITIONS; CONDITION-RESPONSIVE CONTROL IN MICROBIOLOGICAL OR ENZYMOLOGICAL PROCESSES
    • C12Q1/00Measuring or testing processes involving enzymes, nucleic acids or microorganisms; Compositions therefor; Processes of preparing such compositions
    • C12Q1/68Measuring or testing processes involving enzymes, nucleic acids or microorganisms; Compositions therefor; Processes of preparing such compositions involving nucleic acids
    • C12Q1/6813Hybridisation assays
    • C12Q1/6841In situ hybridisation
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01NINVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
    • G01N21/00Investigating or analysing materials by the use of optical means, i.e. using sub-millimetre waves, infrared, visible or ultraviolet light
    • G01N21/62Systems in which the material investigated is excited whereby it emits light or causes a change in wavelength of the incident light
    • G01N21/63Systems in which the material investigated is excited whereby it emits light or causes a change in wavelength of the incident light optically excited
    • G01N21/64Fluorescence; Phosphorescence
    • G01N21/6428Measuring fluorescence of fluorescent products of reactions or of fluorochrome labelled reactive substances, e.g. measuring quenching effects, using measuring "optrodes"
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01NINVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
    • G01N21/00Investigating or analysing materials by the use of optical means, i.e. using sub-millimetre waves, infrared, visible or ultraviolet light
    • G01N21/62Systems in which the material investigated is excited whereby it emits light or causes a change in wavelength of the incident light
    • G01N21/63Systems in which the material investigated is excited whereby it emits light or causes a change in wavelength of the incident light optically excited
    • G01N21/64Fluorescence; Phosphorescence
    • G01N21/645Specially adapted constructive features of fluorimeters
    • G01N21/6456Spatial resolved fluorescence measurements; Imaging
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01NINVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
    • G01N21/00Investigating or analysing materials by the use of optical means, i.e. using sub-millimetre waves, infrared, visible or ultraviolet light
    • G01N21/62Systems in which the material investigated is excited whereby it emits light or causes a change in wavelength of the incident light
    • G01N21/63Systems in which the material investigated is excited whereby it emits light or causes a change in wavelength of the incident light optically excited
    • G01N21/64Fluorescence; Phosphorescence
    • G01N21/645Specially adapted constructive features of fluorimeters
    • G01N21/6456Spatial resolved fluorescence measurements; Imaging
    • G01N21/6458Fluorescence microscopy
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B21/00Microscopes
    • G02B21/0004Microscopes specially adapted for specific applications
    • G02B21/002Scanning microscopes
    • G02B21/0024Confocal scanning microscopes (CSOMs) or confocal "macroscopes"; Accessories which are not restricted to use with CSOMs, e.g. sample holders
    • G02B21/0052Optical details of the image generation
    • G02B21/0076Optical details of the image generation arrangements using fluorescence or luminescence
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B21/00Microscopes
    • G02B21/16Microscopes adapted for ultraviolet illumination ; Fluorescence microscopes
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B21/00Microscopes
    • G02B21/36Microscopes arranged for photographic purposes or projection purposes or digital imaging or video purposes including associated control and data processing arrangements
    • G02B21/365Control or image processing arrangements for digital or video microscopes
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B21/00Microscopes
    • G02B21/36Microscopes arranged for photographic purposes or projection purposes or digital imaging or video purposes including associated control and data processing arrangements
    • G02B21/365Control or image processing arrangements for digital or video microscopes
    • G02B21/367Control or image processing arrangements for digital or video microscopes providing an output produced by processing a plurality of individual source images, e.g. image tiling, montage, composite images, depth sectioning, image comparison
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformations in the plane of the image
    • G06T3/40Scaling of whole images or parts thereof, e.g. expanding or contracting
    • G06T3/4038Image mosaicing, e.g. composing plane images from plane sub-images
    • G06T5/003
    • G06T5/006
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/20Image enhancement or restoration using local operators
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/50Image enhancement or restoration using two or more images, e.g. averaging or subtraction
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/73Deblurring; Sharpening
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/80Geometric correction
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • G06T7/0012Biomedical image inspection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/30Determination of transform parameters for the alignment of images, i.e. image registration
    • G06T7/33Determination of transform parameters for the alignment of images, i.e. image registration using feature-based methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/30Noise filtering
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/50Extraction of image or video features by performing operations within image blocks; by using histograms, e.g. histogram of oriented gradients [HoG]; by summing image-intensity values; Projection analysis
    • G06V10/507Summing image-intensity values; Histogram projection analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/74Image or video pattern matching; Proximity measures in feature spaces
    • G06V10/75Organisation of the matching processes, e.g. simultaneous or sequential comparisons of image or video features; Coarse-fine approaches, e.g. multi-scale approaches; using context analysis; Selection of dictionaries
    • G06V10/758Involving statistics of pixels or of feature values, e.g. histogram matching
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/60Type of objects
    • G06V20/69Microscopic objects, e.g. biological cells or cellular parts
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/60Type of objects
    • G06V20/69Microscopic objects, e.g. biological cells or cellular parts
    • G06V20/698Matching; Classification
    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16BBIOINFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR GENETIC OR PROTEIN-RELATED DATA PROCESSING IN COMPUTATIONAL MOLECULAR BIOLOGY
    • G16B50/00ICT programming tools or database systems specially adapted for bioinformatics
    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16BBIOINFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR GENETIC OR PROTEIN-RELATED DATA PROCESSING IN COMPUTATIONAL MOLECULAR BIOLOGY
    • G16B50/00ICT programming tools or database systems specially adapted for bioinformatics
    • G16B50/40Encryption of genetic data
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01NINVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
    • G01N21/00Investigating or analysing materials by the use of optical means, i.e. using sub-millimetre waves, infrared, visible or ultraviolet light
    • G01N21/62Systems in which the material investigated is excited whereby it emits light or causes a change in wavelength of the incident light
    • G01N21/63Systems in which the material investigated is excited whereby it emits light or causes a change in wavelength of the incident light optically excited
    • G01N21/64Fluorescence; Phosphorescence
    • G01N2021/6417Spectrofluorimetric devices
    • G01N2021/6419Excitation at two or more wavelengths
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01NINVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
    • G01N21/00Investigating or analysing materials by the use of optical means, i.e. using sub-millimetre waves, infrared, visible or ultraviolet light
    • G01N21/62Systems in which the material investigated is excited whereby it emits light or causes a change in wavelength of the incident light
    • G01N21/63Systems in which the material investigated is excited whereby it emits light or causes a change in wavelength of the incident light optically excited
    • G01N21/64Fluorescence; Phosphorescence
    • G01N2021/6417Spectrofluorimetric devices
    • G01N2021/6421Measuring at two or more wavelengths
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01NINVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
    • G01N21/00Investigating or analysing materials by the use of optical means, i.e. using sub-millimetre waves, infrared, visible or ultraviolet light
    • G01N21/62Systems in which the material investigated is excited whereby it emits light or causes a change in wavelength of the incident light
    • G01N21/63Systems in which the material investigated is excited whereby it emits light or causes a change in wavelength of the incident light optically excited
    • G01N21/64Fluorescence; Phosphorescence
    • G01N21/6428Measuring fluorescence of fluorescent products of reactions or of fluorochrome labelled reactive substances, e.g. measuring quenching effects, using measuring "optrodes"
    • G01N2021/6439Measuring fluorescence of fluorescent products of reactions or of fluorochrome labelled reactive substances, e.g. measuring quenching effects, using measuring "optrodes" with indicators, stains, dyes, tags, labels, marks
    • G01N2021/6441Measuring fluorescence of fluorescent products of reactions or of fluorochrome labelled reactive substances, e.g. measuring quenching effects, using measuring "optrodes" with indicators, stains, dyes, tags, labels, marks with two or more labels
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01NINVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
    • G01N21/00Investigating or analysing materials by the use of optical means, i.e. using sub-millimetre waves, infrared, visible or ultraviolet light
    • G01N21/75Systems in which material is subjected to a chemical reaction, the progress or the result of the reaction being investigated
    • G01N2021/755Comparing readings with/without reagents, or before/after reaction
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01NINVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
    • G01N21/00Investigating or analysing materials by the use of optical means, i.e. using sub-millimetre waves, infrared, visible or ultraviolet light
    • G01N21/01Arrangements or apparatus for facilitating the optical investigation
    • G01N21/03Cuvette constructions
    • G01N21/05Flow-through cuvettes
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01NINVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
    • G01N2201/00Features of devices classified in G01N21/00
    • G01N2201/12Circuits of general importance; Signal processing
    • G01N2201/127Calibration; base line adjustment; drift compensation
    • G01N2201/12746Calibration values determination
    • G01N2201/12761Precalibration, e.g. for a given series of reagents
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10056Microscopic image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10064Fluorescence image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30004Biomedical image processing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30168Image quality inspection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30204Marker

Definitions

  • This specification relates multiplexed fluorescence in-situ hybridization imaging.
  • exemplary methods include multiplex imaging of antibodies or oligonucleotides (e.g., RNA or DNA) labeled with a functional domain to a biological sample.
  • Multiplexed fluorescence in-situ hybridization (mFISH) imaging is a powerful technique to determine gene expression in spatial transcriptomics.
  • a sample is exposed to multiple oligonucleotide probes that target RNA of interest. These probes have different labeling schemes that will allow one to distinguish different RNA species when the complementary, fluorescent labeled probes are introduced to the sample.
  • the sequential rounds of fluorescence images are acquired with exposure to excitation light of different wavelengths. For each given pixel, its fluorescence intensities from the different images for the different wavelengths of excitation light form a signal sequence.
  • This sequence is then compared to a library of reference codes from a codebook that associates each code with a gene. The best matching reference code is used to identify an associated gene that is expressed at that pixel in the image.
  • a fluorescent in-situ hybridization imaging system includes a flow cell to contain a sample to be exposed to fluorescent probes in a reagent, a valve to control flow from one of a plurality of reagent sources the flow cell, a pump to cause fluid flow through the flow cell, a fluorescence microscope including a variable frequency excitation light source and a camera positioned to receive fluorescently emitted light from the sample, an actuator to cause relative vertical motion between the flow cell and the fluorescence microscope, a motor to cause to cause relative lateral motion between the flow cell and the fluorescence microscope, and a control system.
  • the control system is configured to, as nested loops, cause the valve to sequentially couple the flow cell to a plurality of different reagent sources to expose the sample to a plurality of different reagents, for each reagent of the plurality of different reagents, cause the motor to sequentially position the fluorescence microscope relative to sample at a plurality of different fields of view, for each field of view of the plurality of different fields of view, cause the variable frequency excitation light source to sequentially emit a plurality of different wavelengths, for each wavelength of the plurality of different wavelengths, cause the actuator to sequentially position the fluorescence microscope relative to sample at a plurality of different vertical heights, and for each vertical height of the plurality of different vertical heights, obtain an image at the respective vertical height covering the respective field of view of the sample having respective fluorescent probes of the respective regent as excited by the respective wavelength.
  • a fluorescent in-situ hybridization imaging and analysis system includes a flow cell to contain a sample to be exposed to fluorescent probes in a reagent, a fluorescence microscope to obtain sequentially collect a plurality of images of the sample at a plurality of different combinations of imaging parameters, and a data processing system.
  • the data processing system includes an online pre-processing system configured to sequentially receive the images from the fluorescence microscope as the images are collected and perform on-the-fly image pre-processing to remove experimental artifacts of the image and to provide RNA image spot sharpening, and an offline processing system configured to, after the plurality of images are collected, perform registration of images having a same field of view and to decode intensity values in the plurality of images to identify expressed genes.
  • a fluorescent in-situ hybridization imaging system in another aspect, includes a flow cell to contain a sample to be exposed to fluorescent probes in a reagent, a fluorescence microscope, and a control system.
  • the fluorescence microscope includes a variable frequency excitation light source to illuminate the sample, a plurality of emission bandpass filters on a filter wheel, an actuator to rotate the filter wheel, and a camera positioned to receive fluorescently emitted light from the sample that has passed through a filter on the filter wheel.
  • the control system is configured to cause the variable frequency excitation light source to emit a light beam having a selected wavelength, cause the actuator to rotate the filter wheel to position a selected filter in a light path between the sample and the camera, obtain an image from the camera, and coordinate the variable frequency excitation light source and filter wheel such that the selected filter has an emission bandpass associated with emission by the fluorescent probes when excited by the selected wavelength when the image is obtained.
  • a computer program product has instructions to cause one or more processors to apply one or more filters to and perform deconvolution on each image of the plurality of images to generate a plurality of filtered images, perform registration of the plurality of filtered images to generate a plurality of registered and filtered images, for each pixel of at least two pixels of the plurality of registered and filtered images decode the pixel by identifying a code word from a plurality of code words in a code book that provides a best match to data values in the plurality of registered and filtered images for the pixel, and for each pixel of at least two pixels, determine a gene associated with the code word and store an indication that the gene is expressed at the pixel.
  • a computer program product has instructions to receive a plurality of initial images of a sample, the plurality of initial images sharing a common field of view. For each initial image of the plurality of initial images, the initial image is split into a plurality of sub-images of equal size, a combined image of size equal to that of a sub-image is calculated from the plurality of sub-images to thereby generate a plurality of reduced-size images with each image of the plurality of images having a corresponding reduced-size image, for at least one reduced-size images of the plurality of reduced-size images a transform is determined to register the plurality of reduced-size images, for each of the at least one reduced-size images the transform determined for the reduced-size image is applied to the corresponding initial image to generate a transformed image of size equal to the initial image and thereby generate a plurality of transformed images, and for each pixel of at least two pixels of the plurality of transformed images the pixel is decoded by identifying a code word from a plurality of code words in
  • a computer program product has instructions to receive a plurality of images representing a common field of view of a sample, perform registration of the plurality of images to generate a plurality of registered images, for each pixel of a plurality of pixels of the registered images decode the pixel by identifying a code word from a plurality of code words in a code book that provides a best match to data values in the plurality of registered images for the pixel, and for each code word identified as a best match for one or more pixels and each pixel of the one or more pixels determine whether a bit ratio for an image word for the pixel meets a threshold for the code word.
  • the image word is formed from the data values in the plurality of registered images for the pixel. For at least one pixel that is determined to meet the threshold a gene associated with the code word is determined and an indication is stored that the gene is expressed at the pixel. Pixels for which the bit ratio does not meet the threshold are screened.
  • a computer program product has instructions to receive a plurality of images representing a common field of view of a sample, perform registration of the plurality of images to generate a plurality of registered images, for each pixel of a plurality of pixels of the registered images decode the pixel by identifying a code word from a plurality of code words in a code book that provides a best match to data values in the plurality of registered images for the pixel, and for each code word identified as a best match for one or more pixels and each pixel of the one or more pixels whether a bit brightness for an image word for the pixel meets a threshold for the code word is determined.
  • the image word is formed from the data values in the plurality of registered images for the pixel. For each pixel that is determined to meet the threshold a gene associated with the code word is determined and an indication is stored that the gene is expressed at the pixel. Pixels for which the bit ratio does not meet the threshold are screened.
  • a computer program product has instructions to receive a plurality of initial images of a sample.
  • the plurality of initial images including a multiplicity of images for each field of view of a plurality of different fields of view.
  • Instructions are stored to perform an image processing algorithm in which one or more of image filtering of the multiplicity of images for the field of view or registration of the multiplicity of images for the field of view.
  • For each field of view one or more initial parameter values are determined, wherein different fields of view have different initial parameter values.
  • the image processing algorithm is performed for each field of view using the different initial decoding parameter values for the different fields of view.
  • One or more subsequent decoding parameter values are determined, and the image processing algorithm and a decoding algorithm are performed for each field of view using the same subsequent parameter values for the different fields of view to generate a set of spatial transcriptomic data.
  • a computer program product has instructions to receive a plurality of initial images of a sample, the plurality of initial images including a multiplicity of images for each field of view of a plurality of different fields of view. For each field of view, a set of spatial transcriptomic data is generated based on the multiplicity of images for the field of view. Each set of spatial transcriptomic data includes one or more data pairs, each data pair of the one or more data pairs including a position and an expressed gene for that position. Regions of overlap of multiplicities of images from adjacent fields of view are determined. From the set of spatial transcriptomic data for each field of view, a combined set of spatial transcriptomic data is generated for a combined image covering the plurality of different fields of view. In generating the combined set of spatial transcriptomic data, at least one double-counted data pair from different fields of view that represents the same expressed gene at the same location within the sample is screened.
  • a computer program product has instructions to receive a plurality of images of a sample, the plurality of images including a multiplicity of images for each field of view of a plurality of different fields of view.
  • an image quality check is performed on the image.
  • artifacts are removed from each image of the plurality of images.
  • image spots are sharpened in each image of the plurality of images.
  • registration of the multiplicity of images is performed to generate a plurality of registered images.
  • the pixel After performing registration, for each multiplicity of images, and for each pixel of a plurality of pixels of the registered images of a respective multiplicity, the pixel is decoded by identifying a code word from a plurality of code words in a code book that provides a best match to data values in the respective multiplicity of registered images for the pixel.
  • mFISH multiplexed fluorescence in-situ hybridization
  • mFISH multiplexed fluorescence in-situ hybridization
  • FOVs fields-of-view
  • Run-to-run consistency in multi-sample or multi-modal research can be improved. Fluorescence images of the same FOV captured at different times can be registered without using separate fiducials, e.g., without requiring fiducial beads, and using fewer computational resources and more accurately than some conventional techniques.
  • FIG. 1 is a schematic diagram of an apparatus for multiplexed fluorescence in-situ hybridization imaging.
  • FIG. 2 is a flow chart of a method of data processing.
  • FIG. 3 schematically illustrates an image registration process that uses a mutual-information algorithm.
  • FIG. 4 is a logarithmic histogram of intensities in an image.
  • FIG. 5 A is a collected image.
  • FIG. 5 B is illustrates a mask generated from the image of FIG. 5 A .
  • FIG. 6 A is a logarithmic histogram of intensities from a filtered and de-convoluted image.
  • FIG. 6 B is a logarithmic histogram of intensities after normalization.
  • FIG. 7 illustrates a method of decoding
  • FIG. 8 A is a logarithmic histogram of bit ratios for pixels identified for a sample code word.
  • FIG. 8 B is a logarithmic histogram of brightness for pixels identified for a sample code word.
  • FIG. 9 is a flow chart of a method of stitching.
  • FIG. 10 is a flow chart of another implementation of a method of data processing in which image processing is performed on-the-fly.
  • mFISH multiplexed fluorescence in-situ hybridization
  • the overall architecture and individual components and steps enable an end-to-end mFISH imaging and decoding process that involves minimal manual steps.
  • the system integrates fully automatic hardware architecture and on-the-fly image processing.
  • the methods include feedback control, image calibration, and global optimization.
  • the end-to-end process time can be reduced and throughput per sample can be increased.
  • the apparatus can be easy to use by a research scientist in an industry context. Control metrics can ensure data quality before downstream analysis; the early detection reduces waste of samples.
  • the algorithm can enable scalability to large imaging areas with uniform and consistent spatial gene expression. Data accuracy and repeatability for large-scale study can be improved; the results can be easily interpreted and further analyzed in software packages.
  • a multiplexed fluorescent in-situ hybridization (mFISH) imaging and image processing apparatus 100 includes a flow cell 110 to hold a sample 10, a fluorescence microscope 120 to obtain images of the sample 10, and a control system 140 to control operation of the various components of the mFISH imaging and image processing apparatus 100 .
  • the control system 140 can include a computer 142 , e.g., having a memory, processor, etc., that executes control software.
  • the fluorescence microscope 120 includes an excitation light source 122 that can generate excitation light 130 of multiple different wavelengths.
  • the excitation light source 122 can generate narrow-bandwidth light beams having different wavelengths at different times.
  • the excitation light source 122 can be provided by a multi-wavelength continuous wave laser system, e.g., multiple laser modules 122 a that can be independently activated to generate laser beams of different wavelengths. Output from the laser modules 122 a can be multiplexed into a common light beam path.
  • the fluorescence microscope 120 includes a microscope body 124 that includes the various optical components to direct the excitation light from the light source 122 to the flow cell 110 .
  • excitation light from the light source 122 can be coupled into a multimode fiber, refocused and expanded by a set of lenses, then directed into the sample by a core imaging component, such as a high numerical aperture (NA) objective lens 136 .
  • NA numerical aperture
  • the excitation channel needs to be switched, one of the multiple laser modules 122 a can be deactivated and another laser module 122 a can be activated, with synchronization among the devices accomplished by one or more microcontrollers 144 , 146 .
  • the objective lens 136 can be installed on vertically movable mount coupled to a Z-drive actuator. Adjustment of the Z-position, e.g., by a microcontroller 146 controlling the Z-drive actuator, can enable fine tuning of focal position.
  • the flow cell 110 or a stage 118 supporting the sample in the flow cell 110
  • a Z-drive actuator 118 b e.g., an axial piezo stage.
  • Such a piezo stage can permit precise and swift multi-plane image acquisition.
  • the sample 10 to be imaged is positioned in the flow cell 110 .
  • the flow cell 110 can be a chamber with cross-sectional area (parallel to the object or image plane of the microscope) with and area of about 2 cm by 2 cm.
  • the sample 10 can be supported on a stage 118 within the flow cell, and the stage (or the entire flow cell) can be laterally movable, e.g., by a pair of linear actuators 118 a to permit XY motion. This permits acquisition of images of the sample 10 in different laterally offset fields of view (FOVs).
  • the microscope body 124 could be carried on a laterally movable stage.
  • An entrance to the flow cell 110 is connected to a set of hybridization reagents sources 112 .
  • a multi-valve positioner 114 can be controlled by the controller 140 to switch between sources to select which reagent 112 a is supplied to the flow cell 110 .
  • Each reagent includes a different set of one or more oligonucleotide probes. Each probe targets a different RNA sequence of interest, and has a different set of one or more fluorescent materials, e.g., phosphors, that are excited by different combinations of wavelengths.
  • a source of a purge fluid 112 b e.g., DI water.
  • An exit to the flow cell 110 is connected to a pump 116 , e.g., a peristaltic pump, which is also controlled by the controller 140 to control flow of liquid, e.g., the reagent or purge fluid, through the flow cell 110 .
  • a pump 116 e.g., a peristaltic pump
  • Used solution from the flow cell 110 can be passed by the pump 116 to a chemical waste management subsystem 119 .
  • the controller 140 causes the light source 122 to emit the excitation light 130 , which causes fluorescence of fluorescent material in the sample 10, e.g., fluorescence of the probes that are bound to RNA in the sample and that are excited by the wavelength of the excitation light.
  • the emitted fluorescent light 132 as well as back propagating excitation light, e.g., excitation light scattered from the sample, stage, etc., is collected by an objective lens 136 of the microscope body 124 .
  • the collected light can be filtered by a multi-band dichroic mirror 138 in the microscope body 124 to separate the emitted fluorescent light from the back propagating illumination light, and the emitted fluorescent light is passed to a camera 134 .
  • the multi-band dichroic mirror 138 can include a pass band for each emission wavelength expected from the probes under the variety of excitation wavelengths. Use of a single multi-band dichroic mirror (as compared to multiple dichroic mirrors or a movable dichroic mirror) can provide improved system stability.
  • the camera 134 can be a high resolution (e.g., 2048 ⁇ 2048 pixel) CMOS (e.g., a scientific CMOS) camera, and can be installed at the immediate image plane of the objective. Other camera types, e.g., CCD, may be possible.
  • image data from the camera can be captured, e.g., sent to an image processing system 150 .
  • the camera 134 can collect a sequence of images from the sample.
  • each laser emission wavelength can be paired with a corresponding band-pass emission filter 128 a .
  • Each filter 128 a can have a wavelength of 10-50 nm, e.g., 14-32 nm.
  • a filter is narrower than the bandwidth of the fluorescent material of the probe resulting from the excitation, e.g., if the fluorescent material of the probe has a long trailing spectral profile.
  • the filters are installed on a high-speed filter wheel 128 that is rotatable by an actuator 128 b .
  • the filter wheel 128 can be installed at the infinity space to minimize optical aberration in the imaging path.
  • the cleaned fluorescence signals can be refocused by a tube lens and captured by the camera 134 .
  • the dichroic mirror 138 can be positioned in the light path between the objective lens 138 and the filter wheel 128 .
  • the control system 140 can include two microcontrollers 144 , 146 that are employed to send trigger signals, e.g., TTL signals, to the components of the fluorescence microscope 120 in a coordinated manner.
  • the first microcontroller 144 is directly run by the computer 142 , and triggers actuator 128 b of the filter wheel 128 to switch emission filters 128 a at different color channels.
  • the first microcontroller 144 also triggers the second microcontroller 146 , which sends digital signals to the light source 122 in order to control which wavelength of light is passed to the sample 10.
  • the second microcontroller 146 can send on/off signals to the individual laser modules of the light source 122 to control which laser module is active, and thus control which wavelength of light is used for the excitation light. After completion of switching to a new excitation channel, the second microcontroller 146 controls the motor for the piezo stage 116 to select the imaging height. Finally the second microcontroller 146 sends a trigger signal to the camera 134 for image acquisition.
  • This control software can integrate drivers of all the device components into a single framework, and thus can allow a user to operate the imaging system as a single instrument (instead of having to separately control many devices).
  • the control software supports interactive operations of the microscope and instant visualization of imaging results.
  • the control software can provide a programming interface which allows users to design and automate their imaging workflow.
  • a set of default workflow scripts can be designated in the scripting language.
  • control system 140 is configured, i.e., by the control software and/or the workflow script, to acquire fluorescence images (also termed simply “collected images” or simply “images”) in loops in the following order (from innermost loop to outermost loop): z-axis, color channel, lateral position, and reagent.
  • fluorescence images also termed simply “collected images” or simply “images”
  • the control system 140 causes the stage 118 to step through multiple vertical positions. Because the vertical position of the stage 118 is controlled by a piezoelectric actuator, the time required to adjust positions is small and each step in this loop is extremely fast.
  • the sample can be sufficiently thick, e.g., a few microns, that multiple image planes through the sample may be desirable. For example, multiple layers of cells can be present, or even within a cell there may be a vertical variation in gene expression. Moreover, for thin samples, the vertical position of the focal plane may not be known in advance, e.g., due to thermal drift. In addition, the sample 10 may vertically drift within the flow cell 110 . Imaging at multiple Z-axis positions can ensure most of the cells in a thick sample are covered, and can help identify the best focal position in a thin sample.
  • the control system 140 causes the light source 122 to step through different wavelengths of excitation light. For example, one of the laser modules is activated, the other laser modules are deactivated, and the emission filter wheel 128 is rotated to bring the appropriate filter into the optical path of the light between the sample 10 and the camera 134 .
  • the control system 140 causes the light source 122 to step through different lateral positions in order to obtain different fields of view (FOVs) of the sample.
  • FOVs fields of view
  • the linear actuators supporting the stage 118 can be driven to shift the stage laterally.
  • the control system 140 number of steps and lateral motion is selected such that the accumulated FOVs to cover the entire sample 10.
  • the lateral motion is selected such that FOVs partially overlap.
  • the control system 140 causes the apparatus 100 to step through multiple different available reagents. For example, at each step of the loop, the control system 140 can control the valve 114 to connect the flow cell 110 to the purge fluid 112 b , cause the pump 116 to draw the purge fluid through the cell for a first period of time to purge the current reagent, then control the valve 114 to connect the flow cell 110 to different new reagent, and then draw the new reagent through the cell for a second period of time sufficient for the probes in the new reagent to bind to the appropriate RNA sequences. Because some time is required to purge the flow cell and for the probes in the new reagent to bind, the time required to adjust reagents in is the longest, as compared to adjusting the lateral position, color channel or z-axis.
  • a data processing system 150 is used to process the images and determine gene expression to generate the spatial transcriptomic data.
  • the data processing system 150 includes a data processing device 152 , e.g., one or more processors controlled by software stored on a computer readable medium, and a local storage device 154 , e.g., non-volatile computer readable media, that receives the images acquired by the camera 134 .
  • the data processing device 152 can be a work station with GPU processors or FPGA boards installed.
  • the data processing system 150 can also be connected through a network to remote storage 156 , e.g., through the Internet to cloud storage.
  • the data processing system 150 performs on-the-fly image processing as the images are received.
  • the data processing device 152 can perform image pre-processing steps, such as filtering and deconvolution, that can be performed on the image data in the storage device 154 but which do not require the entire data set. Because filtering and deconvolution are a major bottleneck in the data processing pipeline, pre-processing as image acquisition is occurring can significantly shorten the offline processing time and thus improve the throughput.
  • FIG. 2 illustrates a flow chart of a method of data processing in which the processing is performed after all of the images have been acquired.
  • the process begins with the system receiving the raw image files and supporting files (step 202 ).
  • the data processing system can receive the full set of raw images from the camera, e.g., an image for each combination of possible values for the z-axis, color channel (excitation wavelength), lateral FOV, and reagent.
  • the data processing system can receive a reference expression file, e.g., a FPKM (fragments per kilobase of sequence per million mapped reads) file, a data schema, and one or more stain images, e.g., DAPI images.
  • a reference expression file e.g., a FPKM (fragments per kilobase of sequence per million mapped reads) file
  • a data schema e.g., DAPI images.
  • stain images e.g., DAPI images.
  • the reference expression file can be used to cross-check between traditional sequence results and the mFISH results.
  • the image files received from the camera can optionally include metadata, the hardware parameter values (such as stage positions, pixel sizes, excitation channels, etc.) at which the image was taken.
  • the data schema provides a rule for ordering the images based on the hardware parameters so that the images are placed into one or more image stacks in the appropriate order. If metadata is not included, the data schema can associate an order of the images with the values for the z-axis, color channel, lateral FOV and reagent used to generate that image.
  • the stain images will be presented to the user with the transcriptomic information overlaid.
  • the collected images can be subjected to one or more quality metrics (step 203 ) before more intensive processing in order to screen out images of insufficient quality. Only images that meet the quality metric(s) are passed on for further processing. This can significantly reduce processing load on the data processing system
  • sharpness quality can be determined for each collected image.
  • sharpness quality can be calculated by summing the gradient in intensity over the image:
  • the threshold can be a preset empirically determined value, or the threshold can be calculated based on the sharpness values for the images with the same FOV. For example, a standard deviation of sharpness quality values can be calculated, the threshold can be set based on the standard deviation.
  • the z-axis position of the focal plane might not be known exactly. Images which are out of focus can be screened out. In some implementations, in each z-axis loop, only a single image is retained, e.g., the image having the best sharpness quality. In some implementation, multiple images from a single each z-axis loop can meet the sharpness quality, and the multiple images are combined, e.g., blended, to form a single image.
  • a brightness quality value can be determined for each collected image.
  • the brightness quality can be used to determine whether any cells are present in the image. For example, the intensity values of all the pixels in the image can be summed and compared to a threshold. If the total is less than the threshold, then this can indicate that there is essentially nothing in the image, i.e., no cells are in the image, and there is no information of interest and the image need not be processed.
  • a standard deviation of the intensity values in the image can be calculated and compared to a threshold. If the standard deviation is below the threshold, this indicates that the image is fairly uniform and may lack bright spots indicating excited probes. Again, such a collected image does not have information of interest and the image need not be processed.
  • the above two quality checks are exemplary and other quality metrics can be used.
  • a quality metric that measures spot quality can be used.
  • sample drift can be detected by a variety of image processing algorithms, e.g., phase-correlation.
  • image processing algorithms e.g., phase-correlation.
  • a collected image can be compared to an image for the same FOV but for a prior reagent. If the images are not sufficiently correlated, e.g., does not meet a threshold correlation, the collected image does not have information of interest and the image need not be processed.
  • each image is processed to remove experimental artifacts (step 204 ). Since each RNA molecule will be hybridized multiple times with probes at different excitation channels, a strict alignment across the multi-channel, multi-round image stack is beneficial for revealing RNA identities over the whole FOV. Removing the experimental artifacts can include field flattening and/or chromatic aberration correction In some implementations, the field flattening is performed before the chromatic aberration correction.
  • Imperfections of the illumination profile e.g., Gaussian or other uneven distributions instead of uniform distribution across the FOV
  • Imperfections of the illumination profile can lead to uneven fluorescent signal level across the FOV, and this spatial unevenness often varies between excitation channels.
  • Such nonuniformity can be passed to from illumination field to fluorescent signal level across the FOV, and consequently bias the spatial transcriptomic analysis results.
  • an illumination profile e.g., test image
  • These measured illumination profiles are then used as channel-specific normalization factors in imaging data.
  • the illumination profile for each color channel (i.e., each emission wavelength to be measured) can be stored as a normalization image, and the intensity value for each pixel in a collected image for a color channel is divided by the intensity value for the corresponding pixel in the normalization image of the associated color channel.
  • Chromatic aberration correction accounts for optical distortions that vary across the different emission channels. That is, because the probes are being measured for different wavelengths, chromatic aberrations in the imaging path, if any, may significantly degrade the downstream analysis results.
  • a software-based solution for chromatic aberration correction can be provided in the data processing pipeline.
  • a test image is generated for each color channel using a target having fluorescent spots in a known pattern, e.g., a rectangular pattern.
  • a transform is calculated that brings the measured spots in the test image back into alignment with the known pattern.
  • RNA image spot sharpening can include applying filters to remove cellular background and/or deconvolution with point spread function to sharpen RNA spots.
  • a low-pass filter is applied to the image, e.g., to the field-flattened and chromatically corrected images to remove cellular background around RNA spots.
  • the filtered images are further de-convolved with a 2-D point spread function (PSF) to sharpen the RNA spots, and convolved with a 2-D Gaussian kernel with half pixel width to slightly smooth the spots.
  • PSF 2-D point spread function
  • the images having the same FOV are registered to align the features, e.g., the cells or cell organelles, therein (step 208 ).
  • features in different rounds of images are aligned, e.g., to sub-pixel precision.
  • sample drifts and stage drifts through hours-long imaging process can transform into image feature shifts, which can undermine the transcriptomic analysis if left unaddressed.
  • the sample may no longer be in the same location in the later image, which can introduce errors into decoding or simply make decoding impossible.
  • fiducial markers e.g., fluorescent beads
  • the sample and the fiducial marker beads will move approximately in unison. These beads can be identified in the image based on their size and shape. Comparison of the positions of the beads permits registration of the two images, e.g., calculation of an affine transformation.
  • MI mutual-information
  • FIG. 3 illustrates a process for using a mutual-information algorithm to align collected images having the same FOV that preserves high precision of the MI registration algorithm while reducing computational load.
  • the set of collected images having the same FOV can be considered an original image stack; the order of images in the stack is set by the data schema discussed above.
  • Each collected image 302 in the image stack may be, for example, 2048 ⁇ 2048 pixels. In general, performing a mutual-information algorithm on an image of this size may require excessive computational power.
  • the collected image 302 is evenly sliced into four image patches 304 a , 304 b , 304 c , 304 d . Assuming each collected image 302 is 2048 ⁇ 2048 pixels, then the image patches are 1024 ⁇ 1024 pixels.
  • Each image patch 304 a , 304 b , 304 c , 304 d is multiplied by a phase factor ⁇ , the phase shifted image patches are overlapped and summed to generate a complex image, and the real and imaginary parts of the complex image are summed up to create a summed image 306 .
  • This summed image 306 has 1 ⁇ 4 size of the original image 302 , e.g., is 1024 ⁇ 1024 pixels.
  • This process is repeated for each original image 302 in the original image stack to generate a stack 308 of images 306 of reduced side.
  • the mutual-information algorithm is then performed using the stack 308 to generate a transform for each reduced image 306 in the stack 306 .
  • the transform can include a translation and/or scaling transform.
  • the transform determined for each reduced image is applied to the corresponding collected image in the original stack. This should result in a stack 312 of collected images 310 of the original size, e.g., 2048 ⁇ 2048 pixels, in which the images 310 for the same FOV are in alignment.
  • This slice-and-overlap strategy preserves most of the features in the image and thus should still have high precision, but reduces the temporary image size for shift calculation and thus has reduced computational load.
  • FIG. 3 and the discussion above focuses on splitting into four patches corresponding to upper right, lower right, upper left and lower left quadrants of the original image
  • another number of patches can be used, e.g., the image can be split into other shapes, e.g., rectangular rather than square, so long as the patches have the same dimensions.
  • the original image could be split into 8 image patches of 256 ⁇ 2048 pixels.
  • the image can be split into a rectangular grid having a plurality of columns and a plurality of rows, e.g., an equal number of columns and rows.
  • step 206 because the filtering and deconvolution processes of step 206 are implemented on individual frames (i.e., bits) independently, they can be conducted prior to the registration of step 208 . Moreover, since image stacks used for registration are background corrected and sharpened to reveal spatial features of RNA spots, the registration precision can be improved. To validate the latter advantage, three data sets from low confluency cell sample, high confluency cell sample, and brain tissue sample were compared for registration precision with registration first vs. filter first approaches. The registration quality was evaluated by summing up the registered image stack across all the layers, then calculating the maximum and the minimum of the summed image. The filter-first approach generally yields a larger maximum and smaller minimum of the flattened image, which indicates an improvement of the registration quality.
  • a registration quality check can be performed after registration. If properly registered, the bright points in each image should overlap so that the total brightness is increased. Referring the stack of unregistered images can be summed, and similarly the stack 312 of registered images 310 (see FIG. 3 ) can be summed, e.g.:
  • the maximum value of I U (x,y) can be compared to the maximum value of I R (x,y). If max(I R (x,y)) > max(I u (x,y)) then registration can be considered a success. On the other hand, if max(I R (x,y)) ⁇ max(I U (x,y)), the data processing system can indicate the registration is a failure and issue an error code. This FOV might be excluded from subsequent analysis and/or new images may need to be obtained.
  • a mask can be calculated for each collected image.
  • the intensity value for each pixel is compared to a threshold value.
  • a corresponding pixel in the mask is set to 1 if the intensity value is above the threshold, and set to 0 if the intensity value is below the threshold.
  • the threshold value can be an empirically determined predetermined value, or can be calculated from the intensity values in the image.
  • the mask can correspond to the location of cells within the sample; spaces between cells should not fluoresce and should have a low intensity.
  • a compressed frame P(x,y) is created from the image sequence for that FOV.
  • the compressed frame P(x,y) can be generated by performing a product operation on the images for that FOV, e.g.:
  • a threshold can be generated based on the histogram of log P(x,y).
  • FIG. 4 illustrates a logarithmic histogram 400 of compressed frame values (P(x,y)) in a collected image.
  • the logarithmic histogram will include an initial sharp increase in intensity (which may be the leading edge of a first local maxima), and a final sharp decrease in intensity (which may be the trailing edge of a second local maxima).
  • the histogram can be subjected to a peak and/or edge-finding algorithm to identify a first peak 402 that is closest to the lower end of the histogram (or an edge of the sharp increase in intensity if there is no clear peak) and a second peak 404 that is closest to the upper end of the histogram (or an edge of the sharp decrease in intensity if there is no clear peak).
  • a threshold 410 can be set based on the intensity values of the two identified points 402 , 404 .
  • the threshold 410 can be an average of the intensity values.
  • the histogram can be subjected to an Otsu threshold method to determine the threshold 410 . A. This technique can be useful because the histogram does not always exhibit two clear peaks.
  • Salt and pepper effects can be removed from the mask.
  • Single “on” pixels which are outside a continuous region of “on” pixels, or single “off” pixels which are inside a continuous region of “on” pixels can be converted to “off” or “on” to match the surrounding area.
  • clusters of adjacent pixels in the mask below a threshold size e.g., 50 pixels, can be indicated as an error and be changed to match the surrounding area.
  • FIG. 5 A A collected image is illustrated in FIG. 5 A
  • FIG. 5 B illustrates a mask resulting from the image of FIG. 5 A .
  • spatial transcriptomic analysis can be performed (step 210 ).
  • FOV normalization can be performed before the spatial transcriptomic analysis in order to make the histogram more consistent.
  • the FOV normalization occurs after registration.
  • FOV normalization can occur before registration.
  • FOV normalization could be considered part of the filtering.
  • a problem in the collected image, even after filtering and deconvolution discussed above, is that there can be a significant variation in brightness in the brightest pixels, e.g., over several orders of magnitude. For example, referring to FIG. 6 A , the histogram 602 of intensity values in a collected image has a very long tail 604 . This can introduce uncertainty in the decoding process. FOV normalization can counteract this source of error.
  • all the intensity values in the image are normalized relative to the maximum intensity value in the image. For example, the maximum intensity value is determined, and all intensity values are divided by the maximum so that intensity values vary between 0 and I MAX , e.g., 1.
  • the intensity values in the image are analyzed to determine an upper quantile that includes the highest intensity values, e.g., the 99% and higher quantile (i.e., upper 1%), 99.5% and higher quantile, or 99.9% and higher quantile, or 99.95% and higher quantile (i.e., upper 0.05%).
  • the cutoff for this quantile is indicated by a broken line 610 in FIG. 6 A .
  • the intensity value at this quantile limit e.g., the intensity value of log(-2) at the 99.95% quantile in the example of FIG. 6 A , can be determined and stored. All pixels having intensity values within the upper quantile are reset to have the maximum intensity value, e.g., 1.
  • FIG. 6 B illustrates the histogram 620 of intensity values resulting from the FOV normalization.
  • the aligned images for a particular FOV can be considered as a stack 312 that includes multiple image layers 310 , with each image layer being X by Y pixels, e.g., 2048 ⁇ 2048 pixels.
  • this image stack can be evaluated as a 2-D matrix 702 of pixel words.
  • Each row 704 corresponds to one of the pixels (the same pixel across the multiple image in the stack), the values from the row 704 provide a pixel word 710 .
  • Each column 706 provides one of the values in the word 710 , i.e., the intensity value from the image layer for that pixel.
  • the values can be normalized, e.g., vary between 0 and I MAX . Different intensity values are represented in FIG. 7 as different degrees of shading of the respective cells.
  • the data processing system 150 stores a code book 722 that is used to decode the image data to identify the gene expressed at the particular pixel.
  • the code book 722 includes multiple reference code words, each reference code word associated with a particular gene. As shown in FIG. 7 , the code book 722 can be represented as a 2D matrix with G rows 724 , where G is the number of code words, e.g., the number of genes (although the same gene could be represented by multiple code words), and B columns 726 . Each row 724 corresponds to one of the reference code words 730 , and each column 706 provides one of the values in the reference code word 730 , as established by prior calibration and testing of known genes.
  • the values in the reference code 730 can be binary, i.e., “on” or “off”.
  • each value can be either 0 or I MAX , e.g., 1.
  • the on and off values are represented in FIG. 7 by light and dark shading of respective cells.
  • a distance d(p,i) is calculated between the pixel word 710 and each reference code word 730 .
  • the distance between the pixel word 710 and reference code word 730 can be calculated as a Euclidean distance, e.g., a sum of squared differences between each value in the pixel word and the corresponding value in the reference code word. This calculation can be expressed as:
  • I p,x are the values from the matrix 702 of pixel words and C i,x are the values from the matrix 722 of reference code words.
  • Other metrics e.g., sum of absolute value of differences, cosine angle, correlation, etc., can be used instead of a Euclidean distance.
  • the data processing apparatus determines min (d(p,1), d(p,2), ... d(p,B)), and determines the value b as the value for i (between 1 and B) that provided the minimum.
  • the gene corresponding to that best matching code word is determined, e.g., from a lookup table that associates code words with genes, and the pixel is tagged as expressing the gene.
  • the data processing apparatus can filter out false callouts.
  • One technique to filter out false callouts is to discard tags where the distance value d(p,b) that indicated expression of a gene is greater than a threshold value, e.g., if d(p,b) > D1 MAX .
  • bit ratio is calculated as the mean of the intensity values from the image word for layers that are supposed to be on (as determined from the code word), divided by the mean of the intensity values from the image word for layers that are supposed to be off (again as determined from the code word). For example, for the code word 730 shown in FIG. 7 , the bits corresponding to columns 3-6 and 12-14 are on, and the bits corresponding to columns 1-2 and 7-11 are off. Thus, assuming this code word is selected as the best match for a pixel p, the bit ratio BR can be calculated as
  • BR I p , 3 + I p , 4 + I p , 5 + I p , 6 + I p , 12 + I p , 13 + I p , 14 / 7 I p , 1 + I p , 2 + I p , 7 + I p , 8 + I p , 9 + I p , 10 + I p , 11 / 7 .
  • the bit ratio BR is compared to a threshold value TH BR .
  • the threshold value TH BR is determined empirically from prior measurements. However, in some implementations, the threshold value TH BR can be calculated automatically for a particular code word based on the measurements obtained from the sample.
  • FIG. 8 A illustrates a logarithmic histogram 802 of bit ratios intensity values for all pixels that were tagged with a particular code word.
  • a threshold value 804 is calculated based on this histogram 802 . For example, a default quantile, e.g., 50%, can be used to calculate an initial threshold value 802 . However, the quantile can be adjusted during optimization discussed below.
  • bit brightness BB is calculated as the mean of the intensity values from the image word for layers that are supposed to be on (as determined from the code word). For example, for the code word 730 shown in FIG. 7 , the bits corresponding to columns 3-6 and 12-14 are on. Thus, assuming this code word is selected as the best match for a pixel p, the bit brightness BB can be calculated as
  • BB I p,3 +I p,4 +I p,5 +I p,6 +I p,12 +I p,13 +I p,14 / 7
  • the bit brightness BB is compared to a threshold value TH BB .
  • the threshold value TH BB is determined empirically from prior measurements. However, in some implementations, the threshold value TH BB can be calculated automatically for a particular code word based on the measurements obtained from the sample.
  • FIG. 8 B illustrates a logarithmic histogram 812 of bit brightness values for all pixels that were tagged with a particular code word.
  • a threshold value 814 is calculated based on this histogram 812 . For example, a default quantile, e.g., 80% or 90%, can be used to calculate an initial threshold value 814 . However, the quantile can be adjusted during optimization discussed below.
  • the data processing apparatus can now perform optimization and re-decoding (step 212 ).
  • the optimization can include machine-learning based optimization of the decoding parameters, followed by returning to step 210 with updated decoding parameters in order to update the spatial transcriptomic analysis. This cycle can be repeated until the decoding parameters have stabilized.
  • the optimization of the decoding parameters will use a merit function, e.g., a FPKM/TPM correlation, spatial correlation, or confidence ratio.
  • Parameters that can be included as variables in the merit function include the shape (e.g., start and end of frequency range, etc.) of the filters used to remove cellular background, the numerical aperture value for the point spread function used to sharpen the RNA spots, the quantile boundary Q used in normalization of the FOV, the bit ratio threshold TH BR , the bit brightness threshold TH BB (or the quantiles used to determine the bit ratio threshold TH BR and bit brightness threshold TH BB ), and/or the maximum distance D1 max at which at which a pixel word can be considered to match a code word.
  • This merit function may be an effectively discontinuous function, so a conventional gradient following algorithm may be insufficient to identify the optimal parameter values.
  • a machine learning model can be used to converge on parameter values.
  • the data processing apparatus can perform unification of the parameter values across all FOVs. Because each FOV is processed individually, each field can experience different normalization, thresholding, and filtering setting. As a result, a high contrast image can result in a histogram with variation that causes false positive callouts in quiet areas. The result of unification is that all FOVs use the same parameter values. This can significantly remove callouts from background noise in quiet area, and can provide a clear and unbiased spatial pattern in large sample area.
  • a variety of approaches are possible to select a parameter value that will be used across all FOVs.
  • One option is to simply pick a predetermined FOV, e.g., the first measured FOV or a FOV near the center of the sample, and use the parameter value for that predetermined FOV.
  • Another option is to average the values for the parameter across multiple FOV and then use the averaged value.
  • Another option is to determine which FOV resulted in the best fit between its pixel words and tagged code words. For example, a FOV with the smallest average distance d(p,b1) between the tagged code words and the pixel words for those code words can be determined and then selected.
  • unification can be performed before the spatial transcriptomic analysis, e.g., between steps 208 and 210 .
  • the unification would be performed for parameters selected before step 210 , e.g., parameters used in steps 204 or 206 , but not for the parameters in steps 210 and 212 .
  • the data processing apparatus can now perform stitching and segmentation (step 214 ).
  • Stitching combines multiple FOV into a single image.
  • Stitching can be performed using a variety of techniques. Referring to FIG. 9 , one approach is, for each row of FOV that together will form the combined image of the sample and each FOV within the row, determine a horizontal shift for each FOV. Once the horizontal shifting is calculated, a vertical shift is calculated for each row of FOV. The horizontal and vertical shifts can be calculated based on cross-correlation, e.g., phase correlation. With the horizontal and vertical shift for each FOV, a single combined image can be generated, and gene coordinates can be transferred to the combined image based on the horizontal and vertical shift.
  • cross-correlation e.g., phase correlation
  • An indication that a gene is expressed at a certain coordinate in the combined fluorescence image can be added, e.g., as metadata. This indication can be termed a “callout.”
  • the stain images e.g., the DAPI images
  • the stain images can be stitched together using the technique discussed with respect to FIG. 9 to generate a combined stain image.
  • the stain image can be registered to the collected fluorescent image(s).
  • An indication that a gene is expressed at a certain coordinate in the combined stain image (as determined from the coordinate in the FOV and the horizontal and vertical shift for that FOV) can be added, e.g., as metadata, to provide a callout.
  • some genes may be double-counted in the overlapping area.
  • a distance e.g., Euclidean distance
  • One of the callouts can be removed if the distance is below a threshold value.
  • More complex techniques can be used if a cluster of pixels are tagged as expressing a gene.
  • Segmentation of the combined image e.g., the image of the stained cell
  • regions corresponding to cells can be performed using various known techniques. Segmentation is typically performed after stitching of the images, but can occur before or after callouts are added to the combined image.
  • the segmented image with callouts indicating positions of gene expression can now be stored and presented to a user, e.g., on a visual display, for analysis.
  • one problem is that the data processing, including the image processing steps, is computationally demanding and can require a significant amount of time.
  • One issue is that in conventional techniques, all images are acquired before the image processing commences. However, it is possible to perform on-the-fly image processing as the images are received.
  • FIG. 10 illustrates a modified data processing method in which image processing is performed on-the-fly. Steps that are not specifically discussed can be performed in a manner similar to the method of FIG. 2 .
  • the data processing system can receive the supporting files (step 202 a ), although this step could be postponed until after image acquisition. It may be necessary for some files, e.g., the data schema, to be loaded to the data processing system for the image processing to be performed.
  • the apparatus begins the process of acquiring new images (step 202 b ), e.g., stepping through the loops to adjust the z-axis, color channel, lateral position, and reagent, and acquiring an image at each combination of possible values for these parameters.
  • a quality control (QC) test is performed to determine whether the image is of sufficient quality to be pre-processed (step 203 ). If it is, on-the-fly pre-processing of the acquired image is performed (step 252 ).
  • “on-the-fly” indicates that the processing of one image can be performed in parallel with acquisition of subsequent images.
  • the on-the-fly pre-processing can include the removing the experimental artifacts (step 204 ) and RNA image spot sharpening (step 206 ). It may be noted that since these image preprocessing steps are performed individually for each image, and do not depend on other images in the same FOV or images for other FOV, shifting processing in this manner does not adversely impact the quality of processing. In general, significant computational load can be shifted to times when the reagent is being changed. As a result, overall time from loading of the sample into the flow cell to receiving the segmented image with callouts is significantly reduced, which can significantly improve throughput and speed research.
  • the images are pre-processed in “real time,” i.e., processing of one image is performed in parallel with collection of a subsequent image by the apparatus and completed before a subsequent image is received at the data processing system.
  • Embodiments of the subject matter and the functional operations described in this specification can be implemented in digital electronic circuitry, in tangibly-embodied computer software or firmware, in computer hardware, including the structures disclosed in this specification and their structural equivalents, or in combinations of one or more of them.
  • Embodiments of the subject matter described in this specification can be implemented as one or more computer programs, i.e., one or more modules of computer program instructions encoded on a tangible non-transitory storage medium for execution by, or to control the operation of, data processing apparatus.
  • the computer storage medium can be a machine-readable storage device, a machine-readable storage substrate, a random or serial access memory device, or a combination of one or more of them.
  • the program instructions can be encoded on an artificially-generated propagated signal, e.g., a machine-generated electrical, optical, or electromagnetic signal, that is generated to encode information for transmission to suitable receiver apparatus for execution by a data processing apparatus.
  • data processing apparatus refers to data processing hardware and encompasses all kinds of apparatus, devices, and machines for processing data, including by way of example a programmable processor, a computer, or multiple processors or computers.
  • the apparatus can also be, or further include, special purpose logic circuitry, e.g., an FPGA (field programmable gate array) or an ASIC (application-specific integrated circuit).
  • the apparatus can optionally include, in addition to hardware, code that creates an execution environment for computer programs, e.g., code that constitutes processor firmware, a protocol stack, a database management system, an operating system, or a combination of one or more of them.
  • a computer program which may also be referred to or described as a program, software, a software application, an app, a module, a software module, a script, or code, can be written in any form of programming language, including compiled or interpreted languages, or declarative or procedural languages; and it can be deployed in any form, including as a stand-alone program or as a module, component, subroutine, or other unit suitable for use in a computing environment.
  • a program may, but need not, correspond to a file in a file system.
  • a program can be stored in a portion of a file that holds other programs or data, e.g., one or more scripts stored in a markup language document, in a single file dedicated to the program in question, or in multiple coordinated files, e.g., files that store one or more modules, sub-programs, or portions of code.
  • a computer program can be deployed to be executed on one computer or on multiple computers that are located at one site or distributed across multiple sites and interconnected by a data communication network.
  • the processes and logic flows described in this specification can be performed by one or more programmable computers executing one or more computer programs to perform functions by operating on input data and generating output.
  • the processes and logic flows can also be performed by special purpose logic circuitry, e.g., an FPGA or an ASIC, or by a combination of special purpose logic circuitry and one or more programmed computers.
  • Computers suitable for the execution of a computer program can be based on general or special purpose microprocessors or both, or any other kind of central processing unit.
  • a central processing unit will receive instructions and data from a read-only memory or a random access memory or both.
  • the essential elements of a computer are a central processing unit for performing or executing instructions and one or more memory devices for storing instructions and data.
  • the central processing unit and the memory can be supplemented by, or incorporated in, special purpose logic circuitry.
  • a computer will also include, or be operatively coupled to receive data from or transfer data to, or both, one or more mass storage devices for storing data, e.g., magnetic, magneto-optical disks, or optical disks. However, a computer need not have such devices.
  • Computer-readable media suitable for storing computer program instructions and data include all forms of non-volatile memory, media and memory devices, including by way of example semiconductor memory devices, e.g., EPROM, EEPROM, and flash memory devices; magnetic disks, e.g., internal hard disks or removable disks; magneto-optical disks; and CD-ROM and DVD-ROM disks.
  • semiconductor memory devices e.g., EPROM, EEPROM, and flash memory devices
  • magnetic disks e.g., internal hard disks or removable disks
  • magneto-optical disks e.g., CD-ROM and DVD-ROM disks.
  • embodiments of the subject matter described in this specification can be implemented on a computer having a display device, e.g., a CRT (cathode ray tube) or LCD (liquid crystal display) monitor, for displaying information to the user and a keyboard and a pointing device, e.g., a mouse or a trackball, by which the user can provide input to the computer.
  • a display device e.g., a CRT (cathode ray tube) or LCD (liquid crystal display) monitor
  • keyboard and a pointing device e.g., a mouse or a trackball
  • Other kinds of devices can be used to provide for interaction with a user as well; for example, feedback provided to the user can be any form of sensory feedback, e.g., visual feedback, auditory feedback, or tactile feedback; and input from the user can be received in any form, including acoustic, speech, or tactile input.
  • a computer can interact with a user by sending documents to and receiving documents from a device that is used by the user; for example, by sending web pages to a web browser on a user’s device in response to requests received from the web browser.
  • a computer can interact with a user by sending text messages or other forms of message to a personal device, e.g., a smartphone that is running a messaging application, and receiving responsive messages from the user in return.
  • Data processing apparatus for implementing machine learning models can also include, for example, special-purpose hardware accelerator units for processing common and compute-intensive parts of machine learning training or production, i.e., inference, workloads.
  • Machine learning models can be implemented and deployed using a machine learning framework, e.g., a TensorFlow framework, a Microsoft Cognitive Toolkit framework, an Apache Singa framework, or an Apache MXNet framework.
  • a machine learning framework e.g., a TensorFlow framework, a Microsoft Cognitive Toolkit framework, an Apache Singa framework, or an Apache MXNet framework.
  • Embodiments of the subject matter described in this specification can be implemented in a computing system that includes a back-end component, e.g., as a data server, or that includes a middleware component, e.g., an application server, or that includes a front-end component, e.g., a client computer having a graphical user interface, a web browser, or an app through which a user can interact with an implementation of the subject matter described in this specification, or any combination of one or more such back-end, middleware, or front-end components.
  • the components of the system can be interconnected by any form or medium of digital data communication, e.g., a communication network. Examples of communication networks include a local area network (LAN) and a wide area network (WAN), e.g., the Internet.
  • LAN local area network
  • WAN wide area network

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Health & Medical Sciences (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Chemical & Material Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Multimedia (AREA)
  • Analytical Chemistry (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Immunology (AREA)
  • Optics & Photonics (AREA)
  • Medical Informatics (AREA)
  • Biochemistry (AREA)
  • Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
  • Molecular Biology (AREA)
  • Spectroscopy & Molecular Physics (AREA)
  • Pathology (AREA)
  • Biophysics (AREA)
  • Bioethics (AREA)
  • Biotechnology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Databases & Information Systems (AREA)
  • Organic Chemistry (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Evolutionary Biology (AREA)
  • Artificial Intelligence (AREA)
  • Evolutionary Computation (AREA)
  • Software Systems (AREA)
  • Biomedical Technology (AREA)
  • Zoology (AREA)
  • Proteomics, Peptides & Aminoacids (AREA)
  • Genetics & Genomics (AREA)
  • Wood Science & Technology (AREA)
  • Quality & Reliability (AREA)
  • Chemical Kinetics & Catalysis (AREA)
  • Signal Processing (AREA)
  • Public Health (AREA)
  • Data Mining & Analysis (AREA)

Abstract

A fluorescent in-situ hybridization imaging system includes a flow cell to contain a sample, a fluorescence microscope, and a control system. The fluorescence microscope includes a variable frequency excitation light source to illuminate the sample, a plurality of emission bandpass filters on a filter wheel, an actuator to rotate the filter wheel, and a camera positioned to receive fluorescently emitted light from the sample. The control system is configured to cause the variable frequency excitation light source to emit a light beam having a selected wavelength, cause the actuator to rotate the filter wheel to position a selected filter in a light path between the sample and the camera, obtain an image from the camera, and coordinate the variable frequency excitation light source and filter wheel such that the selected filter has an emission bandpass associated with emission by the fluorescent probes when excited by the selected wavelength.

Description

    CROSS-REFERENCE TO RELATED APPLICATIONS
  • This application is a continuation of U.S. Pat. Application Serial No. 17/124,285, filed Dec. 16, 2020, which claims priority to U.S. Provisional Pat. Application Serial No. 62/949,391, filed on Dec. 17, 2019, the disclosures of which are incorporated by reference.
  • TECHNICAL FIELD
  • This specification relates multiplexed fluorescence in-situ hybridization imaging.
  • BACKGROUND
  • It is of great interest to the biotech community and pharmaceutical industry to develop methods for visualizing and quantifying multiple biological analytes - e.g., DNA, RNA, and protein – within a biological sample – e.g., tissue resection, biopsy, cells grown in culture. Scientists use such methods to diagnose/monitor disease, validate biomarkers, and investigate treatment. To date, exemplary methods include multiplex imaging of antibodies or oligonucleotides (e.g., RNA or DNA) labeled with a functional domain to a biological sample.
  • Multiplexed fluorescence in-situ hybridization (mFISH) imaging is a powerful technique to determine gene expression in spatial transcriptomics. In brief, a sample is exposed to multiple oligonucleotide probes that target RNA of interest. These probes have different labeling schemes that will allow one to distinguish different RNA species when the complementary, fluorescent labeled probes are introduced to the sample. Then the sequential rounds of fluorescence images are acquired with exposure to excitation light of different wavelengths. For each given pixel, its fluorescence intensities from the different images for the different wavelengths of excitation light form a signal sequence. This sequence is then compared to a library of reference codes from a codebook that associates each code with a gene. The best matching reference code is used to identify an associated gene that is expressed at that pixel in the image.
  • SUMMARY
  • In one aspect, a fluorescent in-situ hybridization imaging system includes a flow cell to contain a sample to be exposed to fluorescent probes in a reagent, a valve to control flow from one of a plurality of reagent sources the flow cell, a pump to cause fluid flow through the flow cell, a fluorescence microscope including a variable frequency excitation light source and a camera positioned to receive fluorescently emitted light from the sample, an actuator to cause relative vertical motion between the flow cell and the fluorescence microscope, a motor to cause to cause relative lateral motion between the flow cell and the fluorescence microscope, and a control system. The control system is configured to, as nested loops, cause the valve to sequentially couple the flow cell to a plurality of different reagent sources to expose the sample to a plurality of different reagents, for each reagent of the plurality of different reagents, cause the motor to sequentially position the fluorescence microscope relative to sample at a plurality of different fields of view, for each field of view of the plurality of different fields of view, cause the variable frequency excitation light source to sequentially emit a plurality of different wavelengths, for each wavelength of the plurality of different wavelengths, cause the actuator to sequentially position the fluorescence microscope relative to sample at a plurality of different vertical heights, and for each vertical height of the plurality of different vertical heights, obtain an image at the respective vertical height covering the respective field of view of the sample having respective fluorescent probes of the respective regent as excited by the respective wavelength.
  • In another aspect, a fluorescent in-situ hybridization imaging and analysis system includes a flow cell to contain a sample to be exposed to fluorescent probes in a reagent, a fluorescence microscope to obtain sequentially collect a plurality of images of the sample at a plurality of different combinations of imaging parameters, and a data processing system. The data processing system includes an online pre-processing system configured to sequentially receive the images from the fluorescence microscope as the images are collected and perform on-the-fly image pre-processing to remove experimental artifacts of the image and to provide RNA image spot sharpening, and an offline processing system configured to, after the plurality of images are collected, perform registration of images having a same field of view and to decode intensity values in the plurality of images to identify expressed genes.
  • In another aspect, a fluorescent in-situ hybridization imaging system includes a flow cell to contain a sample to be exposed to fluorescent probes in a reagent, a fluorescence microscope, and a control system. The fluorescence microscope includes a variable frequency excitation light source to illuminate the sample, a plurality of emission bandpass filters on a filter wheel, an actuator to rotate the filter wheel, and a camera positioned to receive fluorescently emitted light from the sample that has passed through a filter on the filter wheel. The control system is configured to cause the variable frequency excitation light source to emit a light beam having a selected wavelength, cause the actuator to rotate the filter wheel to position a selected filter in a light path between the sample and the camera, obtain an image from the camera, and coordinate the variable frequency excitation light source and filter wheel such that the selected filter has an emission bandpass associated with emission by the fluorescent probes when excited by the selected wavelength when the image is obtained.
  • In another aspect, a computer program product has instructions to cause one or more processors to apply one or more filters to and perform deconvolution on each image of the plurality of images to generate a plurality of filtered images, perform registration of the plurality of filtered images to generate a plurality of registered and filtered images, for each pixel of at least two pixels of the plurality of registered and filtered images decode the pixel by identifying a code word from a plurality of code words in a code book that provides a best match to data values in the plurality of registered and filtered images for the pixel, and for each pixel of at least two pixels, determine a gene associated with the code word and store an indication that the gene is expressed at the pixel.
  • In another aspect, a computer program product has instructions to receive a plurality of initial images of a sample, the plurality of initial images sharing a common field of view. For each initial image of the plurality of initial images, the initial image is split into a plurality of sub-images of equal size, a combined image of size equal to that of a sub-image is calculated from the plurality of sub-images to thereby generate a plurality of reduced-size images with each image of the plurality of images having a corresponding reduced-size image, for at least one reduced-size images of the plurality of reduced-size images a transform is determined to register the plurality of reduced-size images, for each of the at least one reduced-size images the transform determined for the reduced-size image is applied to the corresponding initial image to generate a transformed image of size equal to the initial image and thereby generate a plurality of transformed images, and for each pixel of at least two pixels of the plurality of transformed images the pixel is decoded by identifying a code word from a plurality of code words in a code book that provides a best match to data values in the plurality of transformed images for the pixel. For each pixel of at least two pixels, a gene associated with the code word is determined and an indication that the gene is expressed at the pixel is stored.
  • In another aspect, a computer program product has instructions to receive a plurality of images representing a common field of view of a sample, perform registration of the plurality of images to generate a plurality of registered images, for each pixel of a plurality of pixels of the registered images decode the pixel by identifying a code word from a plurality of code words in a code book that provides a best match to data values in the plurality of registered images for the pixel, and for each code word identified as a best match for one or more pixels and each pixel of the one or more pixels determine whether a bit ratio for an image word for the pixel meets a threshold for the code word. The image word is formed from the data values in the plurality of registered images for the pixel. For at least one pixel that is determined to meet the threshold a gene associated with the code word is determined and an indication is stored that the gene is expressed at the pixel. Pixels for which the bit ratio does not meet the threshold are screened.
  • In another aspect, a computer program product has instructions to receive a plurality of images representing a common field of view of a sample, perform registration of the plurality of images to generate a plurality of registered images, for each pixel of a plurality of pixels of the registered images decode the pixel by identifying a code word from a plurality of code words in a code book that provides a best match to data values in the plurality of registered images for the pixel, and for each code word identified as a best match for one or more pixels and each pixel of the one or more pixels whether a bit brightness for an image word for the pixel meets a threshold for the code word is determined. The image word is formed from the data values in the plurality of registered images for the pixel. For each pixel that is determined to meet the threshold a gene associated with the code word is determined and an indication is stored that the gene is expressed at the pixel. Pixels for which the bit ratio does not meet the threshold are screened.
  • In another aspect, a computer program product has instructions to receive a plurality of initial images of a sample. The plurality of initial images including a multiplicity of images for each field of view of a plurality of different fields of view. Instructions are stored to perform an image processing algorithm in which one or more of image filtering of the multiplicity of images for the field of view or registration of the multiplicity of images for the field of view. For each field of view, one or more initial parameter values are determined, wherein different fields of view have different initial parameter values. The image processing algorithm is performed for each field of view using the different initial decoding parameter values for the different fields of view. One or more subsequent decoding parameter values are determined, and the image processing algorithm and a decoding algorithm are performed for each field of view using the same subsequent parameter values for the different fields of view to generate a set of spatial transcriptomic data.
  • In another aspect, a computer program product has instructions to receive a plurality of initial images of a sample, the plurality of initial images including a multiplicity of images for each field of view of a plurality of different fields of view. For each field of view, a set of spatial transcriptomic data is generated based on the multiplicity of images for the field of view. Each set of spatial transcriptomic data includes one or more data pairs, each data pair of the one or more data pairs including a position and an expressed gene for that position. Regions of overlap of multiplicities of images from adjacent fields of view are determined. From the set of spatial transcriptomic data for each field of view, a combined set of spatial transcriptomic data is generated for a combined image covering the plurality of different fields of view. In generating the combined set of spatial transcriptomic data, at least one double-counted data pair from different fields of view that represents the same expressed gene at the same location within the sample is screened.
  • In another aspect, a computer program product has instructions to receive a plurality of images of a sample, the plurality of images including a multiplicity of images for each field of view of a plurality of different fields of view. On each image of the plurality of images, an image quality check is performed on the image. After performing the image quality check, artifacts are removed from each image of the plurality of images. After removing artifacts, image spots are sharpened in each image of the plurality of images. After sharpening the image spots, for each multiplicity of images, registration of the multiplicity of images is performed to generate a plurality of registered images. After performing registration, for each multiplicity of images, and for each pixel of a plurality of pixels of the registered images of a respective multiplicity, the pixel is decoded by identifying a code word from a plurality of code words in a code book that provides a best match to data values in the respective multiplicity of registered images for the pixel.
  • Advantages of implementations can include, but are not limited to, one or more of the following.
  • The acquisition of multiplexed fluorescence in-situ hybridization (mFISH) images and subsequent processing to determine gene expression can be automated in a manner suitable for industrial adoption. Large sample sizes that require multiple fields-of-view (FOVs) can be processed with uniform image quality. Run-to-run consistency in multi-sample or multi-modal research can be improved. Fluorescence images of the same FOV captured at different times can be registered without using separate fiducials, e.g., without requiring fiducial beads, and using fewer computational resources and more accurately than some conventional techniques.
  • The details of one or more embodiments are set forth in the accompanying drawings and the description below. Other features, aspects, and advantages will become apparent from the description, the drawings, and the claims.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • FIG. 1 is a schematic diagram of an apparatus for multiplexed fluorescence in-situ hybridization imaging.
  • FIG. 2 is a flow chart of a method of data processing.
  • FIG. 3 schematically illustrates an image registration process that uses a mutual-information algorithm.
  • FIG. 4 is a logarithmic histogram of intensities in an image.
  • FIG. 5A is a collected image.
  • FIG. 5B is illustrates a mask generated from the image of FIG. 5A.
  • FIG. 6A is a logarithmic histogram of intensities from a filtered and de-convoluted image.
  • FIG. 6B is a logarithmic histogram of intensities after normalization.
  • FIG. 7 illustrates a method of decoding.
  • FIG. 8A is a logarithmic histogram of bit ratios for pixels identified for a sample code word.
  • FIG. 8B is a logarithmic histogram of brightness for pixels identified for a sample code word.
  • FIG. 9 is a flow chart of a method of stitching.
  • FIG. 10 is a flow chart of another implementation of a method of data processing in which image processing is performed on-the-fly.
  • Like reference numbers and designations in the various drawings indicate like elements.
  • DETAILED DESCRIPTION
  • Although multiplexed fluorescence in-situ hybridization (mFISH) imaging is a powerful technique, it has so far been used primarily in an academic context. Use of the technique for spatial transcriptomics in an industry context, e.g., by a biotechnology company performing research using mFISH for spatial transcriptomic to develop pharmaceuticals intended for productization, poses a variety of issues. For example:
    • The system should be fully or nearly fully automated so that a user, e.g., a research scientist who is familiar with biology but not with image processing or software programming techniques, need merely enter parameters relating to the biological aspects being investigated, and need not make decisions relating to image processing or have programming expertise. Such automation is not, in fact, trivial, and contains a variety of pitfalls.
    • The current state-of-art takes sample images from generic optical microscopes with small image sizes that are suitable for small samples. The samples utilized in the industrial context can be larger and may require multiple fields of view to completely image. Such multiple FOVs can be stitched manually, but semi-automatic approaches suffer from possible issues such as poor image registration, over-counting of expressed genes, or lack of reliability due to inconsistency in processing of different regions of the stitched image.
    • The process needs a sufficiently high throughput to be industrially feasible. However, current state-of-the-art takes considerable imaging time and requires many manual steps in data analysis. A higher imaging throughput and quality control is needed for large-scale life science and clinical research. Moreover, the scalability of data processing is limited by computation time.
    • Because each FOV is processed individually, significant variation across a region of interest can be found. In addition, images from different instruments need to be well-calibrated and converted before being sent to a software pipeline. Conventionally, the image processing parameters need to be adjusted manually to accommodate system or sample variation.
    • Current solutions have a strong assumption of perfect images, which creates a gap between experimental condition and the calibration method.
  • An overall architecture and a variety of individual components and steps are discussed below that can address one or more of the above issues.
  • In general, the overall architecture and individual components and steps enable an end-to-end mFISH imaging and decoding process that involves minimal manual steps. The system integrates fully automatic hardware architecture and on-the-fly image processing. The methods include feedback control, image calibration, and global optimization. The end-to-end process time can be reduced and throughput per sample can be increased. The apparatus can be easy to use by a research scientist in an industry context. Control metrics can ensure data quality before downstream analysis; the early detection reduces waste of samples. The algorithm can enable scalability to large imaging areas with uniform and consistent spatial gene expression. Data accuracy and repeatability for large-scale study can be improved; the results can be easily interpreted and further analyzed in software packages.
  • Referring to FIG. 1 , a multiplexed fluorescent in-situ hybridization (mFISH) imaging and image processing apparatus 100 includes a flow cell 110 to hold a sample 10, a fluorescence microscope 120 to obtain images of the sample 10, and a control system 140 to control operation of the various components of the mFISH imaging and image processing apparatus 100. The control system 140 can include a computer 142, e.g., having a memory, processor, etc., that executes control software.
  • The fluorescence microscope 120 includes an excitation light source 122 that can generate excitation light 130 of multiple different wavelengths. In particular, the excitation light source 122 can generate narrow-bandwidth light beams having different wavelengths at different times. For example, the excitation light source 122 can be provided by a multi-wavelength continuous wave laser system, e.g., multiple laser modules 122 a that can be independently activated to generate laser beams of different wavelengths. Output from the laser modules 122 a can be multiplexed into a common light beam path.
  • The fluorescence microscope 120 includes a microscope body 124 that includes the various optical components to direct the excitation light from the light source 122 to the flow cell 110. For example, excitation light from the light source 122 can be coupled into a multimode fiber, refocused and expanded by a set of lenses, then directed into the sample by a core imaging component, such as a high numerical aperture (NA) objective lens 136. When the excitation channel needs to be switched, one of the multiple laser modules 122 a can be deactivated and another laser module 122 a can be activated, with synchronization among the devices accomplished by one or more microcontrollers 144, 146.
  • The objective lens 136, or the entire microscope body 124, can be installed on vertically movable mount coupled to a Z-drive actuator. Adjustment of the Z-position, e.g., by a microcontroller 146 controlling the Z-drive actuator, can enable fine tuning of focal position. Alternatively, or in addition, the flow cell 110 (or a stage 118 supporting the sample in the flow cell 110) could be vertically movable by a Z-drive actuator 118 b, e.g., an axial piezo stage. Such a piezo stage can permit precise and swift multi-plane image acquisition.
  • The sample 10 to be imaged is positioned in the flow cell 110. The flow cell 110 can be a chamber with cross-sectional area (parallel to the object or image plane of the microscope) with and area of about 2 cm by 2 cm. The sample 10 can be supported on a stage 118 within the flow cell, and the stage (or the entire flow cell) can be laterally movable, e.g., by a pair of linear actuators 118 a to permit XY motion. This permits acquisition of images of the sample 10 in different laterally offset fields of view (FOVs). Alternatively, the microscope body 124 could be carried on a laterally movable stage.
  • An entrance to the flow cell 110 is connected to a set of hybridization reagents sources 112. A multi-valve positioner 114 can be controlled by the controller 140 to switch between sources to select which reagent 112 a is supplied to the flow cell 110. Each reagent includes a different set of one or more oligonucleotide probes. Each probe targets a different RNA sequence of interest, and has a different set of one or more fluorescent materials, e.g., phosphors, that are excited by different combinations of wavelengths. In addition to the reagents 112 a, there can be a source of a purge fluid 112 b, e.g., DI water.
  • An exit to the flow cell 110 is connected to a pump 116, e.g., a peristaltic pump, which is also controlled by the controller 140 to control flow of liquid, e.g., the reagent or purge fluid, through the flow cell 110. Used solution from the flow cell 110 can be passed by the pump 116 to a chemical waste management subsystem 119.
  • In operation, the controller 140 causes the light source 122 to emit the excitation light 130, which causes fluorescence of fluorescent material in the sample 10, e.g., fluorescence of the probes that are bound to RNA in the sample and that are excited by the wavelength of the excitation light. The emitted fluorescent light 132, as well as back propagating excitation light, e.g., excitation light scattered from the sample, stage, etc., is collected by an objective lens 136 of the microscope body 124.
  • The collected light can be filtered by a multi-band dichroic mirror 138 in the microscope body 124 to separate the emitted fluorescent light from the back propagating illumination light, and the emitted fluorescent light is passed to a camera 134. The multi-band dichroic mirror 138 can include a pass band for each emission wavelength expected from the probes under the variety of excitation wavelengths. Use of a single multi-band dichroic mirror (as compared to multiple dichroic mirrors or a movable dichroic mirror) can provide improved system stability.
  • The camera 134 can be a high resolution (e.g., 2048×2048 pixel) CMOS (e.g., a scientific CMOS) camera, and can be installed at the immediate image plane of the objective. Other camera types, e.g., CCD, may be possible. When triggered by a signal, e.g., from a microcontroller, image data from the camera can be captured, e.g., sent to an image processing system 150. Thus, the camera 134 can collect a sequence of images from the sample.
  • To further remove residual excitation light and minimize cross talk between excitation channels, each laser emission wavelength can be paired with a corresponding band-pass emission filter 128 a. Each filter 128 a can have a wavelength of 10-50 nm, e.g., 14-32 nm. In some implementations, a filter is narrower than the bandwidth of the fluorescent material of the probe resulting from the excitation, e.g., if the fluorescent material of the probe has a long trailing spectral profile.
  • The filters are installed on a high-speed filter wheel 128 that is rotatable by an actuator 128 b. The filter wheel 128 can be installed at the infinity space to minimize optical aberration in the imaging path. After passing the emission filter of the filter wheel 128, the cleaned fluorescence signals can be refocused by a tube lens and captured by the camera 134. The dichroic mirror 138 can be positioned in the light path between the objective lens 138 and the filter wheel 128.
  • To facilitate high speed, synchronized operation of the system, the control system 140 can include two microcontrollers 144, 146 that are employed to send trigger signals, e.g., TTL signals, to the components of the fluorescence microscope 120 in a coordinated manner. The first microcontroller 144 is directly run by the computer 142, and triggers actuator 128 b of the filter wheel 128 to switch emission filters 128 a at different color channels. The first microcontroller 144 also triggers the second microcontroller 146, which sends digital signals to the light source 122 in order to control which wavelength of light is passed to the sample 10. For example, the second microcontroller 146 can send on/off signals to the individual laser modules of the light source 122 to control which laser module is active, and thus control which wavelength of light is used for the excitation light. After completion of switching to a new excitation channel, the second microcontroller 146 controls the motor for the piezo stage 116 to select the imaging height. Finally the second microcontroller 146 sends a trigger signal to the camera 134 for image acquisition.
  • Communication between the computer 142 and the device components of the apparatus 100 is coordinated by the control software. This control software can integrate drivers of all the device components into a single framework, and thus can allow a user to operate the imaging system as a single instrument (instead of having to separately control many devices).
  • The control software supports interactive operations of the microscope and instant visualization of imaging results. In addition, the control software can provide a programming interface which allows users to design and automate their imaging workflow. A set of default workflow scripts can be designated in the scripting language.
  • In some implementations, the control system 140 is configured, i.e., by the control software and/or the workflow script, to acquire fluorescence images (also termed simply “collected images” or simply “images”) in loops in the following order (from innermost loop to outermost loop): z-axis, color channel, lateral position, and reagent.
  • These loops may be represented by the following pseudocode:
  • for h = 1:N _hybridization
      % multiple hybridizations
      for f= 1:N_FOVs
        % multiple lateral field-of-views
        for c = 1 :N_channels
           % multiple color channels
           for z = 1:N_planes
             % multiple z planes
             Acquire image(h, f, c, z);
           end % end for z
        end % end for c
      end % end for f
                   end % end for h
  • For the z-axis loop, the control system 140 causes the stage 118 to step through multiple vertical positions. Because the vertical position of the stage 118 is controlled by a piezoelectric actuator, the time required to adjust positions is small and each step in this loop is extremely fast.
  • First, the sample can be sufficiently thick, e.g., a few microns, that multiple image planes through the sample may be desirable. For example, multiple layers of cells can be present, or even within a cell there may be a vertical variation in gene expression. Moreover, for thin samples, the vertical position of the focal plane may not be known in advance, e.g., due to thermal drift. In addition, the sample 10 may vertically drift within the flow cell 110. Imaging at multiple Z-axis positions can ensure most of the cells in a thick sample are covered, and can help identify the best focal position in a thin sample.
  • For the color channel loop, the control system 140 causes the light source 122 to step through different wavelengths of excitation light. For example, one of the laser modules is activated, the other laser modules are deactivated, and the emission filter wheel 128 is rotated to bring the appropriate filter into the optical path of the light between the sample 10 and the camera 134.
  • For the lateral position, the control system 140 causes the light source 122 to step through different lateral positions in order to obtain different fields of view (FOVs) of the sample. For example, at each step of the loop, the linear actuators supporting the stage 118 can be driven to shift the stage laterally. In some implementations, the control system 140 number of steps and lateral motion is selected such that the accumulated FOVs to cover the entire sample 10. In some implementations, the lateral motion is selected such that FOVs partially overlap.
  • For the reagent, the control system 140 causes the apparatus 100 to step through multiple different available reagents. For example, at each step of the loop, the control system 140 can control the valve 114 to connect the flow cell 110 to the purge fluid 112 b, cause the pump 116 to draw the purge fluid through the cell for a first period of time to purge the current reagent, then control the valve 114 to connect the flow cell 110 to different new reagent, and then draw the new reagent through the cell for a second period of time sufficient for the probes in the new reagent to bind to the appropriate RNA sequences. Because some time is required to purge the flow cell and for the probes in the new reagent to bind, the time required to adjust reagents in is the longest, as compared to adjusting the lateral position, color channel or z-axis.
  • As a result, a fluorescence image is acquired for each combination of possible values for the z-axis, color channel (excitation wavelength), lateral FOV, and reagent. Because the innermost loop has the fastest adjustment time, and the successively surrounding loops are of successively slower adjustment time, this configuration provides the most time efficient technique to acquire the images for the combination of values for these parameters.
  • A data processing system 150 is used to process the images and determine gene expression to generate the spatial transcriptomic data. At a minimum, the data processing system 150 includes a data processing device 152, e.g., one or more processors controlled by software stored on a computer readable medium, and a local storage device 154, e.g., non-volatile computer readable media, that receives the images acquired by the camera 134. For example, the data processing device 152 can be a work station with GPU processors or FPGA boards installed. The data processing system 150 can also be connected through a network to remote storage 156, e.g., through the Internet to cloud storage.
  • In some implementations, the data processing system 150 performs on-the-fly image processing as the images are received. In particular, while data acquisition is in progress, the data processing device 152 can perform image pre-processing steps, such as filtering and deconvolution, that can be performed on the image data in the storage device 154 but which do not require the entire data set. Because filtering and deconvolution are a major bottleneck in the data processing pipeline, pre-processing as image acquisition is occurring can significantly shorten the offline processing time and thus improve the throughput.
  • FIG. 2 illustrates a flow chart of a method of data processing in which the processing is performed after all of the images have been acquired. The process begins with the system receiving the raw image files and supporting files (step 202). In particular, the data processing system can receive the full set of raw images from the camera, e.g., an image for each combination of possible values for the z-axis, color channel (excitation wavelength), lateral FOV, and reagent.
  • In addition, the data processing system can receive a reference expression file, e.g., a FPKM (fragments per kilobase of sequence per million mapped reads) file, a data schema, and one or more stain images, e.g., DAPI images. The reference expression file can be used to cross-check between traditional sequence results and the mFISH results.
  • The image files received from the camera can optionally include metadata, the hardware parameter values (such as stage positions, pixel sizes, excitation channels, etc.) at which the image was taken. The data schema provides a rule for ordering the images based on the hardware parameters so that the images are placed into one or more image stacks in the appropriate order. If metadata is not included, the data schema can associate an order of the images with the values for the z-axis, color channel, lateral FOV and reagent used to generate that image.
  • The stain images will be presented to the user with the transcriptomic information overlaid.
  • The collected images can be subjected to one or more quality metrics (step 203) before more intensive processing in order to screen out images of insufficient quality. Only images that meet the quality metric(s) are passed on for further processing. This can significantly reduce processing load on the data processing system
  • In order to detect focusing failures, a sharpness quality value can be determined for each collected image. As an example, sharpness quality can be calculated by summing the gradient in intensity over the image:
  • S = d x d y I x , y 2
  • If the sharpness quality value is below a threshold, then that image can be identified as out of focus. The threshold can be a preset empirically determined value, or the threshold can be calculated based on the sharpness values for the images with the same FOV. For example, a standard deviation of sharpness quality values can be calculated, the threshold can be set based on the standard deviation.
  • As noted above, the z-axis position of the focal plane might not be known exactly. Images which are out of focus can be screened out. In some implementations, in each z-axis loop, only a single image is retained, e.g., the image having the best sharpness quality. In some implementation, multiple images from a single each z-axis loop can meet the sharpness quality, and the multiple images are combined, e.g., blended, to form a single image.
  • In order to detect regions of interest, a brightness quality value can be determined for each collected image. The brightness quality can be used to determine whether any cells are present in the image. For example, the intensity values of all the pixels in the image can be summed and compared to a threshold. If the total is less than the threshold, then this can indicate that there is essentially nothing in the image, i.e., no cells are in the image, and there is no information of interest and the image need not be processed. As another example, a standard deviation of the intensity values in the image can be calculated and compared to a threshold. If the standard deviation is below the threshold, this indicates that the image is fairly uniform and may lack bright spots indicating excited probes. Again, such a collected image does not have information of interest and the image need not be processed.
  • The above two quality checks are exemplary and other quality metrics can be used. For example, a quality metric that measures spot quality can be used.
  • Another quality check can detect inter-hybridization shift. If there is significant sample drift occurring between rounds of imaging, then the subsequent registration may be unreliable. Sample drift can be detected by a variety of image processing algorithms, e.g., phase-correlation. In particular, a collected image can be compared to an image for the same FOV but for a prior reagent. If the images are not sufficiently correlated, e.g., does not meet a threshold correlation, the collected image does not have information of interest and the image need not be processed.
  • Next, each image is processed to remove experimental artifacts (step 204). Since each RNA molecule will be hybridized multiple times with probes at different excitation channels, a strict alignment across the multi-channel, multi-round image stack is beneficial for revealing RNA identities over the whole FOV. Removing the experimental artifacts can include field flattening and/or chromatic aberration correction In some implementations, the field flattening is performed before the chromatic aberration correction.
  • Imperfections of the illumination profile (e.g., Gaussian or other uneven distributions instead of uniform distribution across the FOV) in the excitation light can lead to uneven fluorescent signal level across the FOV, and this spatial unevenness often varies between excitation channels. Such nonuniformity can be passed to from illumination field to fluorescent signal level across the FOV, and consequently bias the spatial transcriptomic analysis results. To eliminate this artifact, an illumination profile, e.g., test image, is measured for each color channel using a fluorescent target with uniform fluorophore density within the FOV. These measured illumination profiles are then used as channel-specific normalization factors in imaging data. For example, the illumination profile for each color channel (i.e., each emission wavelength to be measured) can be stored as a normalization image, and the intensity value for each pixel in a collected image for a color channel is divided by the intensity value for the corresponding pixel in the normalization image of the associated color channel.
  • Chromatic aberration correction accounts for optical distortions that vary across the different emission channels. That is, because the probes are being measured for different wavelengths, chromatic aberrations in the imaging path, if any, may significantly degrade the downstream analysis results. To maintain the robustness of analysis results against this potential hardware defect, a software-based solution for chromatic aberration correction can be provided in the data processing pipeline. In particular, a test image is generated for each color channel using a target having fluorescent spots in a known pattern, e.g., a rectangular pattern. For each color channel, a transform is calculated that brings the measured spots in the test image back into alignment with the known pattern. These transforms are stored, and in normal operation when images of a sample are being collected, an appropriate transform is selected and applied to each collected image based on the color channel for that collected image.
  • Each image is processed to provide RNA image spot sharpening (step 206). RNA image spot sharpening can include applying filters to remove cellular background and/or deconvolution with point spread function to sharpen RNA spots.
  • In order to distinguish RNA spots from relatively bright background, a low-pass filter is applied to the image, e.g., to the field-flattened and chromatically corrected images to remove cellular background around RNA spots. The filtered images are further de-convolved with a 2-D point spread function (PSF) to sharpen the RNA spots, and convolved with a 2-D Gaussian kernel with half pixel width to slightly smooth the spots.
  • The images having the same FOV are registered to align the features, e.g., the cells or cell organelles, therein (step 208). To accurately identify RNA species in the image sequences, features in different rounds of images are aligned, e.g., to sub-pixel precision. However, since an mFISH sample is imaged in aqueous phase and moved around by a motorized stage, sample drifts and stage drifts through hours-long imaging process can transform into image feature shifts, which can undermine the transcriptomic analysis if left unaddressed. In other words, even assuming precise repeatable alignment of the fluorescence microscope to the flow cell or support, the sample may no longer be in the same location in the later image, which can introduce errors into decoding or simply make decoding impossible.
  • One conventional technique to register images is to place fiducial markers, e.g., fluorescent beads, within the carrier material on the slide. In general, the sample and the fiducial marker beads will move approximately in unison. These beads can be identified in the image based on their size and shape. Comparison of the positions of the beads permits registration of the two images, e.g., calculation of an affine transformation.
  • However, it may be possible to register images without the use of fluorescent beads or similar fiducial markers, but simply using the data within the image. In particular, high intensity regions should generally be located at the same position across multiple images of the same FOV. Techniques that can be used for registration between images include phase-correlation algorithms and mutual-information (MI) algorithms. The MI algorithm is generally more precise and robust than the phase-correlation algorithms. Therefore, a mutual-information (MI) based drift correction method can be used for alignment. However, the precision can come at an expense of high computational time.
  • FIG. 3 illustrates a process for using a mutual-information algorithm to align collected images having the same FOV that preserves high precision of the MI registration algorithm while reducing computational load. The set of collected images having the same FOV can be considered an original image stack; the order of images in the stack is set by the data schema discussed above. Each collected image 302 in the image stack may be, for example, 2048×2048 pixels. In general, performing a mutual-information algorithm on an image of this size may require excessive computational power. The collected image 302 is evenly sliced into four image patches 304 a, 304 b, 304 c, 304 d. Assuming each collected image 302 is 2048×2048 pixels, then the image patches are 1024×1024 pixels. Each image patch 304 a, 304 b, 304 c, 304 d is multiplied by a phase factor ϕ, the phase shifted image patches are overlapped and summed to generate a complex image, and the real and imaginary parts of the complex image are summed up to create a summed image 306. This summed image 306 has ¼ size of the original image 302, e.g., is 1024×1024 pixels.
  • This process is repeated for each original image 302 in the original image stack to generate a stack 308 of images 306 of reduced side. The mutual-information algorithm is then performed using the stack 308 to generate a transform for each reduced image 306 in the stack 306. The transform can include a translation and/or scaling transform. To align the images in the original stack, the transform determined for each reduced image is applied to the corresponding collected image in the original stack. This should result in a stack 312 of collected images 310 of the original size, e.g., 2048×2048 pixels, in which the images 310 for the same FOV are in alignment.
  • This slice-and-overlap strategy preserves most of the features in the image and thus should still have high precision, but reduces the temporary image size for shift calculation and thus has reduced computational load. Although FIG. 3 and the discussion above focuses on splitting into four patches corresponding to upper right, lower right, upper left and lower left quadrants of the original image, another number of patches can be used, e.g., the image can be split into other shapes, e.g., rectangular rather than square, so long as the patches have the same dimensions. For example, the original image could be split into 8 image patches of 256×2048 pixels. In general, the image can be split into a rectangular grid having a plurality of columns and a plurality of rows, e.g., an equal number of columns and rows. The image can be split into N images, where N=Z2 and Z is an even number integer (N=2 for the example in FIG. 3 ).
  • Returning to FIG. 2 , because the filtering and deconvolution processes of step 206 are implemented on individual frames (i.e., bits) independently, they can be conducted prior to the registration of step 208. Moreover, since image stacks used for registration are background corrected and sharpened to reveal spatial features of RNA spots, the registration precision can be improved. To validate the latter advantage, three data sets from low confluency cell sample, high confluency cell sample, and brain tissue sample were compared for registration precision with registration first vs. filter first approaches. The registration quality was evaluated by summing up the registered image stack across all the layers, then calculating the maximum and the minimum of the summed image. The filter-first approach generally yields a larger maximum and smaller minimum of the flattened image, which indicates an improvement of the registration quality.
  • A registration quality check can be performed after registration. If properly registered, the bright points in each image should overlap so that the total brightness is increased. Referring the stack of unregistered images can be summed, and similarly the stack 312 of registered images 310 (see FIG. 3 ) can be summed, e.g.:
  • I U x , y = i = 1 B I U x , y , i I R x , y = i = 1 B I R x , y , i
  • The maximum value of IU(x,y) can be compared to the maximum value of IR(x,y). If max(IR(x,y)) > max(Iu(x,y)) then registration can be considered a success. On the other hand, if max(IR(x,y)) ≤ max(IU(x,y)), the data processing system can indicate the registration is a failure and issue an error code. This FOV might be excluded from subsequent analysis and/or new images may need to be obtained.
  • Optionally, after registration, a mask can be calculated for each collected image. In brief, the intensity value for each pixel is compared to a threshold value. A corresponding pixel in the mask is set to 1 if the intensity value is above the threshold, and set to 0 if the intensity value is below the threshold. The threshold value can be an empirically determined predetermined value, or can be calculated from the intensity values in the image. In general, the mask can correspond to the location of cells within the sample; spaces between cells should not fluoresce and should have a low intensity.
  • To create masks, for each field of view, a compressed frame P(x,y) is created from the image sequence for that FOV. The compressed frame P(x,y) can be generated by performing a product operation on the images for that FOV, e.g.:
  • P x , y = i B I x , y , i ,
  • where i represents a layer number in the stack 312 of registered images 310 (see FIG. 3 ). A threshold can be generated based on the histogram of log P(x,y).
  • For example, FIG. 4 illustrates a logarithmic histogram 400 of compressed frame values (P(x,y)) in a collected image. In general, the logarithmic histogram will include an initial sharp increase in intensity (which may be the leading edge of a first local maxima), and a final sharp decrease in intensity (which may be the trailing edge of a second local maxima). The histogram can be subjected to a peak and/or edge-finding algorithm to identify a first peak 402 that is closest to the lower end of the histogram (or an edge of the sharp increase in intensity if there is no clear peak) and a second peak 404 that is closest to the upper end of the histogram (or an edge of the sharp decrease in intensity if there is no clear peak). A threshold 410 can be set based on the intensity values of the two identified points 402, 404. For example, the threshold 410 can be an average of the intensity values. Instead of peak-finding, the histogram can be subjected to an Otsu threshold method to determine the threshold 410. A. This technique can be useful because the histogram does not always exhibit two clear peaks.
  • Salt and pepper effects can be removed from the mask. Single “on” pixels which are outside a continuous region of “on” pixels, or single “off” pixels which are inside a continuous region of “on” pixels can be converted to “off” or “on” to match the surrounding area. In general, clusters of adjacent pixels in the mask below a threshold size, e.g., 50 pixels, can be indicated as an error and be changed to match the surrounding area. By using masks, memory space can be conserved and computational workload can be reduced.
  • A collected image is illustrated in FIG. 5A, and FIG. 5B illustrates a mask resulting from the image of FIG. 5A.
  • Returning to FIG. 2 , after registration of the images in a FOV, spatial transcriptomic analysis can be performed (step 210).
  • FOV normalization can be performed before the spatial transcriptomic analysis in order to make the histogram more consistent. In some implementations, the FOV normalization occurs after registration. Alternatively, FOV normalization can occur before registration. FOV normalization could be considered part of the filtering. A problem in the collected image, even after filtering and deconvolution discussed above, is that there can be a significant variation in brightness in the brightest pixels, e.g., over several orders of magnitude. For example, referring to FIG. 6A, the histogram 602 of intensity values in a collected image has a very long tail 604. This can introduce uncertainty in the decoding process. FOV normalization can counteract this source of error.
  • First, all the intensity values in the image are normalized relative to the maximum intensity value in the image. For example, the maximum intensity value is determined, and all intensity values are divided by the maximum so that intensity values vary between 0 and IMAX, e.g., 1.
  • Next the intensity values in the image are analyzed to determine an upper quantile that includes the highest intensity values, e.g., the 99% and higher quantile (i.e., upper 1%), 99.5% and higher quantile, or 99.9% and higher quantile, or 99.95% and higher quantile (i.e., upper 0.05%). The cutoff for this quantile is indicated by a broken line 610 in FIG. 6A. The intensity value at this quantile limit, e.g., the intensity value of log(-2) at the 99.95% quantile in the example of FIG. 6A, can be determined and stored. All pixels having intensity values within the upper quantile are reset to have the maximum intensity value, e.g., 1. Then the intensity values of the remaining pixels are scaled to run to the same maximum. To accomplish this, intensity values for the pixels that were not in the upper quantile are divided by the stored intensity value for the quantile limit. FIG. 6B illustrates the histogram 620 of intensity values resulting from the FOV normalization.
  • Decoding is explained with reference to FIG. 7 . However, returning momentarily to FIG. 3 , the aligned images for a particular FOV can be considered as a stack 312 that includes multiple image layers 310, with each image layer being X by Y pixels, e.g., 2048×2048 pixels. The number of image layers, B, depends on the combination of the number of color channels (e.g., number of excitation wavelengths) and number of hybridizations (e.g., number of reactants), e.g., B = N_hybridization * N_channels.
  • Turning to FIG. 7 , after normalization, this image stack can be evaluated as a 2-D matrix 702 of pixel words. The matrix 702 can have P rows 704, where P = X * Y, and B columns 706, where B is the number of images in the stack for a given FOV, e.g., N _hybridization * N_channels. Each row 704 corresponds to one of the pixels (the same pixel across the multiple image in the stack), the values from the row 704 provide a pixel word 710. Each column 706 provides one of the values in the word 710, i.e., the intensity value from the image layer for that pixel. As noted above, the values can be normalized, e.g., vary between 0 and IMAX. Different intensity values are represented in FIG. 7 as different degrees of shading of the respective cells.
  • If all the pixels are passed to the decoding step, then all P words will be processed as described below. However, pixels outside cell boundaries can be screened out by the 2-D masks (see FIG. 4B above) and not processed. As result, computational load can be significantly reduced in the following analysis.
  • The data processing system 150 stores a code book 722 that is used to decode the image data to identify the gene expressed at the particular pixel. The code book 722 includes multiple reference code words, each reference code word associated with a particular gene. As shown in FIG. 7 , the code book 722 can be represented as a 2D matrix with G rows 724, where G is the number of code words, e.g., the number of genes (although the same gene could be represented by multiple code words), and B columns 726. Each row 724 corresponds to one of the reference code words 730, and each column 706 provides one of the values in the reference code word 730, as established by prior calibration and testing of known genes. For each column, the values in the reference code 730 can be binary, i.e., “on” or “off”. For example, each value can be either 0 or IMAX, e.g., 1. The on and off values are represented in FIG. 7 by light and dark shading of respective cells.
  • For each pixel to be decoded, a distance d(p,i) is calculated between the pixel word 710 and each reference code word 730. For example, the distance between the pixel word 710 and reference code word 730 can be calculated as a Euclidean distance, e.g., a sum of squared differences between each value in the pixel word and the corresponding value in the reference code word. This calculation can be expressed as:
  • d p , i = x = 1 B I p , x C i , x 2
  • where Ip,x are the values from the matrix 702 of pixel words and Ci,x are the values from the matrix 722 of reference code words. Other metrics, e.g., sum of absolute value of differences, cosine angle, correlation, etc., can be used instead of a Euclidean distance.
  • Once the distance values for each code word are calculated for a given pixel, the smallest distance value is determined, the code word that provides that smallest distance value is selected as the best matching code word. Stated differently, the data processing apparatus determines min (d(p,1), d(p,2), ... d(p,B)), and determines the value b as the value for i (between 1 and B) that provided the minimum. The gene corresponding to that best matching code word is determined, e.g., from a lookup table that associates code words with genes, and the pixel is tagged as expressing the gene.
  • Returning to FIG. 2 , the data processing apparatus can filter out false callouts. One technique to filter out false callouts is to discard tags where the distance value d(p,b) that indicated expression of a gene is greater than a threshold value, e.g., if d(p,b) > D1MAX.
  • Yet another technique for filtering false callouts is to reject code words where a calculated bit ratio falls below a threshold. The bit ratio is calculated as the mean of the intensity values from the image word for layers that are supposed to be on (as determined from the code word), divided by the mean of the intensity values from the image word for layers that are supposed to be off (again as determined from the code word). For example, for the code word 730 shown in FIG. 7 , the bits corresponding to columns 3-6 and 12-14 are on, and the bits corresponding to columns 1-2 and 7-11 are off. Thus, assuming this code word is selected as the best match for a pixel p, the bit ratio BR can be calculated as
  • BR= I p , 3 + I p , 4 + I p , 5 + I p , 6 + I p , 12 + I p , 13 + I p , 14 / 7 I p , 1 + I p , 2 + I p , 7 + I p , 8 + I p , 9 + I p , 10 + I p , 11 / 7 .
  • The bit ratio BR is compared to a threshold value THBR. In some implementations, the threshold value THBR is determined empirically from prior measurements. However, in some implementations, the threshold value THBR can be calculated automatically for a particular code word based on the measurements obtained from the sample. FIG. 8A illustrates a logarithmic histogram 802 of bit ratios intensity values for all pixels that were tagged with a particular code word. A threshold value 804 is calculated based on this histogram 802. For example, a default quantile, e.g., 50%, can be used to calculate an initial threshold value 802. However, the quantile can be adjusted during optimization discussed below.
  • Yet another technique for filtering false callouts is to reject code words where a calculated bit brightness falls below a threshold. The bit brightness is calculated as the mean of the intensity values from the image word for layers that are supposed to be on (as determined from the code word). For example, for the code word 730 shown in FIG. 7 , the bits corresponding to columns 3-6 and 12-14 are on. Thus, assuming this code word is selected as the best match for a pixel p, the bit brightness BB can be calculated as
  • BB= I p,3 +I p,4 +I p,5 +I p,6 +I p,12 +I p,13 +I p,14 / 7
  • The bit brightness BB is compared to a threshold value THBB. In some implementations, the threshold value THBB is determined empirically from prior measurements. However, in some implementations, the threshold value THBB can be calculated automatically for a particular code word based on the measurements obtained from the sample. FIG. 8B illustrates a logarithmic histogram 812 of bit brightness values for all pixels that were tagged with a particular code word. A threshold value 814 is calculated based on this histogram 812. For example, a default quantile, e.g., 80% or 90%, can be used to calculate an initial threshold value 814. However, the quantile can be adjusted during optimization discussed below.
  • Returning to FIG. 2 , the data processing apparatus can now perform optimization and re-decoding (step 212). The optimization can include machine-learning based optimization of the decoding parameters, followed by returning to step 210 with updated decoding parameters in order to update the spatial transcriptomic analysis. This cycle can be repeated until the decoding parameters have stabilized.
  • The optimization of the decoding parameters will use a merit function, e.g., a FPKM/TPM correlation, spatial correlation, or confidence ratio. Parameters that can be included as variables in the merit function include the shape (e.g., start and end of frequency range, etc.) of the filters used to remove cellular background, the numerical aperture value for the point spread function used to sharpen the RNA spots, the quantile boundary Q used in normalization of the FOV, the bit ratio threshold THBR, the bit brightness threshold THBB (or the quantiles used to determine the bit ratio threshold THBR and bit brightness threshold THBB), and/or the maximum distance D1max at which at which a pixel word can be considered to match a code word.
  • This merit function may be an effectively discontinuous function, so a conventional gradient following algorithm may be insufficient to identify the optimal parameter values. A machine learning model can be used to converge on parameter values.
  • Next, the data processing apparatus can perform unification of the parameter values across all FOVs. Because each FOV is processed individually, each field can experience different normalization, thresholding, and filtering setting. As a result, a high contrast image can result in a histogram with variation that causes false positive callouts in quiet areas. The result of unification is that all FOVs use the same parameter values. This can significantly remove callouts from background noise in quiet area, and can provide a clear and unbiased spatial pattern in large sample area.
  • A variety of approaches are possible to select a parameter value that will be used across all FOVs. One option is to simply pick a predetermined FOV, e.g., the first measured FOV or a FOV near the center of the sample, and use the parameter value for that predetermined FOV. Another option is to average the values for the parameter across multiple FOV and then use the averaged value. Another option is to determine which FOV resulted in the best fit between its pixel words and tagged code words. For example, a FOV with the smallest average distance d(p,b1) between the tagged code words and the pixel words for those code words can be determined and then selected.
  • Alternatively, unification can be performed before the spatial transcriptomic analysis, e.g., between steps 208 and 210. In this case, the unification would be performed for parameters selected before step 210, e.g., parameters used in steps 204 or 206, but not for the parameters in steps 210 and 212.
  • The data processing apparatus can now perform stitching and segmentation (step 214). Stitching combines multiple FOV into a single image. Stitching can be performed using a variety of techniques. Referring to FIG. 9 , one approach is, for each row of FOV that together will form the combined image of the sample and each FOV within the row, determine a horizontal shift for each FOV. Once the horizontal shifting is calculated, a vertical shift is calculated for each row of FOV. The horizontal and vertical shifts can be calculated based on cross-correlation, e.g., phase correlation. With the horizontal and vertical shift for each FOV, a single combined image can be generated, and gene coordinates can be transferred to the combined image based on the horizontal and vertical shift.
  • An indication that a gene is expressed at a certain coordinate in the combined fluorescence image (as determined from the coordinate in the FOV and the horizontal and vertical shift for that FOV) can be added, e.g., as metadata. This indication can be termed a “callout.”
  • The stain images, e.g., the DAPI images, can be stitched together using the technique discussed with respect to FIG. 9 to generate a combined stain image. In some implementations, it is not necessary to create a combined fluorescence image from the collected fluorescence images; once the horizontal and vertical shift for each FOV is determined, the gene coordinates within the combined stain image can be calculated. The stain image can be registered to the collected fluorescent image(s). An indication that a gene is expressed at a certain coordinate in the combined stain image (as determined from the coordinate in the FOV and the horizontal and vertical shift for that FOV) can be added, e.g., as metadata, to provide a callout.
  • A potential problem remains in the stitched image. In particular, some genes may be double-counted in the overlapping area. To remove double-counting, a distance, e.g., Euclidean distance, can be calculated between each pixel tagged as expressing a gene and other nearby pixels tagged as expressing the same gene. One of the callouts can be removed if the distance is below a threshold value. More complex techniques can be used if a cluster of pixels are tagged as expressing a gene.
  • Segmentation of the combined image, e.g., the image of the stained cell, into regions corresponding to cells can be performed using various known techniques. Segmentation is typically performed after stitching of the images, but can occur before or after callouts are added to the combined image.
  • The segmented image with callouts indicating positions of gene expression, can now be stored and presented to a user, e.g., on a visual display, for analysis.
  • Although the discussion above assumes that a single z-axis image is used in for each FOV, this is not required. Images from different z-axis positions can be processed separately; effectively the different z-axis positions provide a new set of FOVs.
  • As noted above, one problem is that the data processing, including the image processing steps, is computationally demanding and can require a significant amount of time. One issue is that in conventional techniques, all images are acquired before the image processing commences. However, it is possible to perform on-the-fly image processing as the images are received.
  • FIG. 10 illustrates a modified data processing method in which image processing is performed on-the-fly. Steps that are not specifically discussed can be performed in a manner similar to the method of FIG. 2 . Initially, the data processing system can receive the supporting files (step 202 a), although this step could be postponed until after image acquisition. It may be necessary for some files, e.g., the data schema, to be loaded to the data processing system for the image processing to be performed.
  • The apparatus begins the process of acquiring new images (step 202 b), e.g., stepping through the loops to adjust the z-axis, color channel, lateral position, and reagent, and acquiring an image at each combination of possible values for these parameters.
  • As each image is acquired, a quality control (QC) test is performed to determine whether the image is of sufficient quality to be pre-processed (step 203). If it is, on-the-fly pre-processing of the acquired image is performed (step 252). In this context, “on-the-fly” indicates that the processing of one image can be performed in parallel with acquisition of subsequent images. The on-the-fly pre-processing can include the removing the experimental artifacts (step 204) and RNA image spot sharpening (step 206). It may be noted that since these image preprocessing steps are performed individually for each image, and do not depend on other images in the same FOV or images for other FOV, shifting processing in this manner does not adversely impact the quality of processing. In general, significant computational load can be shifted to times when the reagent is being changed. As a result, overall time from loading of the sample into the flow cell to receiving the segmented image with callouts is significantly reduced, which can significantly improve throughput and speed research.
  • In some implementations, the images are pre-processed in “real time,” i.e., processing of one image is performed in parallel with collection of a subsequent image by the apparatus and completed before a subsequent image is received at the data processing system.
  • This specification uses the term “configured” in connection with systems and computer program components. For a system of one or more computers to be configured to perform particular operations or actions means that the system has installed on it software, firmware, hardware, or a combination of them that in operation cause the system to perform the operations or actions. For one or more computer programs to be configured to perform particular operations or actions means that the one or more programs include instructions that, when executed by data processing apparatus, cause the apparatus to perform the operations or actions.
  • Embodiments of the subject matter and the functional operations described in this specification can be implemented in digital electronic circuitry, in tangibly-embodied computer software or firmware, in computer hardware, including the structures disclosed in this specification and their structural equivalents, or in combinations of one or more of them. Embodiments of the subject matter described in this specification can be implemented as one or more computer programs, i.e., one or more modules of computer program instructions encoded on a tangible non-transitory storage medium for execution by, or to control the operation of, data processing apparatus. The computer storage medium can be a machine-readable storage device, a machine-readable storage substrate, a random or serial access memory device, or a combination of one or more of them. Alternatively or in addition, the program instructions can be encoded on an artificially-generated propagated signal, e.g., a machine-generated electrical, optical, or electromagnetic signal, that is generated to encode information for transmission to suitable receiver apparatus for execution by a data processing apparatus.
  • The term “data processing apparatus” refers to data processing hardware and encompasses all kinds of apparatus, devices, and machines for processing data, including by way of example a programmable processor, a computer, or multiple processors or computers. The apparatus can also be, or further include, special purpose logic circuitry, e.g., an FPGA (field programmable gate array) or an ASIC (application-specific integrated circuit). The apparatus can optionally include, in addition to hardware, code that creates an execution environment for computer programs, e.g., code that constitutes processor firmware, a protocol stack, a database management system, an operating system, or a combination of one or more of them.
  • A computer program, which may also be referred to or described as a program, software, a software application, an app, a module, a software module, a script, or code, can be written in any form of programming language, including compiled or interpreted languages, or declarative or procedural languages; and it can be deployed in any form, including as a stand-alone program or as a module, component, subroutine, or other unit suitable for use in a computing environment. A program may, but need not, correspond to a file in a file system. A program can be stored in a portion of a file that holds other programs or data, e.g., one or more scripts stored in a markup language document, in a single file dedicated to the program in question, or in multiple coordinated files, e.g., files that store one or more modules, sub-programs, or portions of code. A computer program can be deployed to be executed on one computer or on multiple computers that are located at one site or distributed across multiple sites and interconnected by a data communication network.
  • The processes and logic flows described in this specification can be performed by one or more programmable computers executing one or more computer programs to perform functions by operating on input data and generating output. The processes and logic flows can also be performed by special purpose logic circuitry, e.g., an FPGA or an ASIC, or by a combination of special purpose logic circuitry and one or more programmed computers.
  • Computers suitable for the execution of a computer program can be based on general or special purpose microprocessors or both, or any other kind of central processing unit. Generally, a central processing unit will receive instructions and data from a read-only memory or a random access memory or both. The essential elements of a computer are a central processing unit for performing or executing instructions and one or more memory devices for storing instructions and data. The central processing unit and the memory can be supplemented by, or incorporated in, special purpose logic circuitry. Generally, a computer will also include, or be operatively coupled to receive data from or transfer data to, or both, one or more mass storage devices for storing data, e.g., magnetic, magneto-optical disks, or optical disks. However, a computer need not have such devices.
  • Computer-readable media suitable for storing computer program instructions and data include all forms of non-volatile memory, media and memory devices, including by way of example semiconductor memory devices, e.g., EPROM, EEPROM, and flash memory devices; magnetic disks, e.g., internal hard disks or removable disks; magneto-optical disks; and CD-ROM and DVD-ROM disks.
  • To provide for interaction with a user, embodiments of the subject matter described in this specification can be implemented on a computer having a display device, e.g., a CRT (cathode ray tube) or LCD (liquid crystal display) monitor, for displaying information to the user and a keyboard and a pointing device, e.g., a mouse or a trackball, by which the user can provide input to the computer. Other kinds of devices can be used to provide for interaction with a user as well; for example, feedback provided to the user can be any form of sensory feedback, e.g., visual feedback, auditory feedback, or tactile feedback; and input from the user can be received in any form, including acoustic, speech, or tactile input. In addition, a computer can interact with a user by sending documents to and receiving documents from a device that is used by the user; for example, by sending web pages to a web browser on a user’s device in response to requests received from the web browser. Also, a computer can interact with a user by sending text messages or other forms of message to a personal device, e.g., a smartphone that is running a messaging application, and receiving responsive messages from the user in return.
  • Data processing apparatus for implementing machine learning models can also include, for example, special-purpose hardware accelerator units for processing common and compute-intensive parts of machine learning training or production, i.e., inference, workloads.
  • Machine learning models can be implemented and deployed using a machine learning framework, e.g., a TensorFlow framework, a Microsoft Cognitive Toolkit framework, an Apache Singa framework, or an Apache MXNet framework.
  • Embodiments of the subject matter described in this specification can be implemented in a computing system that includes a back-end component, e.g., as a data server, or that includes a middleware component, e.g., an application server, or that includes a front-end component, e.g., a client computer having a graphical user interface, a web browser, or an app through which a user can interact with an implementation of the subject matter described in this specification, or any combination of one or more such back-end, middleware, or front-end components. The components of the system can be interconnected by any form or medium of digital data communication, e.g., a communication network. Examples of communication networks include a local area network (LAN) and a wide area network (WAN), e.g., the Internet.
  • While this specification contains many specific implementation details, these should not be construed as limitations on the scope of any invention or on the scope of what may be claimed, but rather as descriptions of features that may be specific to particular embodiments of particular inventions. Certain features that are described in this specification in the context of separate embodiments can also be implemented in combination in a single embodiment.
  • Conversely, various features that are described in the context of a single embodiment can also be implemented in multiple embodiments separately or in any suitable subcombination. Moreover, although features may be described above as acting in certain combinations and even initially be claimed as such, one or more features from a claimed combination can in some cases be excised from the combination, and the claimed combination may be directed to a subcombination or variation of a subcombination.
  • Similarly, while operations are depicted in the drawings and recited in the claims in a particular order, this should not be understood as requiring that such operations be performed in the particular order shown or in sequential order, or that all illustrated operations be performed, to achieve desirable results. In certain circumstances, multitasking and parallel processing may be advantageous. Moreover, the separation of various system modules and components in the embodiments described above should not be understood as requiring such separation in all embodiments, and it should be understood that the described program components and systems can generally be integrated together in a single software product or packaged into multiple software products.
  • Particular embodiments of the subject matter have been described. Other embodiments are within the scope of the following claims. For example, the actions recited in the claims can be performed in a different order and still achieve desirable results . As one example, the processes depicted in the accompanying figures do not necessarily require the particular order shown, or sequential order, to achieve desirable results. In some cases, multitasking and parallel processing may be advantageous.

Claims (18)

What is claimed is:
1. A fluorescent in-situ hybridization imaging system, comprising:
a flow cell to contain a sample to be exposed to fluorescent probes in a reagent;
a fluorescence microscope including
a variable frequency excitation light source to illuminate the sample,
a plurality of emission bandpass filters on a filter wheel,
a first actuator to rotate the filter wheel,
a camera positioned to receive fluorescently emitted light from the sample that has passed through a filter on the filter wheel;
a control system configured to
cause the variable frequency excitation light source to emit a light beam having a selected wavelength,
cause the first actuator to rotate the filter wheel to position a selected filter in a light path between the sample and the camera,
obtain an image from the camera, and
coordinate the variable frequency excitation light source and filter wheel such that the selected filter has an emission bandpass associated with emission by the fluorescent probes when excited by the selected wavelength when the image is obtained,
wherein the control system comprises a first microcontroller cause the first actuator to rotate the filter wheel in response to a first trigger signal and to generate a second trigger signal, and a second microcontroller to cause the variable frequency excitation light source to emit the light beam having the selected wavelength in response to the second trigger signal.
2. The system of claim 1, wherein the second microcontroller is configured to send a third trigger signal to the camera to cause image acquisition.
3. The system of claim 1, comprising a dichroic filter in the light path between the sample and the camera.
4. The system of claim 3, wherein the dichroic filter is positioned in the light path between the sample and the filter wheel.
5. The system of claim 3, wherein the dichroic mirror is a multi-pass dichroic mirror having a pass band for each emission wavelength for the fluorescent probes.
6. The system of claim 1, wherein each filter of the plurality of emission bandpass filters has a unique band and is associated with a unique wavelength for the variable frequency excitation light source.
7. The system of claim 1, wherein the variable frequency excitation light source is configured to illuminate the sample with excitation light normal to the sample.
8. The system of claim 1, comprising a second actuator to cause relative vertical motion between the flow cell and the fluorescence microscope.
9. The system of claim 8, wherein the second microcontroller is configured to control the second actuator to select an imaging height of the fluorescence microscope above the flow cell.
10. The system of claim 8, wherein the actuator comprises a piezoelectric actuator.
11. The system of claim 1, comprising a valve to control flow from one of a plurality of reagent sources to the flow cell.
12. The system of claim 11, wherein the control system includes a computer to control the valve.
13. The system of claim 12, wherein the first microcontroller receives input from the computer.
14. The system of claim 12, further comprising a motor to cause to cause relative lateral motion between the flow cell and the fluorescence microscope.
15. The system of claim 14, comprising a stage to support the sample in the flow cell, and wherein the motor comprises a pair of linear actuators configured to drive the stage along a pair of perpendicular axes.
16. The system of claim 14, wherein the computer is configured to control the motor.
17. The system of claim 1, wherein the variable frequency excitation light source comprises a plurality of laser modules and the control system is configured to sequentially activate the laser modules to sequentially emit the plurality of different wavelengths.
18. The system of claim 1, further comprising a data processing system configured to, for each field of view of at least two fields of view of the plurality of different fields of view and for each pixel of at least two pixels shared in images of the field of view, decode the pixel by identifying a code word from a plurality of code words in a code book that provides a best match to data values for the pixel in the plurality of images for the field of view.
US18/301,875 2019-12-17 2023-04-17 System for acquisition and processing of multiplexed fluorescence in-situ hybridization images Pending US20230253073A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US18/301,875 US20230253073A1 (en) 2019-12-17 2023-04-17 System for acquisition and processing of multiplexed fluorescence in-situ hybridization images

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
US201962949391P 2019-12-17 2019-12-17
US17/124,285 US11630067B2 (en) 2019-12-17 2020-12-16 System for acquisition and processing of multiplexed fluorescence in-situ hybridization images
US18/301,875 US20230253073A1 (en) 2019-12-17 2023-04-17 System for acquisition and processing of multiplexed fluorescence in-situ hybridization images

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US17/124,285 Continuation US11630067B2 (en) 2019-12-17 2020-12-16 System for acquisition and processing of multiplexed fluorescence in-situ hybridization images

Publications (1)

Publication Number Publication Date
US20230253073A1 true US20230253073A1 (en) 2023-08-10

Family

ID=76316250

Family Applications (5)

Application Number Title Priority Date Filing Date
US17/124,305 Active 2041-04-29 US11624708B2 (en) 2019-12-17 2020-12-16 Image processing techniques in multiplexed fluorescence in-situ hybridization
US17/124,333 Abandoned US20210182531A1 (en) 2019-12-17 2020-12-16 Registration techniques for multiplexed fluorescence in-situ hybridization images
US17/124,285 Active US11630067B2 (en) 2019-12-17 2020-12-16 System for acquisition and processing of multiplexed fluorescence in-situ hybridization images
US17/124,353 Active 2041-03-17 US11783916B2 (en) 2019-12-17 2020-12-16 System and method for acquisition and processing of multiplexed fluorescence in-situ hybridization images
US18/301,875 Pending US20230253073A1 (en) 2019-12-17 2023-04-17 System for acquisition and processing of multiplexed fluorescence in-situ hybridization images

Family Applications Before (4)

Application Number Title Priority Date Filing Date
US17/124,305 Active 2041-04-29 US11624708B2 (en) 2019-12-17 2020-12-16 Image processing techniques in multiplexed fluorescence in-situ hybridization
US17/124,333 Abandoned US20210182531A1 (en) 2019-12-17 2020-12-16 Registration techniques for multiplexed fluorescence in-situ hybridization images
US17/124,285 Active US11630067B2 (en) 2019-12-17 2020-12-16 System for acquisition and processing of multiplexed fluorescence in-situ hybridization images
US17/124,353 Active 2041-03-17 US11783916B2 (en) 2019-12-17 2020-12-16 System and method for acquisition and processing of multiplexed fluorescence in-situ hybridization images

Country Status (7)

Country Link
US (5) US11624708B2 (en)
EP (1) EP4078151A4 (en)
JP (1) JP7420950B2 (en)
KR (1) KR20220110589A (en)
CN (1) CN114868010A (en)
TW (2) TWI782368B (en)
WO (1) WO2021127019A1 (en)

Families Citing this family (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9928403B2 (en) * 2016-02-09 2018-03-27 Molecular Devices, Llc System and method for image analysis of multi-dimensional data
EP3644044B1 (en) * 2018-10-24 2020-12-23 Leica Biosystems Imaging, Inc. Camera exposure control when acquiring fluorescence in situ hybridization images
JP7280107B2 (en) * 2019-05-10 2023-05-23 株式会社エビデント Image processing method, program, image processing device, image processing system, and microscope system
EP4078151A4 (en) 2019-12-17 2024-01-03 Applied Materials, Inc. System and method for acquisition and processing of multiplexed fluorescence in-situ hybridization images
EP4214712A2 (en) 2020-09-16 2023-07-26 10X Genomics, Inc. Methods and systems for barcode error correction
JP7554083B2 (en) * 2020-09-18 2024-09-19 株式会社Screenホールディングス IMAGE PROCESSING METHOD, IMAGE PROCESSING APPARATUS AND IMAGING APPARATUS
DE102021115588A1 (en) * 2021-06-16 2022-12-22 Carl Zeiss Meditec Ag MULTI-TASK LEARNING OF WHITE LIGHT PHOTOGRAPHS FOR A SURGICAL MICROSCOPE
DE112021008418T5 (en) * 2021-11-02 2024-08-22 Leica Microsystems Cms Gmbh Method for providing position information for locating a target position in a microscopic sample, method for examining and/or processing such a target position and means for carrying out these methods
WO2023108139A2 (en) 2021-12-10 2023-06-15 10X Genomics, Inc. Multi-resolution in situ decoding
WO2023172915A1 (en) 2022-03-08 2023-09-14 10X Genomics, Inc. In situ code design methods for minimizing optical crowding
WO2023192302A1 (en) 2022-03-29 2023-10-05 10X Genomics, Inc. Spectral unmixing combined with decoding for super-multiplexed in situ analysis
KR102483745B1 (en) * 2022-04-06 2023-01-04 주식회사 포트래이 Spatial transcriptome analysis apparatus and method using the same
CN115602245B (en) * 2022-09-09 2023-10-03 郑州思昆生物工程有限公司 Method, device, equipment and storage medium for screening fluorescent images
CN116798523B (en) * 2023-06-01 2024-07-30 南京金域医学检验所有限公司 Pattern recognition and judgment system for anti-neutrophil cytoplasmic antibody

Family Cites Families (57)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP3224860B2 (en) 1992-06-19 2001-11-05 ポーラ化成工業株式会社 Microscope image evaluation system
US6007994A (en) * 1995-12-22 1999-12-28 Yale University Multiparametric fluorescence in situ hybridization
US5827660A (en) * 1996-08-09 1998-10-27 University Of Massachusetts Caged fluorochrome-labeled probes and subtraction of autoflourescence background
US6404906B2 (en) 1997-03-03 2002-06-11 Bacus Research Laboratories,Inc. Method and apparatus for acquiring and reconstructing magnified specimen images from a computer-controlled microscope
US5880473A (en) * 1997-07-28 1999-03-09 Applied Imaging, Inc. Multifluor-fluorescence in-situ hybridization (M-FISH) imaging techniques using multiple multiband filters with image registration
US5834203A (en) * 1997-08-25 1998-11-10 Applied Spectral Imaging Method for classification of pixels into groups according to their spectra using a plurality of wide band filters and hardwire therefore
US6900435B1 (en) * 1999-02-14 2005-05-31 Nanoptics, Inc. Deconvolving far-field images using scanned probe data
US6834122B2 (en) * 2000-01-22 2004-12-21 Kairos Scientific, Inc. Visualization and processing of multidimensional data using prefiltering and sorting criteria
US6995841B2 (en) * 2001-08-28 2006-02-07 Rice University Pulsed-multiline excitation for color-blind fluorescence detection
JP2003088399A (en) * 2001-09-14 2003-03-25 Fuji Photo Film Co Ltd Hybridization method and device and unit for biochemical analysis
JP3902939B2 (en) 2001-10-26 2007-04-11 株式会社日立ハイテクノロジーズ Apparatus and method for measuring minute area in specimen
AU2002357791B9 (en) 2001-12-05 2008-09-25 The J. David Gladstone Institutes Robotic microscopy systems
KR100437253B1 (en) * 2002-05-13 2004-06-23 학교법인 한양학원 Microarray copy image creation system and method thereof
EP1534114A2 (en) 2002-06-18 2005-06-01 Lifespan Biosciences, Inc. Computerized image capture of structures of interest within a tissue sample
EP1556506A1 (en) 2002-09-19 2005-07-27 The Chancellor, Masters And Scholars Of The University Of Oxford Molecular arrays and single molecule detection
FR2849196B1 (en) * 2002-12-23 2006-09-15 Imstar Image Et Modelisation S BIOPUCES-LIKE CHIP READER, AND ASSOCIATED METHODS
JP4111099B2 (en) * 2003-08-21 2008-07-02 ソニー株式会社 Biochemical analysis method and apparatus
US20050265588A1 (en) * 2004-02-03 2005-12-01 Bioimagene, Inc. Method and system for digital image based flourescent in situ hybridization (FISH) analysis
WO2006076432A2 (en) * 2005-01-11 2006-07-20 University Of Central Florida Interactive multiple gene expression map system
US20090264299A1 (en) 2006-02-24 2009-10-22 Complete Genomics, Inc. High throughput genome sequencing on DNA arrays
US20070031043A1 (en) 2005-08-02 2007-02-08 Perz Cynthia B System for and method of intelligently directed segmentation analysis for automated microscope systems
CA2659978A1 (en) 2006-08-04 2008-02-14 Ikonisys, Inc. Image processing method for a microscope system
WO2008073168A2 (en) 2006-08-25 2008-06-19 The Trustees Of Columbia University In The City Of New York Systems and methods for high-throughput radiation biodosimetry
CA2663286A1 (en) * 2006-09-15 2008-03-20 The Governors Of The University Of Alberta Automatedfishanalysis,circulating microfluidic chip, and method for immobilizing cells to a microfluidic chip
US8244021B2 (en) * 2006-12-20 2012-08-14 Ventana Medical Systems, Inc. Quantitative, multispectral image analysis of tissue specimens stained with quantum dots
WO2009059022A1 (en) * 2007-10-30 2009-05-07 Complete Genomics, Inc. Apparatus for high throughput sequencing of nucleic acids
US20100034444A1 (en) * 2008-08-07 2010-02-11 Helicos Biosciences Corporation Image analysis
US8488863B2 (en) * 2008-11-06 2013-07-16 Los Alamos National Security, Llc Combinational pixel-by-pixel and object-level classifying, segmenting, and agglomerating in performing quantitative image analysis that distinguishes between healthy non-cancerous and cancerous cell nuclei and delineates nuclear, cytoplasm, and stromal material objects from stained biological tissue materials
US20100330555A1 (en) * 2009-06-25 2010-12-30 General Electric Company Accuracy fluorescence in-situ hybridization assay of samples with apoptotic cells
US10353190B2 (en) * 2009-12-30 2019-07-16 Koninklijke Philips N.V. Sensor for microscopy
JP5497587B2 (en) * 2010-03-23 2014-05-21 株式会社日立ハイテクノロジーズ Microchannel chip and microarray chip
EP2518514B1 (en) 2011-04-29 2014-06-18 F. Hoffmann-La Roche AG A method for operating an automated sample workcell
WO2013148448A1 (en) * 2012-03-30 2013-10-03 Clarient Diagnostic Services, Inc. Immunoflu0rescence and fluorescent-based nucleic acid analysis on a single sample
CN102719357B (en) 2012-05-31 2014-07-09 博奥生物集团有限公司 Hybridization system for real-time monitoring parsing process of micro array chip
US9494521B2 (en) 2012-07-25 2016-11-15 Theranos, Inc. Image analysis and measurement of biological samples
JP5663541B2 (en) * 2012-09-19 2015-02-04 株式会社日立ハイテクノロジーズ Reaction vessel, parallel processing device, and sequencer
US9064304B2 (en) * 2013-03-18 2015-06-23 General Electric Company Image quality assessment of microscopy images
WO2015191978A1 (en) * 2014-06-12 2015-12-17 University Of Notre Dame Microfluidic devices, systems, and methods for imaging tissue samples
CN106715768B (en) * 2014-07-30 2020-06-16 哈佛学院院长及董事 System and method for assaying nucleic acids
CN107209934B (en) * 2014-12-03 2021-03-30 文塔纳医疗系统公司 Methods, systems and devices for quantitative analysis of heterogeneous biomarker distributions
WO2016115537A2 (en) 2015-01-15 2016-07-21 Massachusetts Institute Of Technology Systems, methods, and apparatus for in vitro single-cell identification and recovery
EP3251087B1 (en) * 2015-01-29 2018-11-28 Ventana Medical Systems, Inc. Dot detection, color classification of dots and counting of color classified dots
WO2016120442A1 (en) * 2015-01-30 2016-08-04 Ventana Medical Systems, Inc. Foreground segmentation and nucleus ranking for scoring dual ish images
US10495554B2 (en) * 2016-05-25 2019-12-03 The Board Of Trustees Of The Leland Stanford Junior University Method and system for imaging and analysis of a biological specimen
US10203491B2 (en) 2016-08-01 2019-02-12 Verily Life Sciences Llc Pathology data capture
CN116242687A (en) 2016-11-08 2023-06-09 哈佛学院院长及董事 Multiplex imaging using MERFISH, extended microscopy and related techniques
SG10201609334WA (en) 2016-11-08 2018-06-28 Delta Electronics Intl Singapore Pte Ltd Multi-Channel Fluorescence Detection Device
CN110088804B (en) * 2016-12-22 2023-06-27 文塔纳医疗系统公司 Computer scoring based on primary color and immunohistochemical images
US10573003B2 (en) 2017-02-13 2020-02-25 Amit Sethi Systems and methods for computational pathology using points-of-interest
EP3586181B1 (en) 2017-06-13 2021-09-22 Google LLC Augmented reality microscope for pathology
WO2018229549A2 (en) * 2017-06-16 2018-12-20 Nauto Global Limited System and method for digital environment reconstruction
US11222415B2 (en) * 2018-04-26 2022-01-11 The Regents Of The University Of California Systems and methods for deep learning microscopy
CN112424825A (en) * 2018-05-15 2021-02-26 文塔纳医疗系统公司 Quantification of signal in stained aggregates
JP2021531790A (en) * 2018-07-27 2021-11-25 ベンタナ メディカル システムズ, インコーポレイテッド System for automated in-situ hybrid formation analysis
EP3867868A1 (en) * 2018-10-15 2021-08-25 Ventana Medical Systems, Inc. Image enhancement to enable improved nuclei detection and segmentation
WO2020167632A1 (en) * 2019-02-13 2020-08-20 Ventana Medical Systems, Inc. Systems and methods for computing the contributions of autofluorescence in multichannel images
EP4078151A4 (en) 2019-12-17 2024-01-03 Applied Materials, Inc. System and method for acquisition and processing of multiplexed fluorescence in-situ hybridization images

Also Published As

Publication number Publication date
EP4078151A1 (en) 2022-10-26
JP7420950B2 (en) 2024-01-23
JP2023507595A (en) 2023-02-24
US20210183034A1 (en) 2021-06-17
KR20220110589A (en) 2022-08-08
WO2021127019A1 (en) 2021-06-24
CN114868010A (en) 2022-08-05
TWI835323B (en) 2024-03-11
TW202307420A (en) 2023-02-16
US20210182531A1 (en) 2021-06-17
US11630067B2 (en) 2023-04-18
EP4078151A4 (en) 2024-01-03
TWI782368B (en) 2022-11-01
US11783916B2 (en) 2023-10-10
US20210199584A1 (en) 2021-07-01
TW202138564A (en) 2021-10-16
US11624708B2 (en) 2023-04-11
US20210181111A1 (en) 2021-06-17

Similar Documents

Publication Publication Date Title
US20230253073A1 (en) System for acquisition and processing of multiplexed fluorescence in-situ hybridization images
JP6871970B2 (en) Optical distortion correction for imaging samples
CA2802420C (en) Method and apparatus for single-particle localization using wavelet analysis
EP2976747B1 (en) Image quality assessment of microscopy images
US10809514B2 (en) Low resolution slide imaging and slide label imaging and high resolution slide imaging using dual optical paths and a single imaging sensor
US10955650B2 (en) Two pass macro image
AU2014236055A1 (en) Referencing in multi-acquisition slide imaging
WO2014173547A1 (en) A method for automated platform and/or reference object independent acquisition of positional information and localization of objects of interest in a microscope
EP3884327B1 (en) Real-time focusing in a slide-scanning system
US20220064718A1 (en) Alignment beads for mfish
US20230230234A1 (en) Cell body segmentation using machine learning
US20220310202A1 (en) Utilization of sparce codebook in multiplexed fluorescent in-situ hybridization imaging

Legal Events

Date Code Title Description
STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

AS Assignment

Owner name: APPLIED MATERIALS, INC., CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:CHANG, YUN-CHING;XIE, DAN;KIM, CHLOE;SIGNING DATES FROM 20210922 TO 20210929;REEL/FRAME:064263/0874