[go: up one dir, main page]
More Web Proxy on the site http://driver.im/

US20090263017A1 - Method for reconstruction of pixel color values - Google Patents

Method for reconstruction of pixel color values Download PDF

Info

Publication number
US20090263017A1
US20090263017A1 US12/386,779 US38677909A US2009263017A1 US 20090263017 A1 US20090263017 A1 US 20090263017A1 US 38677909 A US38677909 A US 38677909A US 2009263017 A1 US2009263017 A1 US 2009263017A1
Authority
US
United States
Prior art keywords
color
pixel
kernel
value
relatively
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US12/386,779
Inventor
Anthony Amir Tanbakuchi
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Individual
Original Assignee
Individual
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Individual filed Critical Individual
Priority to US12/386,779 priority Critical patent/US20090263017A1/en
Publication of US20090263017A1 publication Critical patent/US20090263017A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N25/00Circuitry of solid-state image sensors [SSIS]; Control thereof
    • H04N25/60Noise processing, e.g. detecting, correcting, reducing or removing noise
    • H04N25/68Noise processing, e.g. detecting, correcting, reducing or removing noise applied to defects
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/80Camera processing pipelines; Components thereof
    • H04N23/84Camera processing pipelines; Components thereof for processing colour signals
    • H04N23/843Demosaicing, e.g. interpolating colour pixel values
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N25/00Circuitry of solid-state image sensors [SSIS]; Control thereof
    • H04N25/10Circuitry of solid-state image sensors [SSIS]; Control thereof for transforming different wavelengths into image signals
    • H04N25/11Arrangement of colour filter arrays [CFA]; Filter mosaics
    • H04N25/13Arrangement of colour filter arrays [CFA]; Filter mosaics characterised by the spectral characteristics of the filter elements
    • H04N25/134Arrangement of colour filter arrays [CFA]; Filter mosaics characterised by the spectral characteristics of the filter elements based on three different wavelength filter elements

Definitions

  • the present invention relates to a method and apparatus for object-based color reconstruction in a multi-color based sensor arrangement through maintaining coherence within objects and along edges, and a computer program and computer program product for controlling such method, and an image sensing facility comprising such apparatus.
  • the invention relates to a method for executing object-based color reconstruction in a multi-color matrix-based sensor arrangement that estimates color value through determining local gradients among various luminance component values assessed.
  • Multi-color matrix-based sensor arrangements that include color sensors that have one first luminance component sensed at a relatively higher spatial frequency and two further chrominance components sensed at relatively lower spatial frequencies are coming in rapidly expanding use.
  • Fields of application include digital cameras, digital cinematography, video cameras, scientific imaging, broadcast cameras, surveillance, security monitoring, and various others.
  • each object point translates into a single-color of the corresponding image pixel.
  • each pixel lacks two of the three sensor colors
  • reconstruction of three-color pixels requires an appreciable amount of processing.
  • the resulting reconstructed array often had artificially colored, noise-enhanced, or blurred edges.
  • the present embodiment generally utilizes only a small data kernel of optimally 5 ⁇ 5 pixels for Green, or minimally 3 ⁇ 3 pixels for Red and Blue. It will generally minimize perceptional error, it reconstructs frequencies above the Nyquist frequency, it has much lower color aliasing than many other algorithms, it will hardly expand existing defects, and it will only be marginally affected by image noise.
  • United States Patent Application Publication 2002/0063789 A1 to Acharya et al and published on May 30, 2002 discloses a Color Filter Array and Color Interpolation Algorithm, and uses a classification to determine which pixels it will use in the interpolation.
  • the present invention in contradistinction uses a soft decision method that is useful for adaptive adjusting for variations in system noise.
  • the reference singles out a very particular matrix design with a very high percentage of 75% Green pixels sensed. The present invention is much wider applicable.
  • the inventor has found that correct application of the algorithm will not remove fundamental signal noise. Rather, it will prevent enhancing of the noise.
  • the inventor has recognized the corresponding behavior of the gradients among the various sensed colors.
  • a method for object-based color reconstruction in a multi-color matrix-based sensor arrangement is based on color sensors that have one first color sensed at a relatively higher spatial frequency and two further colors sensed at relatively lower spatial frequencies.
  • the method executes the following steps.
  • a particular pixel not sensed in said first color has its first color value estimated through determining local gradients ( 82 ) among various first color values assessed, and in accordance with such gradients such estimating is executed through interpolating ( 84 , 86 ) along relatively stronger edge informations.
  • For a particular pixel not sensed in a particular further color that further color's value is estimated in a direction along with relatively smaller differences evaluated in the first color ( 94 , 96 ).
  • the method has the ability for limiting noise propagation.
  • a method includes a first process for a first pixel and a second process for the first pixel.
  • the first process includes extracting a first kernel from a multi-color matrix, generating first variance weights from the first kernel, and generating a first color based on the first variance weights and adjacent pixel values of the first color.
  • the second process includes extracting a second kernel from the multi-color matrix, generating second variance offsets from the second kernel, and generating a second color based on the second variance offsets and an adjacent pixel of the second color.
  • the invention also relates to a computing machine that includes a processor and a memory.
  • the memory stores a first module for controlling the processor to perform a first process for a first pixel.
  • the first process includes extracting a first kernel from a multi-color matrix, generating first variance weights from the first kernel, and generating a first color based on the first variance weights and adjacent pixel values of the first color.
  • the memory further stores a second module for controlling the processor to perform a second process for the first pixel.
  • the second process includes extracting a second kernel from the multi-color matrix, generating second variance offsets from the second kernel, and generating a second color based on the second variance offsets and an adjacent pixel of the second color.
  • the invention also relates to a computer readable medium having stored thereon a plurality of modules for controlling a processor.
  • the plurality of modules includes a first module and a second module.
  • the first module is for controlling the processor to perform a first process for the first pixel.
  • the first process includes extracting a first kernel from a multi-color matrix, generating first variance weights from the first kernel, and generating a first color based on the first variance weights and adjacent pixel values of the first color.
  • the second module is for controlling the processor to perform a second process for the first pixel.
  • the second process includes extracting a second kernel from the multi-color matrix, generating second variance offsets from the second kernel, and generating a second color based on the second variance offsets and an adjacent pixel of the second color.
  • FIG. 1 is a schematic view of a Bayer matrix of image points
  • FIG. 2 is a schematic view of the deriving of the horizontal Green gradient
  • FIG. 3 is a schematic view of the deriving of the vertical Green gradient
  • FIGS. 4 a , 4 b are schematic views of the relevance of the eight possible Green directions
  • FIG. 5 is a graph of a filter characteristic diagram for removing false color responses.
  • FIG. 6 is a schematic block diagram of a hardware implementation of the present invention.
  • FIG. 1 illustrates a Bayer matrix of image points for use with the present invention.
  • other arrangements could be used, such as through using subtractive colors, such as cyan, magenta, and yellow, or through various other published embodiments.
  • the matrix has Green pixels arranged according to a chessboard organization. The remaining pixels have been used for Red pixels in odd rows and Blue pixels in even rows.
  • the intended end result of the reconstruction embodiment is three complete images of Red, Green, and Blue pixel values, respectively, where the original recording did yield less than three values for each respective pixel, according to the pixel pattern for the respective single color. It is known from advance that at each pixel location, two values must be estimated. In practice, the application most favored is to a Bayer pattern wherein the green pixels form a chessboard pattern, while the non-green fields have alternating red and blue values in the diagonal directions.
  • FIG. 2 illustrates the deriving of the horizontal Green gradient or horizontal green edge indicator ⁇ H in a preferred kernel pattern of 5 ⁇ 5 pixels according to the average absolute differences between pixels pairs joined by an arrow, according to:
  • ⁇ H
  • FIG. 3 illustrates the deriving of the vertical Green gradient or vertical Green edge indicator ⁇ V .
  • the pixels assigned to other colors will be ignored.
  • a local edge direction is determined and a Green value for the missing central point is estimated as follows: if ⁇ V is relatively larger, interpolate horizontally between two nearest green neighbors; if ⁇ H is relatively larger, interpolate vertically between two nearest green neighbors, and if ⁇ H and ⁇ V are roughly equal, interpolate linearly between all four nearest neighbors of the missing Green pixel.
  • ⁇ V is relatively larger, interpolate horizontally between two nearest green neighbors
  • ⁇ H and ⁇ V are roughly equal, interpolate linearly between all four nearest neighbors of the missing Green pixel.
  • G i,j 1/2 ⁇ H ( G i,j ⁇ 1 +G i,j+1 )+1/2 ⁇ v ( G i ⁇ 1,j +G i+1,j ),
  • ⁇ H 1 ⁇ k H /( ⁇ k H + ⁇ k V ), and likewise, for the vertical parameter.
  • exponent constant k increases the algorithm's sensitivity to the differences in ⁇ H and ⁇ V .
  • the algebraic expressions offer more benefits than just for presentation.
  • the non-linear combination of the edge indicators can be adjusted with the value of k to either emphasize edge clarity in low-noise situations, or rather avoid the enhancing of the noise while still maintaining minimal color aliasing in high-noise situations. Therefore, k has a two-sided adaptive effect on the algorithm. With small values (between 1 and 2) noise will be adaptively minimized whilst ensuring that edges will have none to minimal color aliasing. On the other hand, with larger values for k in low-noise situations, more emphasis can be put on removing all color aliasing whilst enhancing edge continuity and sharpness. Further, the use of five arrows in FIGS.
  • the missing points in the Red and Blue images will be estimated.
  • the horizontal neighbors are Blue and the vertical neighbors are Red.
  • the Red value is sought for.
  • the vertically closest Green value to the actual pixel is determined and then its associated Red value copied to the actual pixel.
  • the horizontally closest Green value to the actual pixel is determined and then its associated Blue value copied to the actual pixel.
  • a corresponding version applies to the missing Blue values at pixels that were sampled Red.
  • the color reconstruction according to the present invention's embodiment will generally be correct, the particular arrangement of the Bayer matrix, in combination with non-ideal optical image low-pass filtering and system noise can cause some false color responses. Heuristically, the following description is given.
  • the sampling of a scene through a Bayer matrix will cause two major sampling errors, as follows: (a) The relative displacement between green, red and blue samples introduces a phase error between the various colors; (b) The sampling frequencies of Red and Blue are relatively lower than that of Green. This is perceptually advantageous, because the human eye's chroma and hue resolution is relatively less than its luminance resolution.
  • C ′ ( R ⁇ G ) i,j
  • C ′′ ( B ⁇ G ) i,j .
  • FIG. 5 shows the intended frequency response of a chroma filter.
  • the design of such a-filter that has been shown in solid lines is dependent on the system's optical low-pass filtering as well as on system noise. Ideally, it will be designed to minimize the loss of image information and at the same time maximize the removal of false colors. To an appreciable extent, degradation in sharpness through the alpha-filter is mitigated through a beta-filter shown in FIG. 5 in dashed lines.
  • FIG. 6 illustrates a hardware implementation of the present invention.
  • the output of the co-pending application referred to supra may be fed to a DSP pipeline 20 for processing and/or transient storage, as the case may be.
  • the DSP pipeline is fed by the image sensing facility, not shown for brevity, whereas the pipeline itself has a sufficient amount of memory and processing facilities for implementing the method according to the present invention.
  • the system will evaluate the noise level in the image, and therefrom in block 72 convert the noise level in the k-factor considered supra. From a holding element 74 , this k-factor is provided to such elements of the hardware as are in need of it.
  • command, control and synchronization streams have been left out of the embodiment for brevity and optimum clarification.
  • block 22 represents the captured Bayer image information.
  • block 60 it is detected whether the pixel in question is at least 2 pixels away from the image edge. If not, the procedure hereinafter will not have enough data available, so the process will then move to the next pixel in block 62 .
  • block 64 the system detects whether the sample pixel is Green.
  • the system will undertake to reconstruct an appropriate Green value for the pixel in question, as follows.
  • the system extracts in a kernel of 5 ⁇ 5 pixels disclosed earlier the sensed Green pixels.
  • the horizontal and vertical variance parameters are derived.
  • the horizontal and vertical weight generating functions are derived, through also receiving the actual value of the exponent k from block 74 , as discussed earlier.
  • the system calculates a product of the sum weights and the horizontally and vertically sampled Green pixels. With the so reconstructed Green pixel, the system then proceeds to mixer/accumulator 112 whilst in block 114 storing the result and in block 116 detecting whether the end of the image had been reached, which for the time being is not so.
  • the system shifts by one row and one column and checks for a Red pixel. If no, the pixel thus being Blue, the system will undertake to reconstruct a Red pixel.
  • the Red pixels are extracted in a 3 ⁇ 3 kernel as shown earlier, as well as the Green pixels corresponding thereto.
  • the system generates object-based parameters for sampled Red pixels in the kernel.
  • the system generates weights for the object-based parameter, whilst again using the actual value of exponent k received from block 74 .
  • a sum of the products of weights and measured Red pixels in the kernel generates the Red pixel value.
  • the system checks whether the sample pixel was Blue. If yes, the way to mixer facility 112 lies open. If no, the system reverts to the blue reconstructing column further to the right.
  • the various blocks 100 through 106 for this column correspond to their respective counterparts in the middle column, and further extensive discussion thereof is omitted. This column is also entered from block 68 if the sampled pixel were Red itself.
  • the answer in block 116 will become yes. This means that in block 118 the full RGB image will have been recovered; the remainder of the flow-chart implements a few largely cosmetic refinements.
  • the image is converted to an estimated color space with a chroma component applied.
  • a low pass chroma filter is used to substantially remove false colors.
  • the system executes a back conversion to the RGB color space. This will in block 126 yield the reconstructed RGB image that is subsequently presented to the DSP pipeline for further usage.
  • a method may be regarded as having first and second processes for a first pixel.
  • the first process computes a Green value for the pixel and the second process computes a Red value for the pixel.
  • the first process includes extracting a first kernel from a multi-color matrix.
  • the first kernel is a 5 by 5 sub matrix of the Bayer pattern centered on the exemplary Blue pixel.
  • the first process further includes generating first variance weights (e.g., ⁇ H , ⁇ V ) from the first kernel and then generating a first color (Green, in this example) based on the first variance weights and adjacent pixel values of the first color.
  • first variance weights e.g., ⁇ H , ⁇ V
  • Green Green, in this example
  • the generating of the first variance weights includes first determining horizontal and vertical gradient value averages (e.g., ⁇ H and ⁇ V ).
  • the center of the first kernel is the exemplary Blue pixel, and the kernel includes first and second side pixels of the first color spaced apart horizontally.
  • the first color is Green
  • the first and second side pixels are Green pixels of the kernel that are spaced apart horizontally.
  • the horizontal gradient value average e.g., ⁇ H
  • the first process determines plural horizontal gradients (e.g., pairs of two horizontally spaced Green pixels) and then determines a gradient value for each gradient.
  • a first horizontal gradient value is determined by calculating an absolute value of a difference between a value of the first side pixel (a Green pixel in this example) and a value of the second side pixel (another Green pixel in this example). The calculation of the horizontal gradient value is repeated for each horizontal gradient within the first kernel. Then, the first process calculates an average of the plural horizontal gradient values to determine the horizontal gradient value average (e.g., ⁇ H ). For example, five horizontal gradient values are determined for the gradients depicted in FIG. 2 . Then, the average is determined by summing the values and dividing by 5. The vertical gradient value average (e.g., ⁇ V ) is computed in a similar way. See FIG. 3 .
  • the first variance weights include horizontal and vertical interpolation weights (e.g., ⁇ H , ⁇ V ).
  • the first process determines the horizontal and vertical interpolation weights based on the horizontal and vertical gradient value averages and a predetermined exponent value, k.
  • the horizontal interpolation weight is calculated based on
  • the vertical interpolation weight is calculated based on
  • ⁇ V 1 ⁇ k V ( ⁇ k H + ⁇ k V ),
  • ⁇ k H is the horizontal gradient value average
  • ⁇ k V is the vertical gradient value average
  • k is the predetermined exponent value
  • the second process includes extracting a second kernel from the multi-color matrix.
  • the second kernel is a 3 by 3 sub matrix of the Bayer pattern centered on the exemplary Blue pixel.
  • the second process further includes generating second variance offsets
  • ⁇ g 2 G i ⁇ 1,j+1 ⁇ G i,j
  • ⁇ g 4 G i+1,j+1 ⁇ G i,j
  • ⁇ g 6 G i+1,j ⁇ 1 ⁇ G i,j
  • ⁇ g 8 G i ⁇ 1,j ⁇ 1 ⁇ G i,j ).
  • the second variance offsets are generated from the second kernel.
  • the second kernel is a 3 by 3 sub matrix of the Bayer pattern centered on the exemplary Blue pixel.
  • a Green value has been calculated to become associated with the exemplary Blue pixel, and associated with each of the four Red pixels at the four corners of the second kernel, as discussed herein. Also see FIG. 1 .
  • second variance offsets e.g., ⁇ g 2 , ⁇ g 4 , ⁇ g 6 , and ⁇ g 8
  • the second color e.g., Red in this example
  • generating the second variance offsets includes determining diagonal gradients between pixels in the second kernel (e.g., [i, j] to [i ⁇ 1, j+1], [i, j] to [i+1, j+1], [i, j] to [i+1, j ⁇ 1] and [i, j] to [i ⁇ 1, j ⁇ 1]) and then determining gradient values of the first color (Green in this example)
  • ⁇ g 2 G i ⁇ 1,j+1 ⁇ G i,j
  • ⁇ g 4 G i+1,j+1 ⁇ G i,j
  • ⁇ g 6 G i+1,j ⁇ 1 ⁇ G i,j
  • ⁇ g 8 G i ⁇ 1,j ⁇ 1 ⁇ G i,j ).
  • generating a second color includes choosing a minimum value of the gradient values (e.g., ⁇ g 2 ), selecting the adjacent pixel (e.g., [i ⁇ 1, j+1] to the right and above the center of the second kernel as depicted in FIG.
  • the second color (Red, in this example) based on a gradient that corresponds to the minimum value, and subtracting the minimum value (e.g., ⁇ g 2 ) from a pixel value (e.g., R i ⁇ 1,j+1 ) of the selected adjacent pixel to determine the value for the second color (e.g., R i,j ) at the center of the second kernel (e.g., the exemplary Blue pixel).
  • R i,j ( R i+1,j+1 ⁇ G i+1,j+1 +G i,j ),
  • R i,j ( R i+1,j ⁇ 1 ⁇ G i+1,j ⁇ +G i,j ),
  • R i,j ( R i ⁇ 1,j ⁇ 1 G i ⁇ 1,j ⁇ 1 +G i,j ).
  • the second process would determine a Blue value for the center pixel.
  • the method repeats for every pixel (except pixels on the edge of the matrix where a full kernel cannot be extracted) so as to determine all three color planes for each pixel.
  • the method further includes a third process.
  • the third process includes extracting a third kernel (e.g., a 3 by 3 sub matrix centered on the second pixel) from the multi-color matrix (e.g., the Bayer multi-color matrix described in this example).
  • the third process further includes generating third variance offsets from the third kernel (e.g., ⁇ g 2 , ⁇ g 4 , ⁇ g 6 , and ⁇ g 8 ) in a process substantially the same as is described above with respect to the second process, and generating a third color (Blue in this example) based on the third variance offsets and an adjacent pixel of the third color in a process substantially the same as is described above with respect to the second process.
  • the third kernel e.g., ⁇ g 2 , ⁇ g 4 , ⁇ g 6 , and ⁇ g 8
  • the value of the predetermined exponent, k is determined in step 72 ( FIG. 6 ).
  • the system noise level 70 is a matrix of elements corresponding to the elements of the captured Bayer image data, but representing system noise level. For example, a totally dark image could be captured in the sensor and loaded into matrix 70 , or an imaging device could contain a group of light shielded pixels used to determine k.
  • An exemplary way in which k may be determined is to average the noise value over all elements to determine an overall average, and then compute a variance to this average over all elements of the matrix. The variance becomes input to a look up table where the value of k is read as an output. In this way, different loads of the look up table can be used to emphasize edge clarity in low-noise situations and avoid enhancing noise while maintaining minimal color aliasing in high-noise situations.
  • the image in RGB space is converted to chroma space (C′ and C′′ as discussed above) in block 120 .
  • the chroma components e.g., C′ and C′′
  • a low pass chroma filter e.g., such as a moving window of a 3 by 3 kernel of elements where the chroma value of the center point takes up the average value over the 3 by 3 kernel.
  • the above described method is implemented in a computing machine of some form that includes a processor and a memory.
  • the memory stores a first module for controlling the processor to perform a first process for a first pixel.
  • the first process includes extracting a first kernel from a multi-color matrix, generating first variance weights from the first kernel, and generating a first color based on the first variance weights and adjacent pixel values of the first color.
  • the memory further stores a second module for controlling the processor to perform a second process for the first pixel.
  • the second process includes extracting a second kernel from the multi-color matrix, generating second variance offsets from the second kernel, and generating a second color based on the second variance offsets and an adjacent pixel of the second color.
  • the processor may be a programmable processor that is controlled by programming modules stored on a separate computer readable media or it may be a fixed designed that is controlled by the timing generator of the processor.
  • the computer readable media has stored thereon a plurality of modules for controlling a processor.
  • the plurality of modules include first and second modules.
  • the first module controls the processor to perform a first process for a first pixel.
  • the first process includes extracting a first kernel from a multi-color matrix, generating first variance weights from the first kernel, and generating a first color based on the first variance weights and adjacent pixel values of the first color.
  • the second module controls the processor to perform a second process for the first pixel.
  • the second process includes extracting a second kernel from the multi-color matrix, generating second variance offsets from the second kernel, and generating a second color based on the second variance offsets and an adjacent pixel of the second color.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Physics & Mathematics (AREA)
  • Spectroscopy & Molecular Physics (AREA)
  • Color Television Image Signal Generators (AREA)

Abstract

A method of color reconstruction includes a first process for a first pixel and a second process for the first pixel. The first process includes extracting a first kernel from a multi-color matrix, generating first variance weights from the first kernel, and generating a first color based on the first variance weights and adjacent pixel values of the first color. The second process includes extracting a second kernel from the multi-color matrix, generating second variance offsets from the second kernel, and generating a second color based on the second variance offsets and an adjacent pixel of the second color.

Description

  • The priority benefits of the Oct. 10, 2002 filing dates of U.S. provisional applications Ser. Nos. 60/417,142 and 60/417,152 are hereby claimed.
  • BACKGROUND OF THE INVENTION
  • 1. Field of the Invention
  • The present invention relates to a method and apparatus for object-based color reconstruction in a multi-color based sensor arrangement through maintaining coherence within objects and along edges, and a computer program and computer program product for controlling such method, and an image sensing facility comprising such apparatus. In particular, the invention relates to a method for executing object-based color reconstruction in a multi-color matrix-based sensor arrangement that estimates color value through determining local gradients among various luminance component values assessed.
  • 2. Description of Related Art
  • Multi-color matrix-based sensor arrangements that include color sensors that have one first luminance component sensed at a relatively higher spatial frequency and two further chrominance components sensed at relatively lower spatial frequencies are coming in rapidly expanding use. Fields of application include digital cameras, digital cinematography, video cameras, scientific imaging, broadcast cameras, surveillance, security monitoring, and various others. Now generally, each object point translates into a single-color of the corresponding image pixel. Inasmuch as, in the resulting image each pixel lacks two of the three sensor colors, reconstruction of three-color pixels requires an appreciable amount of processing. However, in earlier realizations, the resulting reconstructed array often had artificially colored, noise-enhanced, or blurred edges. In consequence, there exists a growing need to provide for an algorithmic procedure that will provide high-quality estimations for the color values of pixels that were originally sensed in a single color only, whilst requiring to access only a minimal memory facility and minimal power consumption, needing to access only a small number of image lines/columns at a time, and being independent of other DSP-based (i.e., Digital Signal Processing-based) algorithms, such as those for removing false colors or correcting color balance, overall influencing of hue or luminance, or others. As will be discussed more extensively hereinafter, the present embodiment generally utilizes only a small data kernel of optimally 5×5 pixels for Green, or minimally 3×3 pixels for Red and Blue. It will generally minimize perceptional error, it reconstructs frequencies above the Nyquist frequency, it has much lower color aliasing than many other algorithms, it will hardly expand existing defects, and it will only be marginally affected by image noise.
  • United States Patent Application Publication 2002/0063789 A1 to Acharya et al and published on May 30, 2002 discloses a Color Filter Array and Color Interpolation Algorithm, and uses a classification to determine which pixels it will use in the interpolation. The present invention in contradistinction uses a soft decision method that is useful for adaptive adjusting for variations in system noise. Moreover, the reference singles out a very particular matrix design with a very high percentage of 75% Green pixels sensed. The present invention is much wider applicable.
  • Moreover, the inventor has found that correct application of the algorithm will not remove fundamental signal noise. Rather, it will prevent enhancing of the noise. In particular, the inventor has recognized the corresponding behavior of the gradients among the various sensed colors.
  • SUMMARY OF THE INVENTION
  • In consequence, amongst other things, it is an object of the present invention to use such corresponding behavior to attain an improved result against an investment of only a limited amount of processing complexity.
  • By itself, co-pending patent application by the present inventor, titled “Method And Apparatus For Adaptive Pixel Correction Of A Multi-Color Matrix,” identified by Attorney Docket No. 12546 and assigned to the same assignee as the present application relates to the adaptive correcting of defective pixels and is incorporated herein by reference. The adaptive pixel correction of a multi-color matrix is based on soft decisions across the various color planes, and in particular based on taking into account the spatial response pattern of the human visual system. The result of the co-pending application may be taken as a starting point for applying the present invention, but such represents no express restriction.
  • A method for object-based color reconstruction in a multi-color matrix-based sensor arrangement is based on color sensors that have one first color sensed at a relatively higher spatial frequency and two further colors sensed at relatively lower spatial frequencies.
  • In particular, the method executes the following steps. A particular pixel not sensed in said first color has its first color value estimated through determining local gradients (82) among various first color values assessed, and in accordance with such gradients such estimating is executed through interpolating (84, 86) along relatively stronger edge informations. For a particular pixel not sensed in a particular further color that further color's value is estimated in a direction along with relatively smaller differences evaluated in the first color (94, 96). Advantageously, the method has the ability for limiting noise propagation.
  • Now therefore, according to one of its aspects the invention a method includes a first process for a first pixel and a second process for the first pixel. The first process includes extracting a first kernel from a multi-color matrix, generating first variance weights from the first kernel, and generating a first color based on the first variance weights and adjacent pixel values of the first color. The second process includes extracting a second kernel from the multi-color matrix, generating second variance offsets from the second kernel, and generating a second color based on the second variance offsets and an adjacent pixel of the second color.
  • The invention also relates to a computing machine that includes a processor and a memory. The memory stores a first module for controlling the processor to perform a first process for a first pixel. The first process includes extracting a first kernel from a multi-color matrix, generating first variance weights from the first kernel, and generating a first color based on the first variance weights and adjacent pixel values of the first color. The memory further stores a second module for controlling the processor to perform a second process for the first pixel. The second process includes extracting a second kernel from the multi-color matrix, generating second variance offsets from the second kernel, and generating a second color based on the second variance offsets and an adjacent pixel of the second color.
  • The invention also relates to a computer readable medium having stored thereon a plurality of modules for controlling a processor. The plurality of modules includes a first module and a second module. The first module is for controlling the processor to perform a first process for the first pixel. The first process includes extracting a first kernel from a multi-color matrix, generating first variance weights from the first kernel, and generating a first color based on the first variance weights and adjacent pixel values of the first color. The second module is for controlling the processor to perform a second process for the first pixel. The second process includes extracting a second kernel from the multi-color matrix, generating second variance offsets from the second kernel, and generating a second color based on the second variance offsets and an adjacent pixel of the second color.
  • Further advantageous aspects of the invention are as disclosed and claimed herein.
  • BRIEF DESCRIPTION OF DRAWINGS
  • The invention will be described in detail in the following description of preferred embodiments with reference to the following figures wherein:
  • FIG. 1 is a schematic view of a Bayer matrix of image points;
  • FIG. 2 is a schematic view of the deriving of the horizontal Green gradient;
  • FIG. 3 is a schematic view of the deriving of the vertical Green gradient;
  • FIGS. 4 a, 4 b are schematic views of the relevance of the eight possible Green directions;
  • FIG. 5 is a graph of a filter characteristic diagram for removing false color responses; and
  • FIG. 6 is a schematic block diagram of a hardware implementation of the present invention.
  • DETAILED DESCRIPTION OF PREFERRED EMBODIMENTS
  • FIG. 1 illustrates a Bayer matrix of image points for use with the present invention. Alternatively, other arrangements could be used, such as through using subtractive colors, such as cyan, magenta, and yellow, or through various other published embodiments. As shown, the matrix has Green pixels arranged according to a chessboard organization. The remaining pixels have been used for Red pixels in odd rows and Blue pixels in even rows.
  • Now, the intended end result of the reconstruction embodiment is three complete images of Red, Green, and Blue pixel values, respectively, where the original recording did yield less than three values for each respective pixel, according to the pixel pattern for the respective single color. It is known from advance that at each pixel location, two values must be estimated. In practice, the application most favored is to a Bayer pattern wherein the green pixels form a chessboard pattern, while the non-green fields have alternating red and blue values in the diagonal directions.
  • Therefore, first the Green matrix will be completed. In this respect, FIG. 2 illustrates the deriving of the horizontal Green gradient or horizontal green edge indicator δH in a preferred kernel pattern of 5×5 pixels according to the average absolute differences between pixels pairs joined by an arrow, according to:

  • δH ={|G i−1,j−2 −G i−1,j |+|G i−1,j −G i−1,j+2|+ . . . }/5.
  • Likewise, FIG. 3 illustrates the deriving of the vertical Green gradient or vertical Green edge indicator δV. For the time being, the pixels assigned to other colors will be ignored. In certain alternative environments, it may be advantageous to define the kernel as having a size of 3×3 pixels only, through using the two arrows numbered 23 in FIG. 2, and 27 in FIG. 3, respectively. Apparently, the respective kernels for horizontal and vertical determinations will not anymore coincide here.
  • Then, in a relatively simplistic implementation a local edge direction is determined and a Green value for the missing central point is estimated as follows: if δV is relatively larger, interpolate horizontally between two nearest green neighbors; if δH is relatively larger, interpolate vertically between two nearest green neighbors, and if δH and δV are roughly equal, interpolate linearly between all four nearest neighbors of the missing Green pixel. Of course, the determining of the being “roughly” equal allows to introduce a certain amount of flexibility into the algorithm. The logical method presented here demonstrates the functioning of the actual algebraic method.
  • A fuller realization as an algebraic expression for the interpolated green pixel is:

  • G i,j=1/2λH(G i,j−1 +G i,j+1)+1/2λv(G i−1,j +G i+1,j),
  • wherein λH=1−δk H/(δk Hk V), and likewise, for the vertical parameter. Herein, a greater value of exponent constant k increases the algorithm's sensitivity to the differences in δH and δV.
  • The algebraic expressions offer more benefits than just for presentation. The non-linear combination of the edge indicators can be adjusted with the value of k to either emphasize edge clarity in low-noise situations, or rather avoid the enhancing of the noise while still maintaining minimal color aliasing in high-noise situations. Therefore, k has a two-sided adaptive effect on the algorithm. With small values (between 1 and 2) noise will be adaptively minimized whilst ensuring that edges will have none to minimal color aliasing. On the other hand, with larger values for k in low-noise situations, more emphasis can be put on removing all color aliasing whilst enhancing edge continuity and sharpness. Further, the use of five arrows in FIGS. 2 and 3, instead of the minimal number of one, allows reconstruction of frequencies up to twice the Nyquist frequency fN (i.e. the spatial sample frequency) in both vertical and horizontal directions. In case of the smaller kernels, the limit would lie at the Nyquist frequency.
  • After having completed the Green array Gi,j, the missing points in the Red and Blue images will be estimated. For a Green sample on a Green-Blue row, the horizontal neighbors are Blue and the vertical neighbors are Red. First the Red value is sought for. To ensure the continuity of the object recorded at the given point, the vertically closest Green value to the actual pixel is determined and then its associated Red value copied to the actual pixel. For Blue, among the two neighbors, the horizontally closest Green value to the actual pixel is determined and then its associated Blue value copied to the actual pixel.
  • For a Blue sample, one approach is to choose among the four diagonally adjacent Red values, whilst using the respective Green differences shown in FIG. 4 b, as follows:

  • R i,j=(R i−1,j+1 −G i−1,j+1 +G i,j) if Δg2min
  • i.e. add the minimum Green difference to the Red value, and correspondingly:

  • R i,j=(R i+1,j+1 −G i+1,j+1 +G i,j) if Δg4min

  • R i,j=(R i+1,j−1 −G i+1,j−1 +G i,j) if Δg6min

  • R i,j=(R i−1,j−1 −G i−1,j−1 +G i,j) if Δg8min
  • A corresponding version applies to the missing Blue values at pixels that were sampled Red.
  • Now, although the color reconstruction according to the present invention's embodiment will generally be correct, the particular arrangement of the Bayer matrix, in combination with non-ideal optical image low-pass filtering and system noise can cause some false color responses. Heuristically, the following description is given. The sampling of a scene through a Bayer matrix will cause two major sampling errors, as follows: (a) The relative displacement between green, red and blue samples introduces a phase error between the various colors; (b) The sampling frequencies of Red and Blue are relatively lower than that of Green. This is perceptually advantageous, because the human eye's chroma and hue resolution is relatively less than its luminance resolution. On the other hand, this introduces different aliasing responses for Green versus for Red/Blue, in particular, inasmuch as Green will attain a better high-frequency response. Although this may cause false-color spikes and the like, such phenomenon is relatively rare. Such spikes can be considered high-frequency chromatic points, and low-pass filtering would generally remove them. In fact, a chroma image can be estimated through two Green differences

  • C′=(R−G)i,j, and C″=(B−G)i,j.
  • In this respect, FIG. 5 shows the intended frequency response of a chroma filter. The design of such a-filter that has been shown in solid lines is dependent on the system's optical low-pass filtering as well as on system noise. Ideally, it will be designed to minimize the loss of image information and at the same time maximize the removal of false colors. To an appreciable extent, degradation in sharpness through the alpha-filter is mitigated through a beta-filter shown in FIG. 5 in dashed lines.
  • FIG. 6 illustrates a hardware implementation of the present invention. Now, the output of the co-pending application referred to supra may be fed to a DSP pipeline 20 for processing and/or transient storage, as the case may be. The DSP pipeline is fed by the image sensing facility, not shown for brevity, whereas the pipeline itself has a sufficient amount of memory and processing facilities for implementing the method according to the present invention. In addition thereto, in block 70, the system will evaluate the noise level in the image, and therefrom in block 72 convert the noise level in the k-factor considered supra. From a holding element 74, this k-factor is provided to such elements of the hardware as are in need of it. Furthermore, for simplicity, command, control and synchronization streams have been left out of the embodiment for brevity and optimum clarification.
  • Now, block 22 represents the captured Bayer image information. In block 60, it is detected whether the pixel in question is at least 2 pixels away from the image edge. If not, the procedure hereinafter will not have enough data available, so the process will then move to the next pixel in block 62. In block 64, the system detects whether the sample pixel is Green.
  • If not, the system will undertake to reconstruct an appropriate Green value for the pixel in question, as follows. First, in block 80, the system extracts in a kernel of 5×5 pixels disclosed earlier the sensed Green pixels. In block 82, the horizontal and vertical variance parameters are derived. In block 84, the horizontal and vertical weight generating functions are derived, through also receiving the actual value of the exponent k from block 74, as discussed earlier. Finally, in block 86, the system calculates a product of the sum weights and the horizontally and vertically sampled Green pixels. With the so reconstructed Green pixel, the system then proceeds to mixer/accumulator 112 whilst in block 114 storing the result and in block 116 detecting whether the end of the image had been reached, which for the time being is not so.
  • If in block 64 the sample is not Green, so in block 66 the system shifts by one row and one column and checks for a Red pixel. If no, the pixel thus being Blue, the system will undertake to reconstruct a Red pixel. First, in block 90, the Red pixels are extracted in a 3×3 kernel as shown earlier, as well as the Green pixels corresponding thereto. Next, in block 92, the system generates object-based parameters for sampled Red pixels in the kernel. Next, in block in block 94, the system generates weights for the object-based parameter, whilst again using the actual value of exponent k received from block 74. Finally, in block 96, a sum of the products of weights and measured Red pixels in the kernel generates the Red pixel value. Finally, in block 110, the system checks whether the sample pixel was Blue. If yes, the way to mixer facility 112 lies open. If no, the system reverts to the blue reconstructing column further to the right. The various blocks 100 through 106 for this column correspond to their respective counterparts in the middle column, and further extensive discussion thereof is omitted. This column is also entered from block 68 if the sampled pixel were Red itself.
  • Generally, after numerous rounds through the reconstructing columns, the answer in block 116 will become yes. This means that in block 118 the full RGB image will have been recovered; the remainder of the flow-chart implements a few largely cosmetic refinements. In block 120, the image is converted to an estimated color space with a chroma component applied. In block 122, a low pass chroma filter is used to substantially remove false colors. In block 124, the system executes a back conversion to the RGB color space. This will in block 126 yield the reconstructed RGB image that is subsequently presented to the DSP pipeline for further usage.
  • As an example of the above described embodiment, a method may be regarded as having first and second processes for a first pixel. For example, in the Bayer patterned multi-color matrix, if the first pixel is a Blue pixel, then the first process computes a Green value for the pixel and the second process computes a Red value for the pixel. The first process includes extracting a first kernel from a multi-color matrix. In the example discussed herein, the first kernel is a 5 by 5 sub matrix of the Bayer pattern centered on the exemplary Blue pixel. The first process further includes generating first variance weights (e.g., λH, λV) from the first kernel and then generating a first color (Green, in this example) based on the first variance weights and adjacent pixel values of the first color.
  • The generating of the first variance weights includes first determining horizontal and vertical gradient value averages (e.g., δH and δV). The center of the first kernel is the exemplary Blue pixel, and the kernel includes first and second side pixels of the first color spaced apart horizontally. In this example, the first color is Green, and the first and second side pixels are Green pixels of the kernel that are spaced apart horizontally. To determine the horizontal gradient value average (e.g., δH), the first process determines plural horizontal gradients (e.g., pairs of two horizontally spaced Green pixels) and then determines a gradient value for each gradient. A first horizontal gradient value is determined by calculating an absolute value of a difference between a value of the first side pixel (a Green pixel in this example) and a value of the second side pixel (another Green pixel in this example). The calculation of the horizontal gradient value is repeated for each horizontal gradient within the first kernel. Then, the first process calculates an average of the plural horizontal gradient values to determine the horizontal gradient value average (e.g., δH). For example, five horizontal gradient values are determined for the gradients depicted in FIG. 2. Then, the average is determined by summing the values and dividing by 5. The vertical gradient value average (e.g., δV) is computed in a similar way. See FIG. 3.
  • The first variance weights include horizontal and vertical interpolation weights (e.g., λH, λV). The first process determines the horizontal and vertical interpolation weights based on the horizontal and vertical gradient value averages and a predetermined exponent value, k. The horizontal interpolation weight is calculated based on

  • λH=1−δk H/(δk Hk V),and
  • the vertical interpolation weight is calculated based on

  • λV=1−δk Vk Hk V),
  • where δk H is the horizontal gradient value average, δk V is the vertical gradient value average, and k is the predetermined exponent value.
  • The second process includes extracting a second kernel from the multi-color matrix. In the example discussed herein, the second kernel is a 3 by 3 sub matrix of the Bayer pattern centered on the exemplary Blue pixel. The second process further includes generating second variance offsets

  • (e.g., Δg 2 =G i−1,j+1 −G i,j , Δg 4 =G i+1,j+1 −G i,j , Δg 6 =G i+1,j−1 −G i,j, and Δg 8 =G i−1,j−1 −G i,j).
  • The second variance offsets are generated from the second kernel. In the example discussed herein, the second kernel is a 3 by 3 sub matrix of the Bayer pattern centered on the exemplary Blue pixel. As discussed with respect to the first process, a Green value has been calculated to become associated with the exemplary Blue pixel, and associated with each of the four Red pixels at the four corners of the second kernel, as discussed herein. Also see FIG. 1. With these calculated Green values, second variance offsets (e.g., Δg2, Δg4, Δg6, and Δg8) are determined based on the respective differences between the calculated Green values as depicted in FIGS. 4 a and 4 b. Then, the second color (e.g., Red in this example) is generated based on the second variance offsets and an adjacent pixel of the second color.
  • More specifically, generating the second variance offsets includes determining diagonal gradients between pixels in the second kernel (e.g., [i, j] to [i−1, j+1], [i, j] to [i+1, j+1], [i, j] to [i+1, j−1] and [i, j] to [i−1, j−1]) and then determining gradient values of the first color (Green in this example)

  • (e.g., Δg 2 =G i−1,j+1 −G i,j , Δg 4 =G i+1,j+1 −G i,j , Δg 6 =G i+1,j−1 −G i,j, and Δg 8 =G i−1,j−1 −G i,j).
  • corresponding to the gradients.
  • Then, generating a second color (Red, in this example) includes choosing a minimum value of the gradient values (e.g., Δg2), selecting the adjacent pixel (e.g., [i−1, j+1] to the right and above the center of the second kernel as depicted in FIG. 4 b) of the second color (Red, in this example) based on a gradient that corresponds to the minimum value, and subtracting the minimum value (e.g., Δg2) from a pixel value (e.g., Ri−1,j+1) of the selected adjacent pixel to determine the value for the second color (e.g., Ri,j) at the center of the second kernel (e.g., the exemplary Blue pixel).
  • However, if Δg4min then,

  • R i,j=(R i+1,j+1 −G i+1,j+1 +G i,j),
  • or if Δg6min then,

  • R i,j=(R i+1,j−1 −G i+1,j− +G i,j),
  • or if Δg8min then,

  • R i,j=(R i−1,j−1 G i−1,j−1 +G i,j).
  • On the other hand, if the center of the first and second kernels were a Red pixel, the second process would determine a Blue value for the center pixel.
  • The method repeats for every pixel (except pixels on the edge of the matrix where a full kernel cannot be extracted) so as to determine all three color planes for each pixel. For a second pixel, the method further includes a third process. The third process includes extracting a third kernel (e.g., a 3 by 3 sub matrix centered on the second pixel) from the multi-color matrix (e.g., the Bayer multi-color matrix described in this example). The third process further includes generating third variance offsets from the third kernel (e.g., Δg2, Δg4, Δg6, and Δg8) in a process substantially the same as is described above with respect to the second process, and generating a third color (Blue in this example) based on the third variance offsets and an adjacent pixel of the third color in a process substantially the same as is described above with respect to the second process.
  • The value of the predetermined exponent, k, is determined in step 72 (FIG. 6). The system noise level 70 is a matrix of elements corresponding to the elements of the captured Bayer image data, but representing system noise level. For example, a totally dark image could be captured in the sensor and loaded into matrix 70, or an imaging device could contain a group of light shielded pixels used to determine k. An exemplary way in which k may be determined is to average the noise value over all elements to determine an overall average, and then compute a variance to this average over all elements of the matrix. The variance becomes input to a look up table where the value of k is read as an output. In this way, different loads of the look up table can be used to emphasize edge clarity in low-noise situations and avoid enhancing noise while maintaining minimal color aliasing in high-noise situations.
  • After a full RGB image is recovered in block 118 (FIG. 6), the image in RGB space is converted to chroma space (C′ and C″ as discussed above) in block 120. In block 122, the chroma components (e.g., C′ and C″) are filtered to remove false colors and high frequency spikes in a low pass chroma filter (e.g., such as a moving window of a 3 by 3 kernel of elements where the chroma value of the center point takes up the average value over the 3 by 3 kernel). Then, in block 124, the chroma space image is converted back into RGB space.
  • The processes described above for determining Red and Blue values require that Green values be determined for all pixels in the 3 by 3 kernel used to determine the Red values or the kernel used to determine the Blue values. As can be observed in FIG. 6, a one row and a one column delay is introduced in block 66. This permits Green values to be determined one row and one column ahead of the determination of either Red or Blue values. Thus, when a 3 by 3 kernel is extracted to determine the Red values or to determine the Blue values, the entire kernel already has the Green value determined for each element of the 3 by 3 kernel. In this way, the process lends itself for implementation in a processor (be it general purpose or DSP) or an Application Specific Integrated Circuit (ASIC), but it could always be implemented in discrete circuits connected together. Regardless of the specific technology involved, the above described method is implemented in a computing machine of some form that includes a processor and a memory. The memory stores a first module for controlling the processor to perform a first process for a first pixel. The first process includes extracting a first kernel from a multi-color matrix, generating first variance weights from the first kernel, and generating a first color based on the first variance weights and adjacent pixel values of the first color. The memory further stores a second module for controlling the processor to perform a second process for the first pixel. The second process includes extracting a second kernel from the multi-color matrix, generating second variance offsets from the second kernel, and generating a second color based on the second variance offsets and an adjacent pixel of the second color. The processor may be a programmable processor that is controlled by programming modules stored on a separate computer readable media or it may be a fixed designed that is controlled by the timing generator of the processor.
  • In the event that the processes are implemented on a programmable computing machine (such as a general purpose processor or DSP), the details of above described method are included in program modules stored on a computer readable media. The computer readable media has stored thereon a plurality of modules for controlling a processor. The plurality of modules include first and second modules. The first module controls the processor to perform a first process for a first pixel. The first process includes extracting a first kernel from a multi-color matrix, generating first variance weights from the first kernel, and generating a first color based on the first variance weights and adjacent pixel values of the first color. The second module controls the processor to perform a second process for the first pixel. The second process includes extracting a second kernel from the multi-color matrix, generating second variance offsets from the second kernel, and generating a second color based on the second variance offsets and an adjacent pixel of the second color.
  • Having described preferred embodiments of a novel method and apparatus for reconstructing pixel color values (which are intended to be illustrative and not limiting), it is noted that modifications and variations can be made by persons skilled in the art in light of the above teachings. It is therefore to be understood that changes may be made in the particular embodiments of the invention disclosed which are within the scope and spirit of the invention as defined by the appended claims.
  • Having thus described the invention with the details and particularity required by the patent laws, what is claimed and desired protected by Letters Patent is set forth in the appended claims.

Claims (12)

1. A method for object-based color reconstruction in a multicolor matrix-based sensor arrangement comprising color sensors that have one first luminance component sensed at a relatively higher spatial frequency and two further chrominance components sensed at relatively lower spatial fancies, for a particular pixel not sensed in said first luminance component estimating its first color value through determining local gradients among various first luminance component values a said method being characterized by executing the following steps:
and in accordance with such local gradients executing such estimating through along relatively stronger edge informations, interpolating with a relatively greater weight factor, in favor over interpolating along relatively weaker edge informations with a relatively lesser weight factor,
and for a particular pixel not sensed in a particular further chrominance component value estimating that further chrominance component's value in a direction along with relatively smaller differences evaluated in said first luminance component.
2. A method as claimed in claim 1, wherein further chrominance component's values of said particular pixel are interpolated using neighboring pixels' information as based on whether they are situated within a same imaged object.
3. A method as claimed in claim 1 and applied to a Bayer matrix wherein said first luminance component is green.
4. A method as claimed in Clam 2, when said first luminance component's value is estimated on the basis of a 5×5 pixel kernel centered on said particular point.
5. A met as claimed in claim 1, whilst through exponential gradient values adjusting an exponent value (k) for emphasizing edge clarity in a low-noise situation, or rather limiting noise propagation whilst still mitigating for color aliasing.
6. A method as claimed in claim 1, whilst adding a low-pass filter step after estimating non-sampled colors for mitigating false-color spikes.
7. A method as claimed in claim 6, whilst supplementing said low pass filter step with relatively enhancing spatial high frequencies relatively far from said low-pass filter's discriminatory frequency for edge sharpness enhancement FIG. 5).
8. A computer program comprising program instruction for controlling a computer to implement a method according to one of claim 1.
9. A computer program product as being represented with a tangible read-only computer memory medium or being carried by an electrical signal, and comprising program instructions for controlling a cower to implement a method according to one of claim 1.
10. An apparatus being arranged for implementing a method as claimed in claim 1.
11. An apparatus according to claim 10, and executed as a filter facility for limiting noise propagation.
12. An image facility comprising an image forming facility for forming an image on an as claimed in claim 10, and furthermore comprising a memory facility fed by said apparatus, processing means for dynamically interacting with pixel values in said memory and user output means for outputting a reconstructed user image.
US12/386,779 2002-10-10 2009-04-23 Method for reconstruction of pixel color values Abandoned US20090263017A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US12/386,779 US20090263017A1 (en) 2002-10-10 2009-04-23 Method for reconstruction of pixel color values

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
US41715202P 2002-10-10 2002-10-10
US41714202P 2002-10-10 2002-10-10
US65841303A 2003-09-10 2003-09-10
US12/386,779 US20090263017A1 (en) 2002-10-10 2009-04-23 Method for reconstruction of pixel color values

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US65841303A Division 2002-10-10 2003-09-10

Publications (1)

Publication Number Publication Date
US20090263017A1 true US20090263017A1 (en) 2009-10-22

Family

ID=40223929

Family Applications (2)

Application Number Title Priority Date Filing Date
US10/658,523 Active 2027-07-20 US7477781B1 (en) 2002-10-10 2003-09-10 Method and apparatus for adaptive pixel correction of multi-color matrix
US12/386,779 Abandoned US20090263017A1 (en) 2002-10-10 2009-04-23 Method for reconstruction of pixel color values

Family Applications Before (1)

Application Number Title Priority Date Filing Date
US10/658,523 Active 2027-07-20 US7477781B1 (en) 2002-10-10 2003-09-10 Method and apparatus for adaptive pixel correction of multi-color matrix

Country Status (1)

Country Link
US (2) US7477781B1 (en)

Cited By (67)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20090116750A1 (en) * 2006-05-30 2009-05-07 Ho-Young Lee Color interpolation method and device
US20090310000A1 (en) * 2008-06-17 2009-12-17 Junichi Hosokawa Solid-state imaging device
US8693775B2 (en) * 2011-03-29 2014-04-08 Sony Corporation Image processing apparatus, method, recording medium, and program
US20140253808A1 (en) * 2011-08-31 2014-09-11 Sony Corporation Image processing device, and image processing method, and program
US20140347509A1 (en) * 2008-05-20 2014-11-27 Pelican Imaging Corporation Capturing and Processing of Images Including Occlusions Captured by Arrays of Luma and Chroma Cameras
US8929651B2 (en) 2010-11-16 2015-01-06 Thomson Licensing System and method for the repair of anomalies in images
US8995766B1 (en) * 2013-09-27 2015-03-31 Himax Imaging Limited Image processing method and image processing device
US9025894B2 (en) 2011-09-28 2015-05-05 Pelican Imaging Corporation Systems and methods for decoding light field image files having depth and confidence maps
US9041829B2 (en) 2008-05-20 2015-05-26 Pelican Imaging Corporation Capturing and processing of high dynamic range images using camera arrays
US9041824B2 (en) 2010-12-14 2015-05-26 Pelican Imaging Corporation Systems and methods for dynamic refocusing of high resolution images generated using images captured by a plurality of imagers
US9100635B2 (en) 2012-06-28 2015-08-04 Pelican Imaging Corporation Systems and methods for detecting defective camera arrays and optic arrays
US9100586B2 (en) 2013-03-14 2015-08-04 Pelican Imaging Corporation Systems and methods for photometric normalization in array cameras
US9106784B2 (en) 2013-03-13 2015-08-11 Pelican Imaging Corporation Systems and methods for controlling aliasing in images captured by an array camera for use in super-resolution processing
US9113118B2 (en) 2012-05-31 2015-08-18 Apple Inc. Green non-uniformity correction
US9123117B2 (en) 2012-08-21 2015-09-01 Pelican Imaging Corporation Systems and methods for generating depth maps and corresponding confidence maps indicating depth estimation reliability
US9128228B2 (en) 2011-06-28 2015-09-08 Pelican Imaging Corporation Optical arrangements for use with an array camera
US9143711B2 (en) 2012-11-13 2015-09-22 Pelican Imaging Corporation Systems and methods for array camera focal plane control
US9185276B2 (en) 2013-11-07 2015-11-10 Pelican Imaging Corporation Methods of manufacturing array camera modules incorporating independently aligned lens stacks
US9210392B2 (en) 2012-05-01 2015-12-08 Pelican Imaging Coporation Camera modules patterned with pi filter groups
US9214013B2 (en) 2012-09-14 2015-12-15 Pelican Imaging Corporation Systems and methods for correcting user identified artifacts in light field images
US9247117B2 (en) 2014-04-07 2016-01-26 Pelican Imaging Corporation Systems and methods for correcting for warpage of a sensor array in an array camera module by introducing warpage into a focal plane of a lens stack array
US9253380B2 (en) 2013-02-24 2016-02-02 Pelican Imaging Corporation Thin form factor computational array cameras and modular array cameras
US9264610B2 (en) 2009-11-20 2016-02-16 Pelican Imaging Corporation Capturing and processing of images including occlusions captured by heterogeneous camera arrays
US9412206B2 (en) 2012-02-21 2016-08-09 Pelican Imaging Corporation Systems and methods for the manipulation of captured light field image data
US9426361B2 (en) 2013-11-26 2016-08-23 Pelican Imaging Corporation Array camera configurations incorporating multiple constituent array cameras
US9438888B2 (en) 2013-03-15 2016-09-06 Pelican Imaging Corporation Systems and methods for stereo imaging with camera arrays
US9462164B2 (en) 2013-02-21 2016-10-04 Pelican Imaging Corporation Systems and methods for generating compressed light field representation data using captured light fields, array geometry, and parallax information
US9497370B2 (en) 2013-03-15 2016-11-15 Pelican Imaging Corporation Array camera architecture implementing quantum dot color filters
US9497429B2 (en) 2013-03-15 2016-11-15 Pelican Imaging Corporation Extended color processing on pelican array cameras
US9516222B2 (en) 2011-06-28 2016-12-06 Kip Peli P1 Lp Array cameras incorporating monolithic array camera modules with high MTF lens stacks for capture of images used in super-resolution processing
US9521416B1 (en) 2013-03-11 2016-12-13 Kip Peli P1 Lp Systems and methods for image data compression
US9519972B2 (en) 2013-03-13 2016-12-13 Kip Peli P1 Lp Systems and methods for synthesizing images from image data captured by an array camera using restricted depth of field depth maps in which depth estimation precision varies
US9521319B2 (en) 2014-06-18 2016-12-13 Pelican Imaging Corporation Array cameras and array camera modules including spectral filters disposed outside of a constituent image sensor
US9578259B2 (en) 2013-03-14 2017-02-21 Fotonation Cayman Limited Systems and methods for reducing motion blur in images or video in ultra low light with array cameras
US9633442B2 (en) 2013-03-15 2017-04-25 Fotonation Cayman Limited Array cameras including an array camera module augmented with a separate camera
US9741118B2 (en) 2013-03-13 2017-08-22 Fotonation Cayman Limited System and methods for calibration of an array camera
US9766380B2 (en) 2012-06-30 2017-09-19 Fotonation Cayman Limited Systems and methods for manufacturing camera modules using active alignment of lens stack arrays and sensors
US9774789B2 (en) 2013-03-08 2017-09-26 Fotonation Cayman Limited Systems and methods for high dynamic range imaging using array cameras
US9794476B2 (en) 2011-09-19 2017-10-17 Fotonation Cayman Limited Systems and methods for controlling aliasing in images captured by an array camera for use in super resolution processing using pixel apertures
US9813616B2 (en) 2012-08-23 2017-11-07 Fotonation Cayman Limited Feature based high resolution motion estimation from low resolution images captured using an array source
US9866739B2 (en) 2011-05-11 2018-01-09 Fotonation Cayman Limited Systems and methods for transmitting and receiving array camera image data
US9888194B2 (en) 2013-03-13 2018-02-06 Fotonation Cayman Limited Array camera architecture implementing quantum film image sensors
US9898856B2 (en) 2013-09-27 2018-02-20 Fotonation Cayman Limited Systems and methods for depth-assisted perspective distortion correction
US9936148B2 (en) 2010-05-12 2018-04-03 Fotonation Cayman Limited Imager array interfaces
US9942474B2 (en) 2015-04-17 2018-04-10 Fotonation Cayman Limited Systems and methods for performing high speed video capture and depth estimation using array cameras
US9955070B2 (en) 2013-03-15 2018-04-24 Fotonation Cayman Limited Systems and methods for synthesizing high resolution images using image deconvolution based on motion and depth information
US9986224B2 (en) 2013-03-10 2018-05-29 Fotonation Cayman Limited System and methods for calibration of an array camera
US10089740B2 (en) 2014-03-07 2018-10-02 Fotonation Limited System and methods for depth regularization and semiautomatic interactive matting using RGB-D images
TWI638339B (en) * 2017-11-14 2018-10-11 瑞昱半導體股份有限公司 False color removal method
US10119808B2 (en) 2013-11-18 2018-11-06 Fotonation Limited Systems and methods for estimating depth from projected texture using camera arrays
US10122993B2 (en) 2013-03-15 2018-11-06 Fotonation Limited Autofocus system for a conventional camera that uses depth information from an array camera
US10250871B2 (en) 2014-09-29 2019-04-02 Fotonation Limited Systems and methods for dynamic calibration of array cameras
US10390005B2 (en) 2012-09-28 2019-08-20 Fotonation Limited Generating images from light fields utilizing virtual viewpoints
US10482618B2 (en) 2017-08-21 2019-11-19 Fotonation Limited Systems and methods for hybrid depth regularization
US11270110B2 (en) 2019-09-17 2022-03-08 Boston Polarimetrics, Inc. Systems and methods for surface modeling using polarization cues
US11290658B1 (en) 2021-04-15 2022-03-29 Boston Polarimetrics, Inc. Systems and methods for camera exposure control
US11302012B2 (en) 2019-11-30 2022-04-12 Boston Polarimetrics, Inc. Systems and methods for transparent object segmentation using polarization cues
US11525906B2 (en) 2019-10-07 2022-12-13 Intrinsic Innovation Llc Systems and methods for augmentation of sensor systems and imaging systems with polarization
US11580667B2 (en) 2020-01-29 2023-02-14 Intrinsic Innovation Llc Systems and methods for characterizing object pose detection and measurement systems
US11689813B2 (en) 2021-07-01 2023-06-27 Intrinsic Innovation Llc Systems and methods for high dynamic range imaging using crossed polarizers
US11792538B2 (en) 2008-05-20 2023-10-17 Adeia Imaging Llc Capturing and processing of images including occlusions focused on an image sensor by a lens stack array
US11797863B2 (en) 2020-01-30 2023-10-24 Intrinsic Innovation Llc Systems and methods for synthesizing data for training statistical models on different imaging modalities including polarized images
US11954886B2 (en) 2021-04-15 2024-04-09 Intrinsic Innovation Llc Systems and methods for six-degree of freedom pose estimation of deformable objects
US11953700B2 (en) 2020-05-27 2024-04-09 Intrinsic Innovation Llc Multi-aperture polarization optical systems using beam splitters
US12020455B2 (en) 2021-03-10 2024-06-25 Intrinsic Innovation Llc Systems and methods for high dynamic range image reconstruction
US12067746B2 (en) 2021-05-07 2024-08-20 Intrinsic Innovation Llc Systems and methods for using computer vision to pick up small objects
US12069227B2 (en) 2021-03-10 2024-08-20 Intrinsic Innovation Llc Multi-modal and multi-spectral stereo camera arrays

Families Citing this family (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7620241B2 (en) * 2004-11-30 2009-11-17 Hewlett-Packard Development Company, L.P. Artifact reduction in a digital video
KR101503227B1 (en) * 2007-04-11 2015-03-16 레드.컴 인코포레이티드 Video camera
US8237830B2 (en) 2007-04-11 2012-08-07 Red.Com, Inc. Video camera
US20080273102A1 (en) * 2007-05-01 2008-11-06 Hoya Corporation Detection device for defective pixel in photographic device
KR101340518B1 (en) * 2007-08-23 2013-12-11 삼성전기주식회사 Method and apparatus for compensating chromatic aberration of an image
JP6046927B2 (en) 2011-08-09 2016-12-21 キヤノン株式会社 Image processing apparatus and control method thereof
US8891866B2 (en) 2011-08-31 2014-11-18 Sony Corporation Image processing apparatus, image processing method, and program
WO2014127153A1 (en) 2013-02-14 2014-08-21 Red. Com, Inc. Video camera
US9088740B2 (en) * 2013-06-21 2015-07-21 Himax Imaging Limited System and method of reducing noise
GB2521408B (en) * 2013-12-18 2015-12-16 Imagination Tech Ltd Defective pixel fixing
WO2019010233A1 (en) 2017-07-05 2019-01-10 Red. Com, Llc Video image data processing in electronic devices

Citations (44)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4868655A (en) * 1987-11-09 1989-09-19 Etat Francais Represente Par Le Ministre Delegue Des Postes Et Telecommunications (Centre National D'etudes Des Telecommunications) Method and apparatus for processing picture signals having interlaced field scanning
US5291293A (en) * 1992-06-01 1994-03-01 Eastman Kodak Company Electronic imaging device with defect correction
US5596367A (en) * 1996-02-23 1997-01-21 Eastman Kodak Company Averaging green values for green photosites in electronic cameras
US5629734A (en) * 1995-03-17 1997-05-13 Eastman Kodak Company Adaptive color plan interpolation in single sensor color electronic camera
US5631703A (en) * 1996-05-29 1997-05-20 Eastman Kodak Company Particular pattern of pixels for a color filter array which is used to derive luminance and chrominance values
US5652621A (en) * 1996-02-23 1997-07-29 Eastman Kodak Company Adaptive color plane interpolation in single sensor color electronic camera
US5699167A (en) * 1993-09-30 1997-12-16 Canon Kabushiki Kaisha Image forming apparatus which outputs a color image by separating color image information into at least two color components
US5748803A (en) * 1995-11-22 1998-05-05 Linotype-Hell Ag Method and apparatus for image value correction in optoelectronic transducers
US5805217A (en) * 1996-06-14 1998-09-08 Iterated Systems, Inc. Method and system for interpolating missing picture elements in a single color component array obtained from a single color sensor
US5881182A (en) * 1997-05-12 1999-03-09 Eastman Kodak Company Adaptive process for removing streaks in digital images
US5990950A (en) * 1998-02-11 1999-11-23 Iterated Systems, Inc. Method and system for color filter array multifactor interpolation
US6078686A (en) * 1996-09-30 2000-06-20 Samsung Electronics Co., Ltd. Image quality enhancement circuit and method therefor
US6091851A (en) * 1997-11-03 2000-07-18 Intel Corporation Efficient algorithm for color recovery from 8-bit to 24-bit color pixels
US6229578B1 (en) * 1997-12-08 2001-05-08 Intel Corporation Edge-detection based noise removal algorithm
US6236433B1 (en) * 1998-09-29 2001-05-22 Intel Corporation Scaling algorithm for efficient color representation/recovery in video
US6268939B1 (en) * 1998-01-08 2001-07-31 Xerox Corporation Method and apparatus for correcting luminance and chrominance data in digital color images
US6330029B1 (en) * 1998-03-17 2001-12-11 Eastman Kodak Company Particular pattern of pixels for a color filter array which is used to derive luminance and chrominance values
US20020012476A1 (en) * 2000-06-23 2002-01-31 Dillen Bartholomeus Goverdina Maria Henricus Image sensor signal defect correction
US20020015111A1 (en) * 2000-06-30 2002-02-07 Yoshihito Harada Image processing apparatus and its processing method
US20020031257A1 (en) * 2000-09-12 2002-03-14 Shinichi Kato Image processing apparatus and method
US20020054219A1 (en) * 2000-06-29 2002-05-09 Jaspers Cornelis Antonie Maria Processing a sensor output signal
US6392699B1 (en) * 1998-03-04 2002-05-21 Intel Corporation Integrated color interpolation and color space conversion algorithm from 8-bit bayer pattern RGB color space to 12-bit YCrCb color space
US20020063789A1 (en) * 2000-11-30 2002-05-30 Tinku Acharya Color filter array and color interpolation algorithm
US6421084B1 (en) * 1998-03-02 2002-07-16 Compaq Computer Corporation Method for interpolating a full color image from a single sensor using multiple threshold-based gradients
US20020136463A1 (en) * 2001-01-24 2002-09-26 Taisuke Akahori Image processing apparatus, image forming apparatus, and image processing method
US20020141654A1 (en) * 2001-03-29 2002-10-03 Dean Rosales Providing multiple symmetrical filters
US6487304B1 (en) * 1999-06-16 2002-11-26 Microsoft Corporation Multi-view approach to motion and stereo
US20030002747A1 (en) * 2001-07-02 2003-01-02 Jasc Software,Inc Moire correction in images
US6535632B1 (en) * 1998-12-18 2003-03-18 University Of Washington Image processing in HSI color space using adaptive noise filtering
US6674903B1 (en) * 1998-10-05 2004-01-06 Agfa-Gevaert Method for smoothing staircase effect in enlarged low resolution images
US6721458B1 (en) * 2000-04-14 2004-04-13 Seiko Epson Corporation Artifact reduction using adaptive nonlinear filters
US6798910B1 (en) * 2001-05-17 2004-09-28 The United States Of America As Represented By The Secretary Of The Air Force Self-optimizing edge detection in blurred, high-noise images
US6823086B1 (en) * 2000-08-29 2004-11-23 Analogic Corporation Adaptive spatial filter
US20050047675A1 (en) * 1999-09-16 2005-03-03 Walmsley Simon Robert Method of sharpening image using luminance channel
US6956582B2 (en) * 2001-08-23 2005-10-18 Evans & Sutherland Computer Corporation System and method for auto-adjusting image filtering
US7027637B2 (en) * 2002-02-21 2006-04-11 Siemens Corporate Research, Inc. Adaptive threshold determination for ball grid array component modeling
US7145693B2 (en) * 2001-08-08 2006-12-05 Canon Kabushiki Kaisha Image processing apparatus and method
US20070053585A1 (en) * 2005-05-31 2007-03-08 Microsoft Corporation Accelerated face detection based on prior probability of a view
US7231084B2 (en) * 2002-09-27 2007-06-12 Motorola, Inc. Color data image acquistion and processing
US7256828B2 (en) * 2003-01-16 2007-08-14 Dialog Imaging Systems Gmbh Weighted gradient based and color corrected interpolation
US7272265B2 (en) * 1998-03-13 2007-09-18 The University Of Houston System Methods for performing DAF data filtering and padding
US20070292022A1 (en) * 2003-01-16 2007-12-20 Andreas Nilsson Weighted gradient based and color corrected interpolation
US7409084B2 (en) * 2004-05-06 2008-08-05 Magnachip Semiconductor, Ltd. Method for determining dark condition and method for interpolating color in image sensor
US7630546B2 (en) * 2003-06-12 2009-12-08 Nikon Corporation Image processing method, image processing program and image processor

Family Cites Families (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
BE1007354A3 (en) 1993-07-23 1995-05-23 Philips Electronics Nv Signal correction circuit.
US5778106A (en) 1996-03-14 1998-07-07 Polaroid Corporation Electronic camera with reduced color artifacts
US6697534B1 (en) * 1999-06-09 2004-02-24 Intel Corporation Method and apparatus for adaptively sharpening local image content of an image
US7035475B1 (en) * 1999-06-17 2006-04-25 Raytheon Company Non-traditional adaptive non-uniformity compensation (ADNUC) system employing adaptive feedforward shunting and operating methods therefor
US6832009B1 (en) * 1999-09-24 2004-12-14 Zoran Corporation Method and apparatus for improved image interpolation
US6724945B1 (en) 2000-05-24 2004-04-20 Hewlett-Packard Development Company, L.P. Correcting defect pixels in a digital image
JP3636046B2 (en) 2000-07-31 2005-04-06 株式会社日立国際電気 Pixel defect detection method for solid-state image sensor and imaging device using the method
US7003173B2 (en) * 2001-06-12 2006-02-21 Sharp Laboratories Of America, Inc. Filter for combined de-ringing and edge sharpening
US6868179B2 (en) * 2001-07-06 2005-03-15 Jasc Software, Inc. Automatic saturation adjustment
GB0125774D0 (en) * 2001-10-26 2001-12-19 Cableform Ltd Method and apparatus for image matching
WO2004114063A2 (en) * 2003-06-13 2004-12-29 Georgia Tech Research Corporation Data reconstruction using directional interpolation techniques
US7373013B2 (en) * 2003-12-23 2008-05-13 General Instrument Corporation Directional video filters for locally adaptive spatial noise reduction

Patent Citations (45)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4868655A (en) * 1987-11-09 1989-09-19 Etat Francais Represente Par Le Ministre Delegue Des Postes Et Telecommunications (Centre National D'etudes Des Telecommunications) Method and apparatus for processing picture signals having interlaced field scanning
US5291293A (en) * 1992-06-01 1994-03-01 Eastman Kodak Company Electronic imaging device with defect correction
US5699167A (en) * 1993-09-30 1997-12-16 Canon Kabushiki Kaisha Image forming apparatus which outputs a color image by separating color image information into at least two color components
US5629734A (en) * 1995-03-17 1997-05-13 Eastman Kodak Company Adaptive color plan interpolation in single sensor color electronic camera
US5748803A (en) * 1995-11-22 1998-05-05 Linotype-Hell Ag Method and apparatus for image value correction in optoelectronic transducers
US5652621A (en) * 1996-02-23 1997-07-29 Eastman Kodak Company Adaptive color plane interpolation in single sensor color electronic camera
US5596367A (en) * 1996-02-23 1997-01-21 Eastman Kodak Company Averaging green values for green photosites in electronic cameras
US5631703A (en) * 1996-05-29 1997-05-20 Eastman Kodak Company Particular pattern of pixels for a color filter array which is used to derive luminance and chrominance values
US5805217A (en) * 1996-06-14 1998-09-08 Iterated Systems, Inc. Method and system for interpolating missing picture elements in a single color component array obtained from a single color sensor
US6078686A (en) * 1996-09-30 2000-06-20 Samsung Electronics Co., Ltd. Image quality enhancement circuit and method therefor
US5881182A (en) * 1997-05-12 1999-03-09 Eastman Kodak Company Adaptive process for removing streaks in digital images
US6091851A (en) * 1997-11-03 2000-07-18 Intel Corporation Efficient algorithm for color recovery from 8-bit to 24-bit color pixels
US6229578B1 (en) * 1997-12-08 2001-05-08 Intel Corporation Edge-detection based noise removal algorithm
US6268939B1 (en) * 1998-01-08 2001-07-31 Xerox Corporation Method and apparatus for correcting luminance and chrominance data in digital color images
US5990950A (en) * 1998-02-11 1999-11-23 Iterated Systems, Inc. Method and system for color filter array multifactor interpolation
US6421084B1 (en) * 1998-03-02 2002-07-16 Compaq Computer Corporation Method for interpolating a full color image from a single sensor using multiple threshold-based gradients
US6392699B1 (en) * 1998-03-04 2002-05-21 Intel Corporation Integrated color interpolation and color space conversion algorithm from 8-bit bayer pattern RGB color space to 12-bit YCrCb color space
US7272265B2 (en) * 1998-03-13 2007-09-18 The University Of Houston System Methods for performing DAF data filtering and padding
US6330029B1 (en) * 1998-03-17 2001-12-11 Eastman Kodak Company Particular pattern of pixels for a color filter array which is used to derive luminance and chrominance values
US6236433B1 (en) * 1998-09-29 2001-05-22 Intel Corporation Scaling algorithm for efficient color representation/recovery in video
US6674903B1 (en) * 1998-10-05 2004-01-06 Agfa-Gevaert Method for smoothing staircase effect in enlarged low resolution images
US6535632B1 (en) * 1998-12-18 2003-03-18 University Of Washington Image processing in HSI color space using adaptive noise filtering
US6487304B1 (en) * 1999-06-16 2002-11-26 Microsoft Corporation Multi-view approach to motion and stereo
US20050047675A1 (en) * 1999-09-16 2005-03-03 Walmsley Simon Robert Method of sharpening image using luminance channel
US6721458B1 (en) * 2000-04-14 2004-04-13 Seiko Epson Corporation Artifact reduction using adaptive nonlinear filters
US20020012476A1 (en) * 2000-06-23 2002-01-31 Dillen Bartholomeus Goverdina Maria Henricus Image sensor signal defect correction
US20020054219A1 (en) * 2000-06-29 2002-05-09 Jaspers Cornelis Antonie Maria Processing a sensor output signal
US20020015111A1 (en) * 2000-06-30 2002-02-07 Yoshihito Harada Image processing apparatus and its processing method
US6823086B1 (en) * 2000-08-29 2004-11-23 Analogic Corporation Adaptive spatial filter
US20020031257A1 (en) * 2000-09-12 2002-03-14 Shinichi Kato Image processing apparatus and method
US20020063789A1 (en) * 2000-11-30 2002-05-30 Tinku Acharya Color filter array and color interpolation algorithm
US20020136463A1 (en) * 2001-01-24 2002-09-26 Taisuke Akahori Image processing apparatus, image forming apparatus, and image processing method
US20020141654A1 (en) * 2001-03-29 2002-10-03 Dean Rosales Providing multiple symmetrical filters
US6798910B1 (en) * 2001-05-17 2004-09-28 The United States Of America As Represented By The Secretary Of The Air Force Self-optimizing edge detection in blurred, high-noise images
US20050220360A1 (en) * 2001-07-02 2005-10-06 Zaklika Krzysztof A Moire correction in images
US20030002747A1 (en) * 2001-07-02 2003-01-02 Jasc Software,Inc Moire correction in images
US7145693B2 (en) * 2001-08-08 2006-12-05 Canon Kabushiki Kaisha Image processing apparatus and method
US6956582B2 (en) * 2001-08-23 2005-10-18 Evans & Sutherland Computer Corporation System and method for auto-adjusting image filtering
US7027637B2 (en) * 2002-02-21 2006-04-11 Siemens Corporate Research, Inc. Adaptive threshold determination for ball grid array component modeling
US7231084B2 (en) * 2002-09-27 2007-06-12 Motorola, Inc. Color data image acquistion and processing
US7256828B2 (en) * 2003-01-16 2007-08-14 Dialog Imaging Systems Gmbh Weighted gradient based and color corrected interpolation
US20070292022A1 (en) * 2003-01-16 2007-12-20 Andreas Nilsson Weighted gradient based and color corrected interpolation
US7630546B2 (en) * 2003-06-12 2009-12-08 Nikon Corporation Image processing method, image processing program and image processor
US7409084B2 (en) * 2004-05-06 2008-08-05 Magnachip Semiconductor, Ltd. Method for determining dark condition and method for interpolating color in image sensor
US20070053585A1 (en) * 2005-05-31 2007-03-08 Microsoft Corporation Accelerated face detection based on prior probability of a view

Cited By (180)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20090116750A1 (en) * 2006-05-30 2009-05-07 Ho-Young Lee Color interpolation method and device
US9049411B2 (en) 2008-05-20 2015-06-02 Pelican Imaging Corporation Camera arrays incorporating 3×3 imager configurations
US11412158B2 (en) 2008-05-20 2022-08-09 Fotonation Limited Capturing and processing of images including occlusions focused on an image sensor by a lens stack array
US9712759B2 (en) 2008-05-20 2017-07-18 Fotonation Cayman Limited Systems and methods for generating depth maps using a camera arrays incorporating monochrome and color cameras
US9049391B2 (en) 2008-05-20 2015-06-02 Pelican Imaging Corporation Capturing and processing of near-IR images including occlusions using camera arrays incorporating near-IR light sources
US20140347509A1 (en) * 2008-05-20 2014-11-27 Pelican Imaging Corporation Capturing and Processing of Images Including Occlusions Captured by Arrays of Luma and Chroma Cameras
US9124815B2 (en) * 2008-05-20 2015-09-01 Pelican Imaging Corporation Capturing and processing of images including occlusions captured by arrays of luma and chroma cameras
US9576369B2 (en) 2008-05-20 2017-02-21 Fotonation Cayman Limited Systems and methods for generating depth maps using images captured by camera arrays incorporating cameras having different fields of view
US11792538B2 (en) 2008-05-20 2023-10-17 Adeia Imaging Llc Capturing and processing of images including occlusions focused on an image sensor by a lens stack array
US9485496B2 (en) 2008-05-20 2016-11-01 Pelican Imaging Corporation Systems and methods for measuring depth using images captured by a camera array including cameras surrounding a central camera
US10027901B2 (en) 2008-05-20 2018-07-17 Fotonation Cayman Limited Systems and methods for generating depth maps using a camera arrays incorporating monochrome and color cameras
US10142560B2 (en) 2008-05-20 2018-11-27 Fotonation Limited Capturing and processing of images including occlusions focused on an image sensor by a lens stack array
US9188765B2 (en) 2008-05-20 2015-11-17 Pelican Imaging Corporation Capturing and processing of images including occlusions focused on an image sensor by a lens stack array
US9041829B2 (en) 2008-05-20 2015-05-26 Pelican Imaging Corporation Capturing and processing of high dynamic range images using camera arrays
US9041823B2 (en) 2008-05-20 2015-05-26 Pelican Imaging Corporation Systems and methods for performing post capture refocus using images captured by camera arrays
US9191580B2 (en) 2008-05-20 2015-11-17 Pelican Imaging Corporation Capturing and processing of images including occlusions captured by camera arrays
US9235898B2 (en) 2008-05-20 2016-01-12 Pelican Imaging Corporation Systems and methods for generating depth maps using light focused on an image sensor by a lens element array
US9049390B2 (en) 2008-05-20 2015-06-02 Pelican Imaging Corporation Capturing and processing of images captured by arrays including polychromatic cameras
US9749547B2 (en) 2008-05-20 2017-08-29 Fotonation Cayman Limited Capturing and processing of images using camera array incorperating Bayer cameras having different fields of view
US12022207B2 (en) 2008-05-20 2024-06-25 Adeia Imaging Llc Capturing and processing of images including occlusions focused on an image sensor by a lens stack array
US12041360B2 (en) 2008-05-20 2024-07-16 Adeia Imaging Llc Capturing and processing of images including occlusions focused on an image sensor by a lens stack array
US9049367B2 (en) 2008-05-20 2015-06-02 Pelican Imaging Corporation Systems and methods for synthesizing higher resolution images using images captured by camera arrays
US9049381B2 (en) 2008-05-20 2015-06-02 Pelican Imaging Corporation Systems and methods for normalizing image data captured by camera arrays
US9055233B2 (en) 2008-05-20 2015-06-09 Pelican Imaging Corporation Systems and methods for synthesizing higher resolution images using a set of images containing a baseline image
US9055213B2 (en) 2008-05-20 2015-06-09 Pelican Imaging Corporation Systems and methods for measuring depth using images captured by monolithic camera arrays including at least one bayer camera
US9060124B2 (en) 2008-05-20 2015-06-16 Pelican Imaging Corporation Capturing and processing of images using non-monolithic camera arrays
US9060121B2 (en) 2008-05-20 2015-06-16 Pelican Imaging Corporation Capturing and processing of images captured by camera arrays including cameras dedicated to sampling luma and cameras dedicated to sampling chroma
US9060142B2 (en) 2008-05-20 2015-06-16 Pelican Imaging Corporation Capturing and processing of images captured by camera arrays including heterogeneous optics
US9060120B2 (en) 2008-05-20 2015-06-16 Pelican Imaging Corporation Systems and methods for generating depth maps using images captured by camera arrays
US9077893B2 (en) 2008-05-20 2015-07-07 Pelican Imaging Corporation Capturing and processing of images captured by non-grid camera arrays
US9094661B2 (en) 2008-05-20 2015-07-28 Pelican Imaging Corporation Systems and methods for generating depth maps using a set of images containing a baseline image
US8077253B2 (en) * 2008-06-17 2011-12-13 Kabushiki Kaisha Toshiba Solid-state device having digital signal processing circuit
US20090310000A1 (en) * 2008-06-17 2009-12-17 Junichi Hosokawa Solid-state imaging device
US9264610B2 (en) 2009-11-20 2016-02-16 Pelican Imaging Corporation Capturing and processing of images including occlusions captured by heterogeneous camera arrays
US10306120B2 (en) 2009-11-20 2019-05-28 Fotonation Limited Capturing and processing of images captured by camera arrays incorporating cameras with telephoto and conventional lenses to generate depth maps
US9936148B2 (en) 2010-05-12 2018-04-03 Fotonation Cayman Limited Imager array interfaces
US10455168B2 (en) 2010-05-12 2019-10-22 Fotonation Limited Imager array interfaces
US8929651B2 (en) 2010-11-16 2015-01-06 Thomson Licensing System and method for the repair of anomalies in images
US11423513B2 (en) 2010-12-14 2022-08-23 Fotonation Limited Systems and methods for synthesizing high resolution images using images captured by an array of independently controllable imagers
US10366472B2 (en) 2010-12-14 2019-07-30 Fotonation Limited Systems and methods for synthesizing high resolution images using images captured by an array of independently controllable imagers
US9361662B2 (en) 2010-12-14 2016-06-07 Pelican Imaging Corporation Systems and methods for synthesizing high resolution images using images captured by an array of independently controllable imagers
US11875475B2 (en) 2010-12-14 2024-01-16 Adeia Imaging Llc Systems and methods for synthesizing high resolution images using images captured by an array of independently controllable imagers
US9047684B2 (en) 2010-12-14 2015-06-02 Pelican Imaging Corporation Systems and methods for synthesizing high resolution images using a set of geometrically registered images
US9041824B2 (en) 2010-12-14 2015-05-26 Pelican Imaging Corporation Systems and methods for dynamic refocusing of high resolution images generated using images captured by a plurality of imagers
US8693775B2 (en) * 2011-03-29 2014-04-08 Sony Corporation Image processing apparatus, method, recording medium, and program
US10218889B2 (en) 2011-05-11 2019-02-26 Fotonation Limited Systems and methods for transmitting and receiving array camera image data
US10742861B2 (en) 2011-05-11 2020-08-11 Fotonation Limited Systems and methods for transmitting and receiving array camera image data
US9866739B2 (en) 2011-05-11 2018-01-09 Fotonation Cayman Limited Systems and methods for transmitting and receiving array camera image data
US9128228B2 (en) 2011-06-28 2015-09-08 Pelican Imaging Corporation Optical arrangements for use with an array camera
US9516222B2 (en) 2011-06-28 2016-12-06 Kip Peli P1 Lp Array cameras incorporating monolithic array camera modules with high MTF lens stacks for capture of images used in super-resolution processing
US9578237B2 (en) 2011-06-28 2017-02-21 Fotonation Cayman Limited Array cameras incorporating optics with modulation transfer functions greater than sensor Nyquist frequency for capture of images used in super-resolution processing
US9179113B2 (en) * 2011-08-31 2015-11-03 Sony Corporation Image processing device, and image processing method, and program
US20140253808A1 (en) * 2011-08-31 2014-09-11 Sony Corporation Image processing device, and image processing method, and program
US10375302B2 (en) 2011-09-19 2019-08-06 Fotonation Limited Systems and methods for controlling aliasing in images captured by an array camera for use in super resolution processing using pixel apertures
US9794476B2 (en) 2011-09-19 2017-10-17 Fotonation Cayman Limited Systems and methods for controlling aliasing in images captured by an array camera for use in super resolution processing using pixel apertures
US10275676B2 (en) 2011-09-28 2019-04-30 Fotonation Limited Systems and methods for encoding image files containing depth maps stored as metadata
US9025894B2 (en) 2011-09-28 2015-05-05 Pelican Imaging Corporation Systems and methods for decoding light field image files having depth and confidence maps
US9129183B2 (en) 2011-09-28 2015-09-08 Pelican Imaging Corporation Systems and methods for encoding light field image files
US11729365B2 (en) 2011-09-28 2023-08-15 Adela Imaging LLC Systems and methods for encoding image files containing depth maps stored as metadata
US9042667B2 (en) 2011-09-28 2015-05-26 Pelican Imaging Corporation Systems and methods for decoding light field image files using a depth map
US9031343B2 (en) 2011-09-28 2015-05-12 Pelican Imaging Corporation Systems and methods for encoding light field image files having a depth map
US9031335B2 (en) 2011-09-28 2015-05-12 Pelican Imaging Corporation Systems and methods for encoding light field image files having depth and confidence maps
US9025895B2 (en) 2011-09-28 2015-05-05 Pelican Imaging Corporation Systems and methods for decoding refocusable light field image files
US20180197035A1 (en) 2011-09-28 2018-07-12 Fotonation Cayman Limited Systems and Methods for Encoding Image Files Containing Depth Maps Stored as Metadata
US10019816B2 (en) 2011-09-28 2018-07-10 Fotonation Cayman Limited Systems and methods for decoding image files containing depth maps stored as metadata
US10984276B2 (en) 2011-09-28 2021-04-20 Fotonation Limited Systems and methods for encoding image files containing depth maps stored as metadata
US9036931B2 (en) 2011-09-28 2015-05-19 Pelican Imaging Corporation Systems and methods for decoding structured light field image files
US9864921B2 (en) 2011-09-28 2018-01-09 Fotonation Cayman Limited Systems and methods for encoding image files containing depth maps stored as metadata
US10430682B2 (en) 2011-09-28 2019-10-01 Fotonation Limited Systems and methods for decoding image files containing depth maps stored as metadata
US9536166B2 (en) 2011-09-28 2017-01-03 Kip Peli P1 Lp Systems and methods for decoding image files containing depth maps stored as metadata
US12052409B2 (en) 2011-09-28 2024-07-30 Adela Imaging LLC Systems and methods for encoding image files containing depth maps stored as metadata
US9811753B2 (en) 2011-09-28 2017-11-07 Fotonation Cayman Limited Systems and methods for encoding light field image files
US10311649B2 (en) 2012-02-21 2019-06-04 Fotonation Limited Systems and method for performing depth based image editing
US9412206B2 (en) 2012-02-21 2016-08-09 Pelican Imaging Corporation Systems and methods for the manipulation of captured light field image data
US9754422B2 (en) 2012-02-21 2017-09-05 Fotonation Cayman Limited Systems and method for performing depth based image editing
US9210392B2 (en) 2012-05-01 2015-12-08 Pelican Imaging Coporation Camera modules patterned with pi filter groups
US9706132B2 (en) 2012-05-01 2017-07-11 Fotonation Cayman Limited Camera modules patterned with pi filter groups
US9113118B2 (en) 2012-05-31 2015-08-18 Apple Inc. Green non-uniformity correction
US9807382B2 (en) 2012-06-28 2017-10-31 Fotonation Cayman Limited Systems and methods for detecting defective camera arrays and optic arrays
US9100635B2 (en) 2012-06-28 2015-08-04 Pelican Imaging Corporation Systems and methods for detecting defective camera arrays and optic arrays
US10334241B2 (en) 2012-06-28 2019-06-25 Fotonation Limited Systems and methods for detecting defective camera arrays and optic arrays
US11022725B2 (en) 2012-06-30 2021-06-01 Fotonation Limited Systems and methods for manufacturing camera modules using active alignment of lens stack arrays and sensors
US10261219B2 (en) 2012-06-30 2019-04-16 Fotonation Limited Systems and methods for manufacturing camera modules using active alignment of lens stack arrays and sensors
US9766380B2 (en) 2012-06-30 2017-09-19 Fotonation Cayman Limited Systems and methods for manufacturing camera modules using active alignment of lens stack arrays and sensors
US9858673B2 (en) 2012-08-21 2018-01-02 Fotonation Cayman Limited Systems and methods for estimating depth and visibility from a reference viewpoint for pixels in a set of images captured from different viewpoints
US12002233B2 (en) 2012-08-21 2024-06-04 Adeia Imaging Llc Systems and methods for estimating depth and visibility from a reference viewpoint for pixels in a set of images captured from different viewpoints
US9123117B2 (en) 2012-08-21 2015-09-01 Pelican Imaging Corporation Systems and methods for generating depth maps and corresponding confidence maps indicating depth estimation reliability
US9123118B2 (en) 2012-08-21 2015-09-01 Pelican Imaging Corporation System and methods for measuring depth using an array camera employing a bayer filter
US9129377B2 (en) 2012-08-21 2015-09-08 Pelican Imaging Corporation Systems and methods for measuring depth based upon occlusion patterns in images
US9240049B2 (en) 2012-08-21 2016-01-19 Pelican Imaging Corporation Systems and methods for measuring depth using an array of independently controllable cameras
US10380752B2 (en) 2012-08-21 2019-08-13 Fotonation Limited Systems and methods for estimating depth and visibility from a reference viewpoint for pixels in a set of images captured from different viewpoints
US9147254B2 (en) 2012-08-21 2015-09-29 Pelican Imaging Corporation Systems and methods for measuring depth in the presence of occlusions using a subset of images
US9235900B2 (en) 2012-08-21 2016-01-12 Pelican Imaging Corporation Systems and methods for estimating depth and visibility from a reference viewpoint for pixels in a set of images captured from different viewpoints
US9813616B2 (en) 2012-08-23 2017-11-07 Fotonation Cayman Limited Feature based high resolution motion estimation from low resolution images captured using an array source
US10462362B2 (en) 2012-08-23 2019-10-29 Fotonation Limited Feature based high resolution motion estimation from low resolution images captured using an array source
US9214013B2 (en) 2012-09-14 2015-12-15 Pelican Imaging Corporation Systems and methods for correcting user identified artifacts in light field images
US10390005B2 (en) 2012-09-28 2019-08-20 Fotonation Limited Generating images from light fields utilizing virtual viewpoints
US9143711B2 (en) 2012-11-13 2015-09-22 Pelican Imaging Corporation Systems and methods for array camera focal plane control
US9749568B2 (en) 2012-11-13 2017-08-29 Fotonation Cayman Limited Systems and methods for array camera focal plane control
US9462164B2 (en) 2013-02-21 2016-10-04 Pelican Imaging Corporation Systems and methods for generating compressed light field representation data using captured light fields, array geometry, and parallax information
US10009538B2 (en) 2013-02-21 2018-06-26 Fotonation Cayman Limited Systems and methods for generating compressed light field representation data using captured light fields, array geometry, and parallax information
US9774831B2 (en) 2013-02-24 2017-09-26 Fotonation Cayman Limited Thin form factor computational array cameras and modular array cameras
US9253380B2 (en) 2013-02-24 2016-02-02 Pelican Imaging Corporation Thin form factor computational array cameras and modular array cameras
US9374512B2 (en) 2013-02-24 2016-06-21 Pelican Imaging Corporation Thin form factor computational array cameras and modular array cameras
US9743051B2 (en) 2013-02-24 2017-08-22 Fotonation Cayman Limited Thin form factor computational array cameras and modular array cameras
US9917998B2 (en) 2013-03-08 2018-03-13 Fotonation Cayman Limited Systems and methods for measuring scene information while capturing images using array cameras
US9774789B2 (en) 2013-03-08 2017-09-26 Fotonation Cayman Limited Systems and methods for high dynamic range imaging using array cameras
US9986224B2 (en) 2013-03-10 2018-05-29 Fotonation Cayman Limited System and methods for calibration of an array camera
US10225543B2 (en) 2013-03-10 2019-03-05 Fotonation Limited System and methods for calibration of an array camera
US11570423B2 (en) 2013-03-10 2023-01-31 Adeia Imaging Llc System and methods for calibration of an array camera
US11272161B2 (en) 2013-03-10 2022-03-08 Fotonation Limited System and methods for calibration of an array camera
US10958892B2 (en) 2013-03-10 2021-03-23 Fotonation Limited System and methods for calibration of an array camera
US11985293B2 (en) 2013-03-10 2024-05-14 Adeia Imaging Llc System and methods for calibration of an array camera
US9521416B1 (en) 2013-03-11 2016-12-13 Kip Peli P1 Lp Systems and methods for image data compression
US9519972B2 (en) 2013-03-13 2016-12-13 Kip Peli P1 Lp Systems and methods for synthesizing images from image data captured by an array camera using restricted depth of field depth maps in which depth estimation precision varies
US10127682B2 (en) 2013-03-13 2018-11-13 Fotonation Limited System and methods for calibration of an array camera
US9800856B2 (en) 2013-03-13 2017-10-24 Fotonation Cayman Limited Systems and methods for synthesizing images from image data captured by an array camera using restricted depth of field depth maps in which depth estimation precision varies
US9106784B2 (en) 2013-03-13 2015-08-11 Pelican Imaging Corporation Systems and methods for controlling aliasing in images captured by an array camera for use in super-resolution processing
US9888194B2 (en) 2013-03-13 2018-02-06 Fotonation Cayman Limited Array camera architecture implementing quantum film image sensors
US9733486B2 (en) 2013-03-13 2017-08-15 Fotonation Cayman Limited Systems and methods for controlling aliasing in images captured by an array camera for use in super-resolution processing
US9741118B2 (en) 2013-03-13 2017-08-22 Fotonation Cayman Limited System and methods for calibration of an array camera
US9578259B2 (en) 2013-03-14 2017-02-21 Fotonation Cayman Limited Systems and methods for reducing motion blur in images or video in ultra low light with array cameras
US9787911B2 (en) 2013-03-14 2017-10-10 Fotonation Cayman Limited Systems and methods for photometric normalization in array cameras
US10547772B2 (en) 2013-03-14 2020-01-28 Fotonation Limited Systems and methods for reducing motion blur in images or video in ultra low light with array cameras
US9100586B2 (en) 2013-03-14 2015-08-04 Pelican Imaging Corporation Systems and methods for photometric normalization in array cameras
US10412314B2 (en) 2013-03-14 2019-09-10 Fotonation Limited Systems and methods for photometric normalization in array cameras
US10091405B2 (en) 2013-03-14 2018-10-02 Fotonation Cayman Limited Systems and methods for reducing motion blur in images or video in ultra low light with array cameras
US10182216B2 (en) 2013-03-15 2019-01-15 Fotonation Limited Extended color processing on pelican array cameras
US9800859B2 (en) 2013-03-15 2017-10-24 Fotonation Cayman Limited Systems and methods for estimating depth using stereo array cameras
US9955070B2 (en) 2013-03-15 2018-04-24 Fotonation Cayman Limited Systems and methods for synthesizing high resolution images using image deconvolution based on motion and depth information
US9497429B2 (en) 2013-03-15 2016-11-15 Pelican Imaging Corporation Extended color processing on pelican array cameras
US10455218B2 (en) 2013-03-15 2019-10-22 Fotonation Limited Systems and methods for estimating depth using stereo array cameras
US9633442B2 (en) 2013-03-15 2017-04-25 Fotonation Cayman Limited Array cameras including an array camera module augmented with a separate camera
US9497370B2 (en) 2013-03-15 2016-11-15 Pelican Imaging Corporation Array camera architecture implementing quantum dot color filters
US10542208B2 (en) 2013-03-15 2020-01-21 Fotonation Limited Systems and methods for synthesizing high resolution images using image deconvolution based on motion and depth information
US10674138B2 (en) 2013-03-15 2020-06-02 Fotonation Limited Autofocus system for a conventional camera that uses depth information from an array camera
US9438888B2 (en) 2013-03-15 2016-09-06 Pelican Imaging Corporation Systems and methods for stereo imaging with camera arrays
US10122993B2 (en) 2013-03-15 2018-11-06 Fotonation Limited Autofocus system for a conventional camera that uses depth information from an array camera
US10638099B2 (en) 2013-03-15 2020-04-28 Fotonation Limited Extended color processing on pelican array cameras
US9898856B2 (en) 2013-09-27 2018-02-20 Fotonation Cayman Limited Systems and methods for depth-assisted perspective distortion correction
US8995766B1 (en) * 2013-09-27 2015-03-31 Himax Imaging Limited Image processing method and image processing device
US10540806B2 (en) 2013-09-27 2020-01-21 Fotonation Limited Systems and methods for depth-assisted perspective distortion correction
US9264592B2 (en) 2013-11-07 2016-02-16 Pelican Imaging Corporation Array camera modules incorporating independently aligned lens stacks
US9185276B2 (en) 2013-11-07 2015-11-10 Pelican Imaging Corporation Methods of manufacturing array camera modules incorporating independently aligned lens stacks
US9924092B2 (en) 2013-11-07 2018-03-20 Fotonation Cayman Limited Array cameras incorporating independently aligned lens stacks
US10767981B2 (en) 2013-11-18 2020-09-08 Fotonation Limited Systems and methods for estimating depth from projected texture using camera arrays
US10119808B2 (en) 2013-11-18 2018-11-06 Fotonation Limited Systems and methods for estimating depth from projected texture using camera arrays
US11486698B2 (en) 2013-11-18 2022-11-01 Fotonation Limited Systems and methods for estimating depth from projected texture using camera arrays
US10708492B2 (en) 2013-11-26 2020-07-07 Fotonation Limited Array camera configurations incorporating constituent array cameras and constituent cameras
US9426361B2 (en) 2013-11-26 2016-08-23 Pelican Imaging Corporation Array camera configurations incorporating multiple constituent array cameras
US9813617B2 (en) 2013-11-26 2017-11-07 Fotonation Cayman Limited Array camera configurations incorporating constituent array cameras and constituent cameras
US10574905B2 (en) 2014-03-07 2020-02-25 Fotonation Limited System and methods for depth regularization and semiautomatic interactive matting using RGB-D images
US10089740B2 (en) 2014-03-07 2018-10-02 Fotonation Limited System and methods for depth regularization and semiautomatic interactive matting using RGB-D images
US9247117B2 (en) 2014-04-07 2016-01-26 Pelican Imaging Corporation Systems and methods for correcting for warpage of a sensor array in an array camera module by introducing warpage into a focal plane of a lens stack array
US9521319B2 (en) 2014-06-18 2016-12-13 Pelican Imaging Corporation Array cameras and array camera modules including spectral filters disposed outside of a constituent image sensor
US11546576B2 (en) 2014-09-29 2023-01-03 Adeia Imaging Llc Systems and methods for dynamic calibration of array cameras
US10250871B2 (en) 2014-09-29 2019-04-02 Fotonation Limited Systems and methods for dynamic calibration of array cameras
US9942474B2 (en) 2015-04-17 2018-04-10 Fotonation Cayman Limited Systems and methods for performing high speed video capture and depth estimation using array cameras
US10482618B2 (en) 2017-08-21 2019-11-19 Fotonation Limited Systems and methods for hybrid depth regularization
US11562498B2 (en) 2017-08-21 2023-01-24 Adela Imaging LLC Systems and methods for hybrid depth regularization
US11983893B2 (en) 2017-08-21 2024-05-14 Adeia Imaging Llc Systems and methods for hybrid depth regularization
US10818026B2 (en) 2017-08-21 2020-10-27 Fotonation Limited Systems and methods for hybrid depth regularization
TWI638339B (en) * 2017-11-14 2018-10-11 瑞昱半導體股份有限公司 False color removal method
US11699273B2 (en) 2019-09-17 2023-07-11 Intrinsic Innovation Llc Systems and methods for surface modeling using polarization cues
US11270110B2 (en) 2019-09-17 2022-03-08 Boston Polarimetrics, Inc. Systems and methods for surface modeling using polarization cues
US11525906B2 (en) 2019-10-07 2022-12-13 Intrinsic Innovation Llc Systems and methods for augmentation of sensor systems and imaging systems with polarization
US11982775B2 (en) 2019-10-07 2024-05-14 Intrinsic Innovation Llc Systems and methods for augmentation of sensor systems and imaging systems with polarization
US12099148B2 (en) 2019-10-07 2024-09-24 Intrinsic Innovation Llc Systems and methods for surface normals sensing with polarization
US11302012B2 (en) 2019-11-30 2022-04-12 Boston Polarimetrics, Inc. Systems and methods for transparent object segmentation using polarization cues
US11842495B2 (en) 2019-11-30 2023-12-12 Intrinsic Innovation Llc Systems and methods for transparent object segmentation using polarization cues
US11580667B2 (en) 2020-01-29 2023-02-14 Intrinsic Innovation Llc Systems and methods for characterizing object pose detection and measurement systems
US11797863B2 (en) 2020-01-30 2023-10-24 Intrinsic Innovation Llc Systems and methods for synthesizing data for training statistical models on different imaging modalities including polarized images
US11953700B2 (en) 2020-05-27 2024-04-09 Intrinsic Innovation Llc Multi-aperture polarization optical systems using beam splitters
US12020455B2 (en) 2021-03-10 2024-06-25 Intrinsic Innovation Llc Systems and methods for high dynamic range image reconstruction
US12069227B2 (en) 2021-03-10 2024-08-20 Intrinsic Innovation Llc Multi-modal and multi-spectral stereo camera arrays
US11954886B2 (en) 2021-04-15 2024-04-09 Intrinsic Innovation Llc Systems and methods for six-degree of freedom pose estimation of deformable objects
US11683594B2 (en) 2021-04-15 2023-06-20 Intrinsic Innovation Llc Systems and methods for camera exposure control
US11290658B1 (en) 2021-04-15 2022-03-29 Boston Polarimetrics, Inc. Systems and methods for camera exposure control
US12067746B2 (en) 2021-05-07 2024-08-20 Intrinsic Innovation Llc Systems and methods for using computer vision to pick up small objects
US11689813B2 (en) 2021-07-01 2023-06-27 Intrinsic Innovation Llc Systems and methods for high dynamic range imaging using crossed polarizers

Also Published As

Publication number Publication date
US7477781B1 (en) 2009-01-13

Similar Documents

Publication Publication Date Title
US20090263017A1 (en) Method for reconstruction of pixel color values
US6806902B1 (en) System and method for correcting bad pixel data in a digital camera
US6625325B2 (en) Noise cleaning and interpolating sparsely populated color digital image using a variable noise cleaning kernel
EP0304643B1 (en) Improved method and apparatus for reconstructing missing color
US7486844B2 (en) Color interpolation apparatus and color interpolation method utilizing edge indicators adjusted by stochastic adjustment factors to reconstruct missing colors for image pixels
EP1641283B1 (en) Image processing method, image processing program, image processor
US7728882B2 (en) Green reconstruction for image sensors
US8270774B2 (en) Image processing device for performing interpolation
US20150049223A1 (en) Image processing apparatus, image processing method, and program
WO1999046730A1 (en) Edge-dependent interpolation method for color reconstruction in image processing devices
CN102907103A (en) Image processing device, image processing method and program
Chen et al. Effective demosaicking algorithm based on edge property for color filter arrays
CN113454687A (en) Image processing method, apparatus and system, computer readable storage medium
US8233733B2 (en) Image processing device
US20100134661A1 (en) Image processing apparatus, image processing method and program
JPH0646817B2 (en) Video signal processor
US8363135B2 (en) Method and device for reconstructing a color image
US7573515B2 (en) Method and apparatus for processing a sensor signal having a plurality of pixels from an image sensor, computer program product, computing system and camera
CN104427321B (en) Image processing apparatus and its control method
Wang et al. Demosaicing with improved edge direction detection
JP4178919B2 (en) Noise removal method, imaging apparatus, and noise removal program
JP3115123B2 (en) Image signal processing device
EP1522047B1 (en) Method and apparatus for signal processing, computer program product, computing system and camera
JP4666786B2 (en) Image interpolation device
JP4495355B2 (en) Image interpolation device

Legal Events

Date Code Title Description
STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION