US20090263017A1 - Method for reconstruction of pixel color values - Google Patents
Method for reconstruction of pixel color values Download PDFInfo
- Publication number
- US20090263017A1 US20090263017A1 US12/386,779 US38677909A US2009263017A1 US 20090263017 A1 US20090263017 A1 US 20090263017A1 US 38677909 A US38677909 A US 38677909A US 2009263017 A1 US2009263017 A1 US 2009263017A1
- Authority
- US
- United States
- Prior art keywords
- color
- pixel
- kernel
- value
- relatively
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
- 238000000034 method Methods 0.000 title claims abstract description 79
- 239000011159 matrix material Substances 0.000 claims abstract description 39
- 239000003086 colorant Substances 0.000 claims description 11
- 238000012545 processing Methods 0.000 claims description 6
- 230000002708 enhancing effect Effects 0.000 claims description 5
- 238000004590 computer program Methods 0.000 claims description 4
- 230000000116 mitigating effect Effects 0.000 claims 2
- 230000001502 supplementing effect Effects 0.000 claims 1
- 230000008569 process Effects 0.000 abstract description 47
- 238000004422 calculation algorithm Methods 0.000 description 7
- 230000004044 response Effects 0.000 description 6
- 230000003044 adaptive effect Effects 0.000 description 5
- 238000001914 filtration Methods 0.000 description 3
- 238000005070 sampling Methods 0.000 description 3
- 230000006399 behavior Effects 0.000 description 2
- 238000012937 correction Methods 0.000 description 2
- 238000013461 design Methods 0.000 description 2
- 238000010586 diagram Methods 0.000 description 2
- 230000014509 gene expression Effects 0.000 description 2
- 238000003384 imaging method Methods 0.000 description 2
- 230000003287 optical effect Effects 0.000 description 2
- 238000013459 approach Methods 0.000 description 1
- 238000004364 calculation method Methods 0.000 description 1
- 230000015556 catabolic process Effects 0.000 description 1
- 238000006243 chemical reaction Methods 0.000 description 1
- 238000005352 clarification Methods 0.000 description 1
- 239000002537 cosmetic Substances 0.000 description 1
- 230000007547 defect Effects 0.000 description 1
- 230000002950 deficient Effects 0.000 description 1
- 238000006731 degradation reaction Methods 0.000 description 1
- 230000001419 dependent effect Effects 0.000 description 1
- 238000006073 displacement reaction Methods 0.000 description 1
- 239000000284 extract Substances 0.000 description 1
- 230000006870 function Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000012544 monitoring process Methods 0.000 description 1
- 230000008520 organization Effects 0.000 description 1
- 230000035945 sensitivity Effects 0.000 description 1
- 230000001052 transient effect Effects 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N25/00—Circuitry of solid-state image sensors [SSIS]; Control thereof
- H04N25/60—Noise processing, e.g. detecting, correcting, reducing or removing noise
- H04N25/68—Noise processing, e.g. detecting, correcting, reducing or removing noise applied to defects
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/80—Camera processing pipelines; Components thereof
- H04N23/84—Camera processing pipelines; Components thereof for processing colour signals
- H04N23/843—Demosaicing, e.g. interpolating colour pixel values
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N25/00—Circuitry of solid-state image sensors [SSIS]; Control thereof
- H04N25/10—Circuitry of solid-state image sensors [SSIS]; Control thereof for transforming different wavelengths into image signals
- H04N25/11—Arrangement of colour filter arrays [CFA]; Filter mosaics
- H04N25/13—Arrangement of colour filter arrays [CFA]; Filter mosaics characterised by the spectral characteristics of the filter elements
- H04N25/134—Arrangement of colour filter arrays [CFA]; Filter mosaics characterised by the spectral characteristics of the filter elements based on three different wavelength filter elements
Definitions
- the present invention relates to a method and apparatus for object-based color reconstruction in a multi-color based sensor arrangement through maintaining coherence within objects and along edges, and a computer program and computer program product for controlling such method, and an image sensing facility comprising such apparatus.
- the invention relates to a method for executing object-based color reconstruction in a multi-color matrix-based sensor arrangement that estimates color value through determining local gradients among various luminance component values assessed.
- Multi-color matrix-based sensor arrangements that include color sensors that have one first luminance component sensed at a relatively higher spatial frequency and two further chrominance components sensed at relatively lower spatial frequencies are coming in rapidly expanding use.
- Fields of application include digital cameras, digital cinematography, video cameras, scientific imaging, broadcast cameras, surveillance, security monitoring, and various others.
- each object point translates into a single-color of the corresponding image pixel.
- each pixel lacks two of the three sensor colors
- reconstruction of three-color pixels requires an appreciable amount of processing.
- the resulting reconstructed array often had artificially colored, noise-enhanced, or blurred edges.
- the present embodiment generally utilizes only a small data kernel of optimally 5 ⁇ 5 pixels for Green, or minimally 3 ⁇ 3 pixels for Red and Blue. It will generally minimize perceptional error, it reconstructs frequencies above the Nyquist frequency, it has much lower color aliasing than many other algorithms, it will hardly expand existing defects, and it will only be marginally affected by image noise.
- United States Patent Application Publication 2002/0063789 A1 to Acharya et al and published on May 30, 2002 discloses a Color Filter Array and Color Interpolation Algorithm, and uses a classification to determine which pixels it will use in the interpolation.
- the present invention in contradistinction uses a soft decision method that is useful for adaptive adjusting for variations in system noise.
- the reference singles out a very particular matrix design with a very high percentage of 75% Green pixels sensed. The present invention is much wider applicable.
- the inventor has found that correct application of the algorithm will not remove fundamental signal noise. Rather, it will prevent enhancing of the noise.
- the inventor has recognized the corresponding behavior of the gradients among the various sensed colors.
- a method for object-based color reconstruction in a multi-color matrix-based sensor arrangement is based on color sensors that have one first color sensed at a relatively higher spatial frequency and two further colors sensed at relatively lower spatial frequencies.
- the method executes the following steps.
- a particular pixel not sensed in said first color has its first color value estimated through determining local gradients ( 82 ) among various first color values assessed, and in accordance with such gradients such estimating is executed through interpolating ( 84 , 86 ) along relatively stronger edge informations.
- For a particular pixel not sensed in a particular further color that further color's value is estimated in a direction along with relatively smaller differences evaluated in the first color ( 94 , 96 ).
- the method has the ability for limiting noise propagation.
- a method includes a first process for a first pixel and a second process for the first pixel.
- the first process includes extracting a first kernel from a multi-color matrix, generating first variance weights from the first kernel, and generating a first color based on the first variance weights and adjacent pixel values of the first color.
- the second process includes extracting a second kernel from the multi-color matrix, generating second variance offsets from the second kernel, and generating a second color based on the second variance offsets and an adjacent pixel of the second color.
- the invention also relates to a computing machine that includes a processor and a memory.
- the memory stores a first module for controlling the processor to perform a first process for a first pixel.
- the first process includes extracting a first kernel from a multi-color matrix, generating first variance weights from the first kernel, and generating a first color based on the first variance weights and adjacent pixel values of the first color.
- the memory further stores a second module for controlling the processor to perform a second process for the first pixel.
- the second process includes extracting a second kernel from the multi-color matrix, generating second variance offsets from the second kernel, and generating a second color based on the second variance offsets and an adjacent pixel of the second color.
- the invention also relates to a computer readable medium having stored thereon a plurality of modules for controlling a processor.
- the plurality of modules includes a first module and a second module.
- the first module is for controlling the processor to perform a first process for the first pixel.
- the first process includes extracting a first kernel from a multi-color matrix, generating first variance weights from the first kernel, and generating a first color based on the first variance weights and adjacent pixel values of the first color.
- the second module is for controlling the processor to perform a second process for the first pixel.
- the second process includes extracting a second kernel from the multi-color matrix, generating second variance offsets from the second kernel, and generating a second color based on the second variance offsets and an adjacent pixel of the second color.
- FIG. 1 is a schematic view of a Bayer matrix of image points
- FIG. 2 is a schematic view of the deriving of the horizontal Green gradient
- FIG. 3 is a schematic view of the deriving of the vertical Green gradient
- FIGS. 4 a , 4 b are schematic views of the relevance of the eight possible Green directions
- FIG. 5 is a graph of a filter characteristic diagram for removing false color responses.
- FIG. 6 is a schematic block diagram of a hardware implementation of the present invention.
- FIG. 1 illustrates a Bayer matrix of image points for use with the present invention.
- other arrangements could be used, such as through using subtractive colors, such as cyan, magenta, and yellow, or through various other published embodiments.
- the matrix has Green pixels arranged according to a chessboard organization. The remaining pixels have been used for Red pixels in odd rows and Blue pixels in even rows.
- the intended end result of the reconstruction embodiment is three complete images of Red, Green, and Blue pixel values, respectively, where the original recording did yield less than three values for each respective pixel, according to the pixel pattern for the respective single color. It is known from advance that at each pixel location, two values must be estimated. In practice, the application most favored is to a Bayer pattern wherein the green pixels form a chessboard pattern, while the non-green fields have alternating red and blue values in the diagonal directions.
- FIG. 2 illustrates the deriving of the horizontal Green gradient or horizontal green edge indicator ⁇ H in a preferred kernel pattern of 5 ⁇ 5 pixels according to the average absolute differences between pixels pairs joined by an arrow, according to:
- ⁇ H ⁇
- FIG. 3 illustrates the deriving of the vertical Green gradient or vertical Green edge indicator ⁇ V .
- the pixels assigned to other colors will be ignored.
- a local edge direction is determined and a Green value for the missing central point is estimated as follows: if ⁇ V is relatively larger, interpolate horizontally between two nearest green neighbors; if ⁇ H is relatively larger, interpolate vertically between two nearest green neighbors, and if ⁇ H and ⁇ V are roughly equal, interpolate linearly between all four nearest neighbors of the missing Green pixel.
- ⁇ V is relatively larger, interpolate horizontally between two nearest green neighbors
- ⁇ H and ⁇ V are roughly equal, interpolate linearly between all four nearest neighbors of the missing Green pixel.
- G i,j 1/2 ⁇ H ( G i,j ⁇ 1 +G i,j+1 )+1/2 ⁇ v ( G i ⁇ 1,j +G i+1,j ),
- ⁇ H 1 ⁇ k H /( ⁇ k H + ⁇ k V ), and likewise, for the vertical parameter.
- exponent constant k increases the algorithm's sensitivity to the differences in ⁇ H and ⁇ V .
- the algebraic expressions offer more benefits than just for presentation.
- the non-linear combination of the edge indicators can be adjusted with the value of k to either emphasize edge clarity in low-noise situations, or rather avoid the enhancing of the noise while still maintaining minimal color aliasing in high-noise situations. Therefore, k has a two-sided adaptive effect on the algorithm. With small values (between 1 and 2) noise will be adaptively minimized whilst ensuring that edges will have none to minimal color aliasing. On the other hand, with larger values for k in low-noise situations, more emphasis can be put on removing all color aliasing whilst enhancing edge continuity and sharpness. Further, the use of five arrows in FIGS.
- the missing points in the Red and Blue images will be estimated.
- the horizontal neighbors are Blue and the vertical neighbors are Red.
- the Red value is sought for.
- the vertically closest Green value to the actual pixel is determined and then its associated Red value copied to the actual pixel.
- the horizontally closest Green value to the actual pixel is determined and then its associated Blue value copied to the actual pixel.
- a corresponding version applies to the missing Blue values at pixels that were sampled Red.
- the color reconstruction according to the present invention's embodiment will generally be correct, the particular arrangement of the Bayer matrix, in combination with non-ideal optical image low-pass filtering and system noise can cause some false color responses. Heuristically, the following description is given.
- the sampling of a scene through a Bayer matrix will cause two major sampling errors, as follows: (a) The relative displacement between green, red and blue samples introduces a phase error between the various colors; (b) The sampling frequencies of Red and Blue are relatively lower than that of Green. This is perceptually advantageous, because the human eye's chroma and hue resolution is relatively less than its luminance resolution.
- C ′ ( R ⁇ G ) i,j
- C ′′ ( B ⁇ G ) i,j .
- FIG. 5 shows the intended frequency response of a chroma filter.
- the design of such a-filter that has been shown in solid lines is dependent on the system's optical low-pass filtering as well as on system noise. Ideally, it will be designed to minimize the loss of image information and at the same time maximize the removal of false colors. To an appreciable extent, degradation in sharpness through the alpha-filter is mitigated through a beta-filter shown in FIG. 5 in dashed lines.
- FIG. 6 illustrates a hardware implementation of the present invention.
- the output of the co-pending application referred to supra may be fed to a DSP pipeline 20 for processing and/or transient storage, as the case may be.
- the DSP pipeline is fed by the image sensing facility, not shown for brevity, whereas the pipeline itself has a sufficient amount of memory and processing facilities for implementing the method according to the present invention.
- the system will evaluate the noise level in the image, and therefrom in block 72 convert the noise level in the k-factor considered supra. From a holding element 74 , this k-factor is provided to such elements of the hardware as are in need of it.
- command, control and synchronization streams have been left out of the embodiment for brevity and optimum clarification.
- block 22 represents the captured Bayer image information.
- block 60 it is detected whether the pixel in question is at least 2 pixels away from the image edge. If not, the procedure hereinafter will not have enough data available, so the process will then move to the next pixel in block 62 .
- block 64 the system detects whether the sample pixel is Green.
- the system will undertake to reconstruct an appropriate Green value for the pixel in question, as follows.
- the system extracts in a kernel of 5 ⁇ 5 pixels disclosed earlier the sensed Green pixels.
- the horizontal and vertical variance parameters are derived.
- the horizontal and vertical weight generating functions are derived, through also receiving the actual value of the exponent k from block 74 , as discussed earlier.
- the system calculates a product of the sum weights and the horizontally and vertically sampled Green pixels. With the so reconstructed Green pixel, the system then proceeds to mixer/accumulator 112 whilst in block 114 storing the result and in block 116 detecting whether the end of the image had been reached, which for the time being is not so.
- the system shifts by one row and one column and checks for a Red pixel. If no, the pixel thus being Blue, the system will undertake to reconstruct a Red pixel.
- the Red pixels are extracted in a 3 ⁇ 3 kernel as shown earlier, as well as the Green pixels corresponding thereto.
- the system generates object-based parameters for sampled Red pixels in the kernel.
- the system generates weights for the object-based parameter, whilst again using the actual value of exponent k received from block 74 .
- a sum of the products of weights and measured Red pixels in the kernel generates the Red pixel value.
- the system checks whether the sample pixel was Blue. If yes, the way to mixer facility 112 lies open. If no, the system reverts to the blue reconstructing column further to the right.
- the various blocks 100 through 106 for this column correspond to their respective counterparts in the middle column, and further extensive discussion thereof is omitted. This column is also entered from block 68 if the sampled pixel were Red itself.
- the answer in block 116 will become yes. This means that in block 118 the full RGB image will have been recovered; the remainder of the flow-chart implements a few largely cosmetic refinements.
- the image is converted to an estimated color space with a chroma component applied.
- a low pass chroma filter is used to substantially remove false colors.
- the system executes a back conversion to the RGB color space. This will in block 126 yield the reconstructed RGB image that is subsequently presented to the DSP pipeline for further usage.
- a method may be regarded as having first and second processes for a first pixel.
- the first process computes a Green value for the pixel and the second process computes a Red value for the pixel.
- the first process includes extracting a first kernel from a multi-color matrix.
- the first kernel is a 5 by 5 sub matrix of the Bayer pattern centered on the exemplary Blue pixel.
- the first process further includes generating first variance weights (e.g., ⁇ H , ⁇ V ) from the first kernel and then generating a first color (Green, in this example) based on the first variance weights and adjacent pixel values of the first color.
- first variance weights e.g., ⁇ H , ⁇ V
- Green Green, in this example
- the generating of the first variance weights includes first determining horizontal and vertical gradient value averages (e.g., ⁇ H and ⁇ V ).
- the center of the first kernel is the exemplary Blue pixel, and the kernel includes first and second side pixels of the first color spaced apart horizontally.
- the first color is Green
- the first and second side pixels are Green pixels of the kernel that are spaced apart horizontally.
- the horizontal gradient value average e.g., ⁇ H
- the first process determines plural horizontal gradients (e.g., pairs of two horizontally spaced Green pixels) and then determines a gradient value for each gradient.
- a first horizontal gradient value is determined by calculating an absolute value of a difference between a value of the first side pixel (a Green pixel in this example) and a value of the second side pixel (another Green pixel in this example). The calculation of the horizontal gradient value is repeated for each horizontal gradient within the first kernel. Then, the first process calculates an average of the plural horizontal gradient values to determine the horizontal gradient value average (e.g., ⁇ H ). For example, five horizontal gradient values are determined for the gradients depicted in FIG. 2 . Then, the average is determined by summing the values and dividing by 5. The vertical gradient value average (e.g., ⁇ V ) is computed in a similar way. See FIG. 3 .
- the first variance weights include horizontal and vertical interpolation weights (e.g., ⁇ H , ⁇ V ).
- the first process determines the horizontal and vertical interpolation weights based on the horizontal and vertical gradient value averages and a predetermined exponent value, k.
- the horizontal interpolation weight is calculated based on
- the vertical interpolation weight is calculated based on
- ⁇ V 1 ⁇ k V ( ⁇ k H + ⁇ k V ),
- ⁇ k H is the horizontal gradient value average
- ⁇ k V is the vertical gradient value average
- k is the predetermined exponent value
- the second process includes extracting a second kernel from the multi-color matrix.
- the second kernel is a 3 by 3 sub matrix of the Bayer pattern centered on the exemplary Blue pixel.
- the second process further includes generating second variance offsets
- ⁇ g 2 G i ⁇ 1,j+1 ⁇ G i,j
- ⁇ g 4 G i+1,j+1 ⁇ G i,j
- ⁇ g 6 G i+1,j ⁇ 1 ⁇ G i,j
- ⁇ g 8 G i ⁇ 1,j ⁇ 1 ⁇ G i,j ).
- the second variance offsets are generated from the second kernel.
- the second kernel is a 3 by 3 sub matrix of the Bayer pattern centered on the exemplary Blue pixel.
- a Green value has been calculated to become associated with the exemplary Blue pixel, and associated with each of the four Red pixels at the four corners of the second kernel, as discussed herein. Also see FIG. 1 .
- second variance offsets e.g., ⁇ g 2 , ⁇ g 4 , ⁇ g 6 , and ⁇ g 8
- the second color e.g., Red in this example
- generating the second variance offsets includes determining diagonal gradients between pixels in the second kernel (e.g., [i, j] to [i ⁇ 1, j+1], [i, j] to [i+1, j+1], [i, j] to [i+1, j ⁇ 1] and [i, j] to [i ⁇ 1, j ⁇ 1]) and then determining gradient values of the first color (Green in this example)
- ⁇ g 2 G i ⁇ 1,j+1 ⁇ G i,j
- ⁇ g 4 G i+1,j+1 ⁇ G i,j
- ⁇ g 6 G i+1,j ⁇ 1 ⁇ G i,j
- ⁇ g 8 G i ⁇ 1,j ⁇ 1 ⁇ G i,j ).
- generating a second color includes choosing a minimum value of the gradient values (e.g., ⁇ g 2 ), selecting the adjacent pixel (e.g., [i ⁇ 1, j+1] to the right and above the center of the second kernel as depicted in FIG.
- the second color (Red, in this example) based on a gradient that corresponds to the minimum value, and subtracting the minimum value (e.g., ⁇ g 2 ) from a pixel value (e.g., R i ⁇ 1,j+1 ) of the selected adjacent pixel to determine the value for the second color (e.g., R i,j ) at the center of the second kernel (e.g., the exemplary Blue pixel).
- R i,j ( R i+1,j+1 ⁇ G i+1,j+1 +G i,j ),
- R i,j ( R i+1,j ⁇ 1 ⁇ G i+1,j ⁇ +G i,j ),
- R i,j ( R i ⁇ 1,j ⁇ 1 G i ⁇ 1,j ⁇ 1 +G i,j ).
- the second process would determine a Blue value for the center pixel.
- the method repeats for every pixel (except pixels on the edge of the matrix where a full kernel cannot be extracted) so as to determine all three color planes for each pixel.
- the method further includes a third process.
- the third process includes extracting a third kernel (e.g., a 3 by 3 sub matrix centered on the second pixel) from the multi-color matrix (e.g., the Bayer multi-color matrix described in this example).
- the third process further includes generating third variance offsets from the third kernel (e.g., ⁇ g 2 , ⁇ g 4 , ⁇ g 6 , and ⁇ g 8 ) in a process substantially the same as is described above with respect to the second process, and generating a third color (Blue in this example) based on the third variance offsets and an adjacent pixel of the third color in a process substantially the same as is described above with respect to the second process.
- the third kernel e.g., ⁇ g 2 , ⁇ g 4 , ⁇ g 6 , and ⁇ g 8
- the value of the predetermined exponent, k is determined in step 72 ( FIG. 6 ).
- the system noise level 70 is a matrix of elements corresponding to the elements of the captured Bayer image data, but representing system noise level. For example, a totally dark image could be captured in the sensor and loaded into matrix 70 , or an imaging device could contain a group of light shielded pixels used to determine k.
- An exemplary way in which k may be determined is to average the noise value over all elements to determine an overall average, and then compute a variance to this average over all elements of the matrix. The variance becomes input to a look up table where the value of k is read as an output. In this way, different loads of the look up table can be used to emphasize edge clarity in low-noise situations and avoid enhancing noise while maintaining minimal color aliasing in high-noise situations.
- the image in RGB space is converted to chroma space (C′ and C′′ as discussed above) in block 120 .
- the chroma components e.g., C′ and C′′
- a low pass chroma filter e.g., such as a moving window of a 3 by 3 kernel of elements where the chroma value of the center point takes up the average value over the 3 by 3 kernel.
- the above described method is implemented in a computing machine of some form that includes a processor and a memory.
- the memory stores a first module for controlling the processor to perform a first process for a first pixel.
- the first process includes extracting a first kernel from a multi-color matrix, generating first variance weights from the first kernel, and generating a first color based on the first variance weights and adjacent pixel values of the first color.
- the memory further stores a second module for controlling the processor to perform a second process for the first pixel.
- the second process includes extracting a second kernel from the multi-color matrix, generating second variance offsets from the second kernel, and generating a second color based on the second variance offsets and an adjacent pixel of the second color.
- the processor may be a programmable processor that is controlled by programming modules stored on a separate computer readable media or it may be a fixed designed that is controlled by the timing generator of the processor.
- the computer readable media has stored thereon a plurality of modules for controlling a processor.
- the plurality of modules include first and second modules.
- the first module controls the processor to perform a first process for a first pixel.
- the first process includes extracting a first kernel from a multi-color matrix, generating first variance weights from the first kernel, and generating a first color based on the first variance weights and adjacent pixel values of the first color.
- the second module controls the processor to perform a second process for the first pixel.
- the second process includes extracting a second kernel from the multi-color matrix, generating second variance offsets from the second kernel, and generating a second color based on the second variance offsets and an adjacent pixel of the second color.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Physics & Mathematics (AREA)
- Spectroscopy & Molecular Physics (AREA)
- Color Television Image Signal Generators (AREA)
Abstract
A method of color reconstruction includes a first process for a first pixel and a second process for the first pixel. The first process includes extracting a first kernel from a multi-color matrix, generating first variance weights from the first kernel, and generating a first color based on the first variance weights and adjacent pixel values of the first color. The second process includes extracting a second kernel from the multi-color matrix, generating second variance offsets from the second kernel, and generating a second color based on the second variance offsets and an adjacent pixel of the second color.
Description
- The priority benefits of the Oct. 10, 2002 filing dates of U.S. provisional applications Ser. Nos. 60/417,142 and 60/417,152 are hereby claimed.
- 1. Field of the Invention
- The present invention relates to a method and apparatus for object-based color reconstruction in a multi-color based sensor arrangement through maintaining coherence within objects and along edges, and a computer program and computer program product for controlling such method, and an image sensing facility comprising such apparatus. In particular, the invention relates to a method for executing object-based color reconstruction in a multi-color matrix-based sensor arrangement that estimates color value through determining local gradients among various luminance component values assessed.
- 2. Description of Related Art
- Multi-color matrix-based sensor arrangements that include color sensors that have one first luminance component sensed at a relatively higher spatial frequency and two further chrominance components sensed at relatively lower spatial frequencies are coming in rapidly expanding use. Fields of application include digital cameras, digital cinematography, video cameras, scientific imaging, broadcast cameras, surveillance, security monitoring, and various others. Now generally, each object point translates into a single-color of the corresponding image pixel. Inasmuch as, in the resulting image each pixel lacks two of the three sensor colors, reconstruction of three-color pixels requires an appreciable amount of processing. However, in earlier realizations, the resulting reconstructed array often had artificially colored, noise-enhanced, or blurred edges. In consequence, there exists a growing need to provide for an algorithmic procedure that will provide high-quality estimations for the color values of pixels that were originally sensed in a single color only, whilst requiring to access only a minimal memory facility and minimal power consumption, needing to access only a small number of image lines/columns at a time, and being independent of other DSP-based (i.e., Digital Signal Processing-based) algorithms, such as those for removing false colors or correcting color balance, overall influencing of hue or luminance, or others. As will be discussed more extensively hereinafter, the present embodiment generally utilizes only a small data kernel of optimally 5×5 pixels for Green, or minimally 3×3 pixels for Red and Blue. It will generally minimize perceptional error, it reconstructs frequencies above the Nyquist frequency, it has much lower color aliasing than many other algorithms, it will hardly expand existing defects, and it will only be marginally affected by image noise.
- United States Patent Application Publication 2002/0063789 A1 to Acharya et al and published on May 30, 2002 discloses a Color Filter Array and Color Interpolation Algorithm, and uses a classification to determine which pixels it will use in the interpolation. The present invention in contradistinction uses a soft decision method that is useful for adaptive adjusting for variations in system noise. Moreover, the reference singles out a very particular matrix design with a very high percentage of 75% Green pixels sensed. The present invention is much wider applicable.
- Moreover, the inventor has found that correct application of the algorithm will not remove fundamental signal noise. Rather, it will prevent enhancing of the noise. In particular, the inventor has recognized the corresponding behavior of the gradients among the various sensed colors.
- In consequence, amongst other things, it is an object of the present invention to use such corresponding behavior to attain an improved result against an investment of only a limited amount of processing complexity.
- By itself, co-pending patent application by the present inventor, titled “Method And Apparatus For Adaptive Pixel Correction Of A Multi-Color Matrix,” identified by Attorney Docket No. 12546 and assigned to the same assignee as the present application relates to the adaptive correcting of defective pixels and is incorporated herein by reference. The adaptive pixel correction of a multi-color matrix is based on soft decisions across the various color planes, and in particular based on taking into account the spatial response pattern of the human visual system. The result of the co-pending application may be taken as a starting point for applying the present invention, but such represents no express restriction.
- A method for object-based color reconstruction in a multi-color matrix-based sensor arrangement is based on color sensors that have one first color sensed at a relatively higher spatial frequency and two further colors sensed at relatively lower spatial frequencies.
- In particular, the method executes the following steps. A particular pixel not sensed in said first color has its first color value estimated through determining local gradients (82) among various first color values assessed, and in accordance with such gradients such estimating is executed through interpolating (84, 86) along relatively stronger edge informations. For a particular pixel not sensed in a particular further color that further color's value is estimated in a direction along with relatively smaller differences evaluated in the first color (94, 96). Advantageously, the method has the ability for limiting noise propagation.
- Now therefore, according to one of its aspects the invention a method includes a first process for a first pixel and a second process for the first pixel. The first process includes extracting a first kernel from a multi-color matrix, generating first variance weights from the first kernel, and generating a first color based on the first variance weights and adjacent pixel values of the first color. The second process includes extracting a second kernel from the multi-color matrix, generating second variance offsets from the second kernel, and generating a second color based on the second variance offsets and an adjacent pixel of the second color.
- The invention also relates to a computing machine that includes a processor and a memory. The memory stores a first module for controlling the processor to perform a first process for a first pixel. The first process includes extracting a first kernel from a multi-color matrix, generating first variance weights from the first kernel, and generating a first color based on the first variance weights and adjacent pixel values of the first color. The memory further stores a second module for controlling the processor to perform a second process for the first pixel. The second process includes extracting a second kernel from the multi-color matrix, generating second variance offsets from the second kernel, and generating a second color based on the second variance offsets and an adjacent pixel of the second color.
- The invention also relates to a computer readable medium having stored thereon a plurality of modules for controlling a processor. The plurality of modules includes a first module and a second module. The first module is for controlling the processor to perform a first process for the first pixel. The first process includes extracting a first kernel from a multi-color matrix, generating first variance weights from the first kernel, and generating a first color based on the first variance weights and adjacent pixel values of the first color. The second module is for controlling the processor to perform a second process for the first pixel. The second process includes extracting a second kernel from the multi-color matrix, generating second variance offsets from the second kernel, and generating a second color based on the second variance offsets and an adjacent pixel of the second color.
- Further advantageous aspects of the invention are as disclosed and claimed herein.
- The invention will be described in detail in the following description of preferred embodiments with reference to the following figures wherein:
-
FIG. 1 is a schematic view of a Bayer matrix of image points; -
FIG. 2 is a schematic view of the deriving of the horizontal Green gradient; -
FIG. 3 is a schematic view of the deriving of the vertical Green gradient; -
FIGS. 4 a, 4 b are schematic views of the relevance of the eight possible Green directions; -
FIG. 5 is a graph of a filter characteristic diagram for removing false color responses; and -
FIG. 6 is a schematic block diagram of a hardware implementation of the present invention. -
FIG. 1 illustrates a Bayer matrix of image points for use with the present invention. Alternatively, other arrangements could be used, such as through using subtractive colors, such as cyan, magenta, and yellow, or through various other published embodiments. As shown, the matrix has Green pixels arranged according to a chessboard organization. The remaining pixels have been used for Red pixels in odd rows and Blue pixels in even rows. - Now, the intended end result of the reconstruction embodiment is three complete images of Red, Green, and Blue pixel values, respectively, where the original recording did yield less than three values for each respective pixel, according to the pixel pattern for the respective single color. It is known from advance that at each pixel location, two values must be estimated. In practice, the application most favored is to a Bayer pattern wherein the green pixels form a chessboard pattern, while the non-green fields have alternating red and blue values in the diagonal directions.
- Therefore, first the Green matrix will be completed. In this respect,
FIG. 2 illustrates the deriving of the horizontal Green gradient or horizontal green edge indicator δH in a preferred kernel pattern of 5×5 pixels according to the average absolute differences between pixels pairs joined by an arrow, according to: -
δH ={|G i−1,j−2 −G i−1,j |+|G i−1,j −G i−1,j+2|+ . . . }/5. - Likewise,
FIG. 3 illustrates the deriving of the vertical Green gradient or vertical Green edge indicator δV. For the time being, the pixels assigned to other colors will be ignored. In certain alternative environments, it may be advantageous to define the kernel as having a size of 3×3 pixels only, through using the two arrows numbered 23 inFIG. 2 , and 27 inFIG. 3 , respectively. Apparently, the respective kernels for horizontal and vertical determinations will not anymore coincide here. - Then, in a relatively simplistic implementation a local edge direction is determined and a Green value for the missing central point is estimated as follows: if δV is relatively larger, interpolate horizontally between two nearest green neighbors; if δH is relatively larger, interpolate vertically between two nearest green neighbors, and if δH and δV are roughly equal, interpolate linearly between all four nearest neighbors of the missing Green pixel. Of course, the determining of the being “roughly” equal allows to introduce a certain amount of flexibility into the algorithm. The logical method presented here demonstrates the functioning of the actual algebraic method.
- A fuller realization as an algebraic expression for the interpolated green pixel is:
-
G i,j=1/2λH(G i,j−1 +G i,j+1)+1/2λv(G i−1,j +G i+1,j), - wherein λH=1−δk H/(δk H+δk V), and likewise, for the vertical parameter. Herein, a greater value of exponent constant k increases the algorithm's sensitivity to the differences in δH and δV.
- The algebraic expressions offer more benefits than just for presentation. The non-linear combination of the edge indicators can be adjusted with the value of k to either emphasize edge clarity in low-noise situations, or rather avoid the enhancing of the noise while still maintaining minimal color aliasing in high-noise situations. Therefore, k has a two-sided adaptive effect on the algorithm. With small values (between 1 and 2) noise will be adaptively minimized whilst ensuring that edges will have none to minimal color aliasing. On the other hand, with larger values for k in low-noise situations, more emphasis can be put on removing all color aliasing whilst enhancing edge continuity and sharpness. Further, the use of five arrows in
FIGS. 2 and 3 , instead of the minimal number of one, allows reconstruction of frequencies up to twice the Nyquist frequency fN (i.e. the spatial sample frequency) in both vertical and horizontal directions. In case of the smaller kernels, the limit would lie at the Nyquist frequency. - After having completed the Green array Gi,j, the missing points in the Red and Blue images will be estimated. For a Green sample on a Green-Blue row, the horizontal neighbors are Blue and the vertical neighbors are Red. First the Red value is sought for. To ensure the continuity of the object recorded at the given point, the vertically closest Green value to the actual pixel is determined and then its associated Red value copied to the actual pixel. For Blue, among the two neighbors, the horizontally closest Green value to the actual pixel is determined and then its associated Blue value copied to the actual pixel.
- For a Blue sample, one approach is to choose among the four diagonally adjacent Red values, whilst using the respective Green differences shown in
FIG. 4 b, as follows: -
R i,j=(R i−1,j+1 −G i−1,j+1 +G i,j) if Δg2=Δmin - i.e. add the minimum Green difference to the Red value, and correspondingly:
-
R i,j=(R i+1,j+1 −G i+1,j+1 +G i,j) if Δg4=Δmin -
R i,j=(R i+1,j−1 −G i+1,j−1 +G i,j) if Δg6=Δmin -
R i,j=(R i−1,j−1 −G i−1,j−1 +G i,j) if Δg8=Δmin - A corresponding version applies to the missing Blue values at pixels that were sampled Red.
- Now, although the color reconstruction according to the present invention's embodiment will generally be correct, the particular arrangement of the Bayer matrix, in combination with non-ideal optical image low-pass filtering and system noise can cause some false color responses. Heuristically, the following description is given. The sampling of a scene through a Bayer matrix will cause two major sampling errors, as follows: (a) The relative displacement between green, red and blue samples introduces a phase error between the various colors; (b) The sampling frequencies of Red and Blue are relatively lower than that of Green. This is perceptually advantageous, because the human eye's chroma and hue resolution is relatively less than its luminance resolution. On the other hand, this introduces different aliasing responses for Green versus for Red/Blue, in particular, inasmuch as Green will attain a better high-frequency response. Although this may cause false-color spikes and the like, such phenomenon is relatively rare. Such spikes can be considered high-frequency chromatic points, and low-pass filtering would generally remove them. In fact, a chroma image can be estimated through two Green differences
-
C′=(R−G)i,j, and C″=(B−G)i,j. - In this respect,
FIG. 5 shows the intended frequency response of a chroma filter. The design of such a-filter that has been shown in solid lines is dependent on the system's optical low-pass filtering as well as on system noise. Ideally, it will be designed to minimize the loss of image information and at the same time maximize the removal of false colors. To an appreciable extent, degradation in sharpness through the alpha-filter is mitigated through a beta-filter shown inFIG. 5 in dashed lines. -
FIG. 6 illustrates a hardware implementation of the present invention. Now, the output of the co-pending application referred to supra may be fed to aDSP pipeline 20 for processing and/or transient storage, as the case may be. The DSP pipeline is fed by the image sensing facility, not shown for brevity, whereas the pipeline itself has a sufficient amount of memory and processing facilities for implementing the method according to the present invention. In addition thereto, inblock 70, the system will evaluate the noise level in the image, and therefrom inblock 72 convert the noise level in the k-factor considered supra. From a holdingelement 74, this k-factor is provided to such elements of the hardware as are in need of it. Furthermore, for simplicity, command, control and synchronization streams have been left out of the embodiment for brevity and optimum clarification. - Now, block 22 represents the captured Bayer image information. In
block 60, it is detected whether the pixel in question is at least 2 pixels away from the image edge. If not, the procedure hereinafter will not have enough data available, so the process will then move to the next pixel inblock 62. Inblock 64, the system detects whether the sample pixel is Green. - If not, the system will undertake to reconstruct an appropriate Green value for the pixel in question, as follows. First, in
block 80, the system extracts in a kernel of 5×5 pixels disclosed earlier the sensed Green pixels. Inblock 82, the horizontal and vertical variance parameters are derived. Inblock 84, the horizontal and vertical weight generating functions are derived, through also receiving the actual value of the exponent k fromblock 74, as discussed earlier. Finally, inblock 86, the system calculates a product of the sum weights and the horizontally and vertically sampled Green pixels. With the so reconstructed Green pixel, the system then proceeds to mixer/accumulator 112 whilst inblock 114 storing the result and inblock 116 detecting whether the end of the image had been reached, which for the time being is not so. - If in
block 64 the sample is not Green, so inblock 66 the system shifts by one row and one column and checks for a Red pixel. If no, the pixel thus being Blue, the system will undertake to reconstruct a Red pixel. First, inblock 90, the Red pixels are extracted in a 3×3 kernel as shown earlier, as well as the Green pixels corresponding thereto. Next, inblock 92, the system generates object-based parameters for sampled Red pixels in the kernel. Next, in block inblock 94, the system generates weights for the object-based parameter, whilst again using the actual value of exponent k received fromblock 74. Finally, inblock 96, a sum of the products of weights and measured Red pixels in the kernel generates the Red pixel value. Finally, inblock 110, the system checks whether the sample pixel was Blue. If yes, the way tomixer facility 112 lies open. If no, the system reverts to the blue reconstructing column further to the right. Thevarious blocks 100 through 106 for this column correspond to their respective counterparts in the middle column, and further extensive discussion thereof is omitted. This column is also entered fromblock 68 if the sampled pixel were Red itself. - Generally, after numerous rounds through the reconstructing columns, the answer in
block 116 will become yes. This means that inblock 118 the full RGB image will have been recovered; the remainder of the flow-chart implements a few largely cosmetic refinements. Inblock 120, the image is converted to an estimated color space with a chroma component applied. Inblock 122, a low pass chroma filter is used to substantially remove false colors. Inblock 124, the system executes a back conversion to the RGB color space. This will inblock 126 yield the reconstructed RGB image that is subsequently presented to the DSP pipeline for further usage. - As an example of the above described embodiment, a method may be regarded as having first and second processes for a first pixel. For example, in the Bayer patterned multi-color matrix, if the first pixel is a Blue pixel, then the first process computes a Green value for the pixel and the second process computes a Red value for the pixel. The first process includes extracting a first kernel from a multi-color matrix. In the example discussed herein, the first kernel is a 5 by 5 sub matrix of the Bayer pattern centered on the exemplary Blue pixel. The first process further includes generating first variance weights (e.g., λH, λV) from the first kernel and then generating a first color (Green, in this example) based on the first variance weights and adjacent pixel values of the first color.
- The generating of the first variance weights includes first determining horizontal and vertical gradient value averages (e.g., δH and δV). The center of the first kernel is the exemplary Blue pixel, and the kernel includes first and second side pixels of the first color spaced apart horizontally. In this example, the first color is Green, and the first and second side pixels are Green pixels of the kernel that are spaced apart horizontally. To determine the horizontal gradient value average (e.g., δH), the first process determines plural horizontal gradients (e.g., pairs of two horizontally spaced Green pixels) and then determines a gradient value for each gradient. A first horizontal gradient value is determined by calculating an absolute value of a difference between a value of the first side pixel (a Green pixel in this example) and a value of the second side pixel (another Green pixel in this example). The calculation of the horizontal gradient value is repeated for each horizontal gradient within the first kernel. Then, the first process calculates an average of the plural horizontal gradient values to determine the horizontal gradient value average (e.g., δH). For example, five horizontal gradient values are determined for the gradients depicted in
FIG. 2 . Then, the average is determined by summing the values and dividing by 5. The vertical gradient value average (e.g., δV) is computed in a similar way. SeeFIG. 3 . - The first variance weights include horizontal and vertical interpolation weights (e.g., λH, λV). The first process determines the horizontal and vertical interpolation weights based on the horizontal and vertical gradient value averages and a predetermined exponent value, k. The horizontal interpolation weight is calculated based on
-
λH=1−δk H/(δk H+δk V),and - the vertical interpolation weight is calculated based on
-
λV=1−δk V(δk H+δk V), - where δk H is the horizontal gradient value average, δk V is the vertical gradient value average, and k is the predetermined exponent value.
- The second process includes extracting a second kernel from the multi-color matrix. In the example discussed herein, the second kernel is a 3 by 3 sub matrix of the Bayer pattern centered on the exemplary Blue pixel. The second process further includes generating second variance offsets
-
(e.g., Δg 2 =G i−1,j+1 −G i,j , Δg 4 =G i+1,j+1 −G i,j , Δg 6 =G i+1,j−1 −G i,j, and Δg 8 =G i−1,j−1 −G i,j). - The second variance offsets are generated from the second kernel. In the example discussed herein, the second kernel is a 3 by 3 sub matrix of the Bayer pattern centered on the exemplary Blue pixel. As discussed with respect to the first process, a Green value has been calculated to become associated with the exemplary Blue pixel, and associated with each of the four Red pixels at the four corners of the second kernel, as discussed herein. Also see
FIG. 1 . With these calculated Green values, second variance offsets (e.g., Δg2, Δg4, Δg6, and Δg8) are determined based on the respective differences between the calculated Green values as depicted inFIGS. 4 a and 4 b. Then, the second color (e.g., Red in this example) is generated based on the second variance offsets and an adjacent pixel of the second color. - More specifically, generating the second variance offsets includes determining diagonal gradients between pixels in the second kernel (e.g., [i, j] to [i−1, j+1], [i, j] to [i+1, j+1], [i, j] to [i+1, j−1] and [i, j] to [i−1, j−1]) and then determining gradient values of the first color (Green in this example)
-
(e.g., Δg 2 =G i−1,j+1 −G i,j , Δg 4 =G i+1,j+1 −G i,j , Δg 6 =G i+1,j−1 −G i,j, and Δg 8 =G i−1,j−1 −G i,j). - corresponding to the gradients.
- Then, generating a second color (Red, in this example) includes choosing a minimum value of the gradient values (e.g., Δg2), selecting the adjacent pixel (e.g., [i−1, j+1] to the right and above the center of the second kernel as depicted in
FIG. 4 b) of the second color (Red, in this example) based on a gradient that corresponds to the minimum value, and subtracting the minimum value (e.g., Δg2) from a pixel value (e.g., Ri−1,j+1) of the selected adjacent pixel to determine the value for the second color (e.g., Ri,j) at the center of the second kernel (e.g., the exemplary Blue pixel). - However, if Δg4=Δmin then,
-
R i,j=(R i+1,j+1 −G i+1,j+1 +G i,j), - or if Δg6=Δmin then,
-
R i,j=(R i+1,j−1 −G i+1,j− +G i,j), - or if Δg8=Δmin then,
-
R i,j=(R i−1,j−1 G i−1,j−1 +G i,j). - On the other hand, if the center of the first and second kernels were a Red pixel, the second process would determine a Blue value for the center pixel.
- The method repeats for every pixel (except pixels on the edge of the matrix where a full kernel cannot be extracted) so as to determine all three color planes for each pixel. For a second pixel, the method further includes a third process. The third process includes extracting a third kernel (e.g., a 3 by 3 sub matrix centered on the second pixel) from the multi-color matrix (e.g., the Bayer multi-color matrix described in this example). The third process further includes generating third variance offsets from the third kernel (e.g., Δg2, Δg4, Δg6, and Δg8) in a process substantially the same as is described above with respect to the second process, and generating a third color (Blue in this example) based on the third variance offsets and an adjacent pixel of the third color in a process substantially the same as is described above with respect to the second process.
- The value of the predetermined exponent, k, is determined in step 72 (
FIG. 6 ). Thesystem noise level 70 is a matrix of elements corresponding to the elements of the captured Bayer image data, but representing system noise level. For example, a totally dark image could be captured in the sensor and loaded intomatrix 70, or an imaging device could contain a group of light shielded pixels used to determine k. An exemplary way in which k may be determined is to average the noise value over all elements to determine an overall average, and then compute a variance to this average over all elements of the matrix. The variance becomes input to a look up table where the value of k is read as an output. In this way, different loads of the look up table can be used to emphasize edge clarity in low-noise situations and avoid enhancing noise while maintaining minimal color aliasing in high-noise situations. - After a full RGB image is recovered in block 118 (
FIG. 6 ), the image in RGB space is converted to chroma space (C′ and C″ as discussed above) inblock 120. Inblock 122, the chroma components (e.g., C′ and C″) are filtered to remove false colors and high frequency spikes in a low pass chroma filter (e.g., such as a moving window of a 3 by 3 kernel of elements where the chroma value of the center point takes up the average value over the 3 by 3 kernel). Then, inblock 124, the chroma space image is converted back into RGB space. - The processes described above for determining Red and Blue values require that Green values be determined for all pixels in the 3 by 3 kernel used to determine the Red values or the kernel used to determine the Blue values. As can be observed in
FIG. 6 , a one row and a one column delay is introduced inblock 66. This permits Green values to be determined one row and one column ahead of the determination of either Red or Blue values. Thus, when a 3 by 3 kernel is extracted to determine the Red values or to determine the Blue values, the entire kernel already has the Green value determined for each element of the 3 by 3 kernel. In this way, the process lends itself for implementation in a processor (be it general purpose or DSP) or an Application Specific Integrated Circuit (ASIC), but it could always be implemented in discrete circuits connected together. Regardless of the specific technology involved, the above described method is implemented in a computing machine of some form that includes a processor and a memory. The memory stores a first module for controlling the processor to perform a first process for a first pixel. The first process includes extracting a first kernel from a multi-color matrix, generating first variance weights from the first kernel, and generating a first color based on the first variance weights and adjacent pixel values of the first color. The memory further stores a second module for controlling the processor to perform a second process for the first pixel. The second process includes extracting a second kernel from the multi-color matrix, generating second variance offsets from the second kernel, and generating a second color based on the second variance offsets and an adjacent pixel of the second color. The processor may be a programmable processor that is controlled by programming modules stored on a separate computer readable media or it may be a fixed designed that is controlled by the timing generator of the processor. - In the event that the processes are implemented on a programmable computing machine (such as a general purpose processor or DSP), the details of above described method are included in program modules stored on a computer readable media. The computer readable media has stored thereon a plurality of modules for controlling a processor. The plurality of modules include first and second modules. The first module controls the processor to perform a first process for a first pixel. The first process includes extracting a first kernel from a multi-color matrix, generating first variance weights from the first kernel, and generating a first color based on the first variance weights and adjacent pixel values of the first color. The second module controls the processor to perform a second process for the first pixel. The second process includes extracting a second kernel from the multi-color matrix, generating second variance offsets from the second kernel, and generating a second color based on the second variance offsets and an adjacent pixel of the second color.
- Having described preferred embodiments of a novel method and apparatus for reconstructing pixel color values (which are intended to be illustrative and not limiting), it is noted that modifications and variations can be made by persons skilled in the art in light of the above teachings. It is therefore to be understood that changes may be made in the particular embodiments of the invention disclosed which are within the scope and spirit of the invention as defined by the appended claims.
- Having thus described the invention with the details and particularity required by the patent laws, what is claimed and desired protected by Letters Patent is set forth in the appended claims.
Claims (12)
1. A method for object-based color reconstruction in a multicolor matrix-based sensor arrangement comprising color sensors that have one first luminance component sensed at a relatively higher spatial frequency and two further chrominance components sensed at relatively lower spatial fancies, for a particular pixel not sensed in said first luminance component estimating its first color value through determining local gradients among various first luminance component values a said method being characterized by executing the following steps:
and in accordance with such local gradients executing such estimating through along relatively stronger edge informations, interpolating with a relatively greater weight factor, in favor over interpolating along relatively weaker edge informations with a relatively lesser weight factor,
and for a particular pixel not sensed in a particular further chrominance component value estimating that further chrominance component's value in a direction along with relatively smaller differences evaluated in said first luminance component.
2. A method as claimed in claim 1 , wherein further chrominance component's values of said particular pixel are interpolated using neighboring pixels' information as based on whether they are situated within a same imaged object.
3. A method as claimed in claim 1 and applied to a Bayer matrix wherein said first luminance component is green.
4. A method as claimed in Clam 2, when said first luminance component's value is estimated on the basis of a 5×5 pixel kernel centered on said particular point.
5. A met as claimed in claim 1 , whilst through exponential gradient values adjusting an exponent value (k) for emphasizing edge clarity in a low-noise situation, or rather limiting noise propagation whilst still mitigating for color aliasing.
6. A method as claimed in claim 1 , whilst adding a low-pass filter step after estimating non-sampled colors for mitigating false-color spikes.
7. A method as claimed in claim 6 , whilst supplementing said low pass filter step with relatively enhancing spatial high frequencies relatively far from said low-pass filter's discriminatory frequency for edge sharpness enhancement FIG. 5 ).
8. A computer program comprising program instruction for controlling a computer to implement a method according to one of claim 1 .
9. A computer program product as being represented with a tangible read-only computer memory medium or being carried by an electrical signal, and comprising program instructions for controlling a cower to implement a method according to one of claim 1 .
10. An apparatus being arranged for implementing a method as claimed in claim 1 .
11. An apparatus according to claim 10 , and executed as a filter facility for limiting noise propagation.
12. An image facility comprising an image forming facility for forming an image on an as claimed in claim 10 , and furthermore comprising a memory facility fed by said apparatus, processing means for dynamically interacting with pixel values in said memory and user output means for outputting a reconstructed user image.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US12/386,779 US20090263017A1 (en) | 2002-10-10 | 2009-04-23 | Method for reconstruction of pixel color values |
Applications Claiming Priority (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US41715202P | 2002-10-10 | 2002-10-10 | |
US41714202P | 2002-10-10 | 2002-10-10 | |
US65841303A | 2003-09-10 | 2003-09-10 | |
US12/386,779 US20090263017A1 (en) | 2002-10-10 | 2009-04-23 | Method for reconstruction of pixel color values |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US65841303A Division | 2002-10-10 | 2003-09-10 |
Publications (1)
Publication Number | Publication Date |
---|---|
US20090263017A1 true US20090263017A1 (en) | 2009-10-22 |
Family
ID=40223929
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US10/658,523 Active 2027-07-20 US7477781B1 (en) | 2002-10-10 | 2003-09-10 | Method and apparatus for adaptive pixel correction of multi-color matrix |
US12/386,779 Abandoned US20090263017A1 (en) | 2002-10-10 | 2009-04-23 | Method for reconstruction of pixel color values |
Family Applications Before (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US10/658,523 Active 2027-07-20 US7477781B1 (en) | 2002-10-10 | 2003-09-10 | Method and apparatus for adaptive pixel correction of multi-color matrix |
Country Status (1)
Country | Link |
---|---|
US (2) | US7477781B1 (en) |
Cited By (67)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20090116750A1 (en) * | 2006-05-30 | 2009-05-07 | Ho-Young Lee | Color interpolation method and device |
US20090310000A1 (en) * | 2008-06-17 | 2009-12-17 | Junichi Hosokawa | Solid-state imaging device |
US8693775B2 (en) * | 2011-03-29 | 2014-04-08 | Sony Corporation | Image processing apparatus, method, recording medium, and program |
US20140253808A1 (en) * | 2011-08-31 | 2014-09-11 | Sony Corporation | Image processing device, and image processing method, and program |
US20140347509A1 (en) * | 2008-05-20 | 2014-11-27 | Pelican Imaging Corporation | Capturing and Processing of Images Including Occlusions Captured by Arrays of Luma and Chroma Cameras |
US8929651B2 (en) | 2010-11-16 | 2015-01-06 | Thomson Licensing | System and method for the repair of anomalies in images |
US8995766B1 (en) * | 2013-09-27 | 2015-03-31 | Himax Imaging Limited | Image processing method and image processing device |
US9025894B2 (en) | 2011-09-28 | 2015-05-05 | Pelican Imaging Corporation | Systems and methods for decoding light field image files having depth and confidence maps |
US9041829B2 (en) | 2008-05-20 | 2015-05-26 | Pelican Imaging Corporation | Capturing and processing of high dynamic range images using camera arrays |
US9041824B2 (en) | 2010-12-14 | 2015-05-26 | Pelican Imaging Corporation | Systems and methods for dynamic refocusing of high resolution images generated using images captured by a plurality of imagers |
US9100635B2 (en) | 2012-06-28 | 2015-08-04 | Pelican Imaging Corporation | Systems and methods for detecting defective camera arrays and optic arrays |
US9100586B2 (en) | 2013-03-14 | 2015-08-04 | Pelican Imaging Corporation | Systems and methods for photometric normalization in array cameras |
US9106784B2 (en) | 2013-03-13 | 2015-08-11 | Pelican Imaging Corporation | Systems and methods for controlling aliasing in images captured by an array camera for use in super-resolution processing |
US9113118B2 (en) | 2012-05-31 | 2015-08-18 | Apple Inc. | Green non-uniformity correction |
US9123117B2 (en) | 2012-08-21 | 2015-09-01 | Pelican Imaging Corporation | Systems and methods for generating depth maps and corresponding confidence maps indicating depth estimation reliability |
US9128228B2 (en) | 2011-06-28 | 2015-09-08 | Pelican Imaging Corporation | Optical arrangements for use with an array camera |
US9143711B2 (en) | 2012-11-13 | 2015-09-22 | Pelican Imaging Corporation | Systems and methods for array camera focal plane control |
US9185276B2 (en) | 2013-11-07 | 2015-11-10 | Pelican Imaging Corporation | Methods of manufacturing array camera modules incorporating independently aligned lens stacks |
US9210392B2 (en) | 2012-05-01 | 2015-12-08 | Pelican Imaging Coporation | Camera modules patterned with pi filter groups |
US9214013B2 (en) | 2012-09-14 | 2015-12-15 | Pelican Imaging Corporation | Systems and methods for correcting user identified artifacts in light field images |
US9247117B2 (en) | 2014-04-07 | 2016-01-26 | Pelican Imaging Corporation | Systems and methods for correcting for warpage of a sensor array in an array camera module by introducing warpage into a focal plane of a lens stack array |
US9253380B2 (en) | 2013-02-24 | 2016-02-02 | Pelican Imaging Corporation | Thin form factor computational array cameras and modular array cameras |
US9264610B2 (en) | 2009-11-20 | 2016-02-16 | Pelican Imaging Corporation | Capturing and processing of images including occlusions captured by heterogeneous camera arrays |
US9412206B2 (en) | 2012-02-21 | 2016-08-09 | Pelican Imaging Corporation | Systems and methods for the manipulation of captured light field image data |
US9426361B2 (en) | 2013-11-26 | 2016-08-23 | Pelican Imaging Corporation | Array camera configurations incorporating multiple constituent array cameras |
US9438888B2 (en) | 2013-03-15 | 2016-09-06 | Pelican Imaging Corporation | Systems and methods for stereo imaging with camera arrays |
US9462164B2 (en) | 2013-02-21 | 2016-10-04 | Pelican Imaging Corporation | Systems and methods for generating compressed light field representation data using captured light fields, array geometry, and parallax information |
US9497370B2 (en) | 2013-03-15 | 2016-11-15 | Pelican Imaging Corporation | Array camera architecture implementing quantum dot color filters |
US9497429B2 (en) | 2013-03-15 | 2016-11-15 | Pelican Imaging Corporation | Extended color processing on pelican array cameras |
US9516222B2 (en) | 2011-06-28 | 2016-12-06 | Kip Peli P1 Lp | Array cameras incorporating monolithic array camera modules with high MTF lens stacks for capture of images used in super-resolution processing |
US9521416B1 (en) | 2013-03-11 | 2016-12-13 | Kip Peli P1 Lp | Systems and methods for image data compression |
US9519972B2 (en) | 2013-03-13 | 2016-12-13 | Kip Peli P1 Lp | Systems and methods for synthesizing images from image data captured by an array camera using restricted depth of field depth maps in which depth estimation precision varies |
US9521319B2 (en) | 2014-06-18 | 2016-12-13 | Pelican Imaging Corporation | Array cameras and array camera modules including spectral filters disposed outside of a constituent image sensor |
US9578259B2 (en) | 2013-03-14 | 2017-02-21 | Fotonation Cayman Limited | Systems and methods for reducing motion blur in images or video in ultra low light with array cameras |
US9633442B2 (en) | 2013-03-15 | 2017-04-25 | Fotonation Cayman Limited | Array cameras including an array camera module augmented with a separate camera |
US9741118B2 (en) | 2013-03-13 | 2017-08-22 | Fotonation Cayman Limited | System and methods for calibration of an array camera |
US9766380B2 (en) | 2012-06-30 | 2017-09-19 | Fotonation Cayman Limited | Systems and methods for manufacturing camera modules using active alignment of lens stack arrays and sensors |
US9774789B2 (en) | 2013-03-08 | 2017-09-26 | Fotonation Cayman Limited | Systems and methods for high dynamic range imaging using array cameras |
US9794476B2 (en) | 2011-09-19 | 2017-10-17 | Fotonation Cayman Limited | Systems and methods for controlling aliasing in images captured by an array camera for use in super resolution processing using pixel apertures |
US9813616B2 (en) | 2012-08-23 | 2017-11-07 | Fotonation Cayman Limited | Feature based high resolution motion estimation from low resolution images captured using an array source |
US9866739B2 (en) | 2011-05-11 | 2018-01-09 | Fotonation Cayman Limited | Systems and methods for transmitting and receiving array camera image data |
US9888194B2 (en) | 2013-03-13 | 2018-02-06 | Fotonation Cayman Limited | Array camera architecture implementing quantum film image sensors |
US9898856B2 (en) | 2013-09-27 | 2018-02-20 | Fotonation Cayman Limited | Systems and methods for depth-assisted perspective distortion correction |
US9936148B2 (en) | 2010-05-12 | 2018-04-03 | Fotonation Cayman Limited | Imager array interfaces |
US9942474B2 (en) | 2015-04-17 | 2018-04-10 | Fotonation Cayman Limited | Systems and methods for performing high speed video capture and depth estimation using array cameras |
US9955070B2 (en) | 2013-03-15 | 2018-04-24 | Fotonation Cayman Limited | Systems and methods for synthesizing high resolution images using image deconvolution based on motion and depth information |
US9986224B2 (en) | 2013-03-10 | 2018-05-29 | Fotonation Cayman Limited | System and methods for calibration of an array camera |
US10089740B2 (en) | 2014-03-07 | 2018-10-02 | Fotonation Limited | System and methods for depth regularization and semiautomatic interactive matting using RGB-D images |
TWI638339B (en) * | 2017-11-14 | 2018-10-11 | 瑞昱半導體股份有限公司 | False color removal method |
US10119808B2 (en) | 2013-11-18 | 2018-11-06 | Fotonation Limited | Systems and methods for estimating depth from projected texture using camera arrays |
US10122993B2 (en) | 2013-03-15 | 2018-11-06 | Fotonation Limited | Autofocus system for a conventional camera that uses depth information from an array camera |
US10250871B2 (en) | 2014-09-29 | 2019-04-02 | Fotonation Limited | Systems and methods for dynamic calibration of array cameras |
US10390005B2 (en) | 2012-09-28 | 2019-08-20 | Fotonation Limited | Generating images from light fields utilizing virtual viewpoints |
US10482618B2 (en) | 2017-08-21 | 2019-11-19 | Fotonation Limited | Systems and methods for hybrid depth regularization |
US11270110B2 (en) | 2019-09-17 | 2022-03-08 | Boston Polarimetrics, Inc. | Systems and methods for surface modeling using polarization cues |
US11290658B1 (en) | 2021-04-15 | 2022-03-29 | Boston Polarimetrics, Inc. | Systems and methods for camera exposure control |
US11302012B2 (en) | 2019-11-30 | 2022-04-12 | Boston Polarimetrics, Inc. | Systems and methods for transparent object segmentation using polarization cues |
US11525906B2 (en) | 2019-10-07 | 2022-12-13 | Intrinsic Innovation Llc | Systems and methods for augmentation of sensor systems and imaging systems with polarization |
US11580667B2 (en) | 2020-01-29 | 2023-02-14 | Intrinsic Innovation Llc | Systems and methods for characterizing object pose detection and measurement systems |
US11689813B2 (en) | 2021-07-01 | 2023-06-27 | Intrinsic Innovation Llc | Systems and methods for high dynamic range imaging using crossed polarizers |
US11792538B2 (en) | 2008-05-20 | 2023-10-17 | Adeia Imaging Llc | Capturing and processing of images including occlusions focused on an image sensor by a lens stack array |
US11797863B2 (en) | 2020-01-30 | 2023-10-24 | Intrinsic Innovation Llc | Systems and methods for synthesizing data for training statistical models on different imaging modalities including polarized images |
US11954886B2 (en) | 2021-04-15 | 2024-04-09 | Intrinsic Innovation Llc | Systems and methods for six-degree of freedom pose estimation of deformable objects |
US11953700B2 (en) | 2020-05-27 | 2024-04-09 | Intrinsic Innovation Llc | Multi-aperture polarization optical systems using beam splitters |
US12020455B2 (en) | 2021-03-10 | 2024-06-25 | Intrinsic Innovation Llc | Systems and methods for high dynamic range image reconstruction |
US12067746B2 (en) | 2021-05-07 | 2024-08-20 | Intrinsic Innovation Llc | Systems and methods for using computer vision to pick up small objects |
US12069227B2 (en) | 2021-03-10 | 2024-08-20 | Intrinsic Innovation Llc | Multi-modal and multi-spectral stereo camera arrays |
Families Citing this family (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7620241B2 (en) * | 2004-11-30 | 2009-11-17 | Hewlett-Packard Development Company, L.P. | Artifact reduction in a digital video |
KR101503227B1 (en) * | 2007-04-11 | 2015-03-16 | 레드.컴 인코포레이티드 | Video camera |
US8237830B2 (en) | 2007-04-11 | 2012-08-07 | Red.Com, Inc. | Video camera |
US20080273102A1 (en) * | 2007-05-01 | 2008-11-06 | Hoya Corporation | Detection device for defective pixel in photographic device |
KR101340518B1 (en) * | 2007-08-23 | 2013-12-11 | 삼성전기주식회사 | Method and apparatus for compensating chromatic aberration of an image |
JP6046927B2 (en) | 2011-08-09 | 2016-12-21 | キヤノン株式会社 | Image processing apparatus and control method thereof |
US8891866B2 (en) | 2011-08-31 | 2014-11-18 | Sony Corporation | Image processing apparatus, image processing method, and program |
WO2014127153A1 (en) | 2013-02-14 | 2014-08-21 | Red. Com, Inc. | Video camera |
US9088740B2 (en) * | 2013-06-21 | 2015-07-21 | Himax Imaging Limited | System and method of reducing noise |
GB2521408B (en) * | 2013-12-18 | 2015-12-16 | Imagination Tech Ltd | Defective pixel fixing |
WO2019010233A1 (en) | 2017-07-05 | 2019-01-10 | Red. Com, Llc | Video image data processing in electronic devices |
Citations (44)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4868655A (en) * | 1987-11-09 | 1989-09-19 | Etat Francais Represente Par Le Ministre Delegue Des Postes Et Telecommunications (Centre National D'etudes Des Telecommunications) | Method and apparatus for processing picture signals having interlaced field scanning |
US5291293A (en) * | 1992-06-01 | 1994-03-01 | Eastman Kodak Company | Electronic imaging device with defect correction |
US5596367A (en) * | 1996-02-23 | 1997-01-21 | Eastman Kodak Company | Averaging green values for green photosites in electronic cameras |
US5629734A (en) * | 1995-03-17 | 1997-05-13 | Eastman Kodak Company | Adaptive color plan interpolation in single sensor color electronic camera |
US5631703A (en) * | 1996-05-29 | 1997-05-20 | Eastman Kodak Company | Particular pattern of pixels for a color filter array which is used to derive luminance and chrominance values |
US5652621A (en) * | 1996-02-23 | 1997-07-29 | Eastman Kodak Company | Adaptive color plane interpolation in single sensor color electronic camera |
US5699167A (en) * | 1993-09-30 | 1997-12-16 | Canon Kabushiki Kaisha | Image forming apparatus which outputs a color image by separating color image information into at least two color components |
US5748803A (en) * | 1995-11-22 | 1998-05-05 | Linotype-Hell Ag | Method and apparatus for image value correction in optoelectronic transducers |
US5805217A (en) * | 1996-06-14 | 1998-09-08 | Iterated Systems, Inc. | Method and system for interpolating missing picture elements in a single color component array obtained from a single color sensor |
US5881182A (en) * | 1997-05-12 | 1999-03-09 | Eastman Kodak Company | Adaptive process for removing streaks in digital images |
US5990950A (en) * | 1998-02-11 | 1999-11-23 | Iterated Systems, Inc. | Method and system for color filter array multifactor interpolation |
US6078686A (en) * | 1996-09-30 | 2000-06-20 | Samsung Electronics Co., Ltd. | Image quality enhancement circuit and method therefor |
US6091851A (en) * | 1997-11-03 | 2000-07-18 | Intel Corporation | Efficient algorithm for color recovery from 8-bit to 24-bit color pixels |
US6229578B1 (en) * | 1997-12-08 | 2001-05-08 | Intel Corporation | Edge-detection based noise removal algorithm |
US6236433B1 (en) * | 1998-09-29 | 2001-05-22 | Intel Corporation | Scaling algorithm for efficient color representation/recovery in video |
US6268939B1 (en) * | 1998-01-08 | 2001-07-31 | Xerox Corporation | Method and apparatus for correcting luminance and chrominance data in digital color images |
US6330029B1 (en) * | 1998-03-17 | 2001-12-11 | Eastman Kodak Company | Particular pattern of pixels for a color filter array which is used to derive luminance and chrominance values |
US20020012476A1 (en) * | 2000-06-23 | 2002-01-31 | Dillen Bartholomeus Goverdina Maria Henricus | Image sensor signal defect correction |
US20020015111A1 (en) * | 2000-06-30 | 2002-02-07 | Yoshihito Harada | Image processing apparatus and its processing method |
US20020031257A1 (en) * | 2000-09-12 | 2002-03-14 | Shinichi Kato | Image processing apparatus and method |
US20020054219A1 (en) * | 2000-06-29 | 2002-05-09 | Jaspers Cornelis Antonie Maria | Processing a sensor output signal |
US6392699B1 (en) * | 1998-03-04 | 2002-05-21 | Intel Corporation | Integrated color interpolation and color space conversion algorithm from 8-bit bayer pattern RGB color space to 12-bit YCrCb color space |
US20020063789A1 (en) * | 2000-11-30 | 2002-05-30 | Tinku Acharya | Color filter array and color interpolation algorithm |
US6421084B1 (en) * | 1998-03-02 | 2002-07-16 | Compaq Computer Corporation | Method for interpolating a full color image from a single sensor using multiple threshold-based gradients |
US20020136463A1 (en) * | 2001-01-24 | 2002-09-26 | Taisuke Akahori | Image processing apparatus, image forming apparatus, and image processing method |
US20020141654A1 (en) * | 2001-03-29 | 2002-10-03 | Dean Rosales | Providing multiple symmetrical filters |
US6487304B1 (en) * | 1999-06-16 | 2002-11-26 | Microsoft Corporation | Multi-view approach to motion and stereo |
US20030002747A1 (en) * | 2001-07-02 | 2003-01-02 | Jasc Software,Inc | Moire correction in images |
US6535632B1 (en) * | 1998-12-18 | 2003-03-18 | University Of Washington | Image processing in HSI color space using adaptive noise filtering |
US6674903B1 (en) * | 1998-10-05 | 2004-01-06 | Agfa-Gevaert | Method for smoothing staircase effect in enlarged low resolution images |
US6721458B1 (en) * | 2000-04-14 | 2004-04-13 | Seiko Epson Corporation | Artifact reduction using adaptive nonlinear filters |
US6798910B1 (en) * | 2001-05-17 | 2004-09-28 | The United States Of America As Represented By The Secretary Of The Air Force | Self-optimizing edge detection in blurred, high-noise images |
US6823086B1 (en) * | 2000-08-29 | 2004-11-23 | Analogic Corporation | Adaptive spatial filter |
US20050047675A1 (en) * | 1999-09-16 | 2005-03-03 | Walmsley Simon Robert | Method of sharpening image using luminance channel |
US6956582B2 (en) * | 2001-08-23 | 2005-10-18 | Evans & Sutherland Computer Corporation | System and method for auto-adjusting image filtering |
US7027637B2 (en) * | 2002-02-21 | 2006-04-11 | Siemens Corporate Research, Inc. | Adaptive threshold determination for ball grid array component modeling |
US7145693B2 (en) * | 2001-08-08 | 2006-12-05 | Canon Kabushiki Kaisha | Image processing apparatus and method |
US20070053585A1 (en) * | 2005-05-31 | 2007-03-08 | Microsoft Corporation | Accelerated face detection based on prior probability of a view |
US7231084B2 (en) * | 2002-09-27 | 2007-06-12 | Motorola, Inc. | Color data image acquistion and processing |
US7256828B2 (en) * | 2003-01-16 | 2007-08-14 | Dialog Imaging Systems Gmbh | Weighted gradient based and color corrected interpolation |
US7272265B2 (en) * | 1998-03-13 | 2007-09-18 | The University Of Houston System | Methods for performing DAF data filtering and padding |
US20070292022A1 (en) * | 2003-01-16 | 2007-12-20 | Andreas Nilsson | Weighted gradient based and color corrected interpolation |
US7409084B2 (en) * | 2004-05-06 | 2008-08-05 | Magnachip Semiconductor, Ltd. | Method for determining dark condition and method for interpolating color in image sensor |
US7630546B2 (en) * | 2003-06-12 | 2009-12-08 | Nikon Corporation | Image processing method, image processing program and image processor |
Family Cites Families (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
BE1007354A3 (en) | 1993-07-23 | 1995-05-23 | Philips Electronics Nv | Signal correction circuit. |
US5778106A (en) | 1996-03-14 | 1998-07-07 | Polaroid Corporation | Electronic camera with reduced color artifacts |
US6697534B1 (en) * | 1999-06-09 | 2004-02-24 | Intel Corporation | Method and apparatus for adaptively sharpening local image content of an image |
US7035475B1 (en) * | 1999-06-17 | 2006-04-25 | Raytheon Company | Non-traditional adaptive non-uniformity compensation (ADNUC) system employing adaptive feedforward shunting and operating methods therefor |
US6832009B1 (en) * | 1999-09-24 | 2004-12-14 | Zoran Corporation | Method and apparatus for improved image interpolation |
US6724945B1 (en) | 2000-05-24 | 2004-04-20 | Hewlett-Packard Development Company, L.P. | Correcting defect pixels in a digital image |
JP3636046B2 (en) | 2000-07-31 | 2005-04-06 | 株式会社日立国際電気 | Pixel defect detection method for solid-state image sensor and imaging device using the method |
US7003173B2 (en) * | 2001-06-12 | 2006-02-21 | Sharp Laboratories Of America, Inc. | Filter for combined de-ringing and edge sharpening |
US6868179B2 (en) * | 2001-07-06 | 2005-03-15 | Jasc Software, Inc. | Automatic saturation adjustment |
GB0125774D0 (en) * | 2001-10-26 | 2001-12-19 | Cableform Ltd | Method and apparatus for image matching |
WO2004114063A2 (en) * | 2003-06-13 | 2004-12-29 | Georgia Tech Research Corporation | Data reconstruction using directional interpolation techniques |
US7373013B2 (en) * | 2003-12-23 | 2008-05-13 | General Instrument Corporation | Directional video filters for locally adaptive spatial noise reduction |
-
2003
- 2003-09-10 US US10/658,523 patent/US7477781B1/en active Active
-
2009
- 2009-04-23 US US12/386,779 patent/US20090263017A1/en not_active Abandoned
Patent Citations (45)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4868655A (en) * | 1987-11-09 | 1989-09-19 | Etat Francais Represente Par Le Ministre Delegue Des Postes Et Telecommunications (Centre National D'etudes Des Telecommunications) | Method and apparatus for processing picture signals having interlaced field scanning |
US5291293A (en) * | 1992-06-01 | 1994-03-01 | Eastman Kodak Company | Electronic imaging device with defect correction |
US5699167A (en) * | 1993-09-30 | 1997-12-16 | Canon Kabushiki Kaisha | Image forming apparatus which outputs a color image by separating color image information into at least two color components |
US5629734A (en) * | 1995-03-17 | 1997-05-13 | Eastman Kodak Company | Adaptive color plan interpolation in single sensor color electronic camera |
US5748803A (en) * | 1995-11-22 | 1998-05-05 | Linotype-Hell Ag | Method and apparatus for image value correction in optoelectronic transducers |
US5652621A (en) * | 1996-02-23 | 1997-07-29 | Eastman Kodak Company | Adaptive color plane interpolation in single sensor color electronic camera |
US5596367A (en) * | 1996-02-23 | 1997-01-21 | Eastman Kodak Company | Averaging green values for green photosites in electronic cameras |
US5631703A (en) * | 1996-05-29 | 1997-05-20 | Eastman Kodak Company | Particular pattern of pixels for a color filter array which is used to derive luminance and chrominance values |
US5805217A (en) * | 1996-06-14 | 1998-09-08 | Iterated Systems, Inc. | Method and system for interpolating missing picture elements in a single color component array obtained from a single color sensor |
US6078686A (en) * | 1996-09-30 | 2000-06-20 | Samsung Electronics Co., Ltd. | Image quality enhancement circuit and method therefor |
US5881182A (en) * | 1997-05-12 | 1999-03-09 | Eastman Kodak Company | Adaptive process for removing streaks in digital images |
US6091851A (en) * | 1997-11-03 | 2000-07-18 | Intel Corporation | Efficient algorithm for color recovery from 8-bit to 24-bit color pixels |
US6229578B1 (en) * | 1997-12-08 | 2001-05-08 | Intel Corporation | Edge-detection based noise removal algorithm |
US6268939B1 (en) * | 1998-01-08 | 2001-07-31 | Xerox Corporation | Method and apparatus for correcting luminance and chrominance data in digital color images |
US5990950A (en) * | 1998-02-11 | 1999-11-23 | Iterated Systems, Inc. | Method and system for color filter array multifactor interpolation |
US6421084B1 (en) * | 1998-03-02 | 2002-07-16 | Compaq Computer Corporation | Method for interpolating a full color image from a single sensor using multiple threshold-based gradients |
US6392699B1 (en) * | 1998-03-04 | 2002-05-21 | Intel Corporation | Integrated color interpolation and color space conversion algorithm from 8-bit bayer pattern RGB color space to 12-bit YCrCb color space |
US7272265B2 (en) * | 1998-03-13 | 2007-09-18 | The University Of Houston System | Methods for performing DAF data filtering and padding |
US6330029B1 (en) * | 1998-03-17 | 2001-12-11 | Eastman Kodak Company | Particular pattern of pixels for a color filter array which is used to derive luminance and chrominance values |
US6236433B1 (en) * | 1998-09-29 | 2001-05-22 | Intel Corporation | Scaling algorithm for efficient color representation/recovery in video |
US6674903B1 (en) * | 1998-10-05 | 2004-01-06 | Agfa-Gevaert | Method for smoothing staircase effect in enlarged low resolution images |
US6535632B1 (en) * | 1998-12-18 | 2003-03-18 | University Of Washington | Image processing in HSI color space using adaptive noise filtering |
US6487304B1 (en) * | 1999-06-16 | 2002-11-26 | Microsoft Corporation | Multi-view approach to motion and stereo |
US20050047675A1 (en) * | 1999-09-16 | 2005-03-03 | Walmsley Simon Robert | Method of sharpening image using luminance channel |
US6721458B1 (en) * | 2000-04-14 | 2004-04-13 | Seiko Epson Corporation | Artifact reduction using adaptive nonlinear filters |
US20020012476A1 (en) * | 2000-06-23 | 2002-01-31 | Dillen Bartholomeus Goverdina Maria Henricus | Image sensor signal defect correction |
US20020054219A1 (en) * | 2000-06-29 | 2002-05-09 | Jaspers Cornelis Antonie Maria | Processing a sensor output signal |
US20020015111A1 (en) * | 2000-06-30 | 2002-02-07 | Yoshihito Harada | Image processing apparatus and its processing method |
US6823086B1 (en) * | 2000-08-29 | 2004-11-23 | Analogic Corporation | Adaptive spatial filter |
US20020031257A1 (en) * | 2000-09-12 | 2002-03-14 | Shinichi Kato | Image processing apparatus and method |
US20020063789A1 (en) * | 2000-11-30 | 2002-05-30 | Tinku Acharya | Color filter array and color interpolation algorithm |
US20020136463A1 (en) * | 2001-01-24 | 2002-09-26 | Taisuke Akahori | Image processing apparatus, image forming apparatus, and image processing method |
US20020141654A1 (en) * | 2001-03-29 | 2002-10-03 | Dean Rosales | Providing multiple symmetrical filters |
US6798910B1 (en) * | 2001-05-17 | 2004-09-28 | The United States Of America As Represented By The Secretary Of The Air Force | Self-optimizing edge detection in blurred, high-noise images |
US20050220360A1 (en) * | 2001-07-02 | 2005-10-06 | Zaklika Krzysztof A | Moire correction in images |
US20030002747A1 (en) * | 2001-07-02 | 2003-01-02 | Jasc Software,Inc | Moire correction in images |
US7145693B2 (en) * | 2001-08-08 | 2006-12-05 | Canon Kabushiki Kaisha | Image processing apparatus and method |
US6956582B2 (en) * | 2001-08-23 | 2005-10-18 | Evans & Sutherland Computer Corporation | System and method for auto-adjusting image filtering |
US7027637B2 (en) * | 2002-02-21 | 2006-04-11 | Siemens Corporate Research, Inc. | Adaptive threshold determination for ball grid array component modeling |
US7231084B2 (en) * | 2002-09-27 | 2007-06-12 | Motorola, Inc. | Color data image acquistion and processing |
US7256828B2 (en) * | 2003-01-16 | 2007-08-14 | Dialog Imaging Systems Gmbh | Weighted gradient based and color corrected interpolation |
US20070292022A1 (en) * | 2003-01-16 | 2007-12-20 | Andreas Nilsson | Weighted gradient based and color corrected interpolation |
US7630546B2 (en) * | 2003-06-12 | 2009-12-08 | Nikon Corporation | Image processing method, image processing program and image processor |
US7409084B2 (en) * | 2004-05-06 | 2008-08-05 | Magnachip Semiconductor, Ltd. | Method for determining dark condition and method for interpolating color in image sensor |
US20070053585A1 (en) * | 2005-05-31 | 2007-03-08 | Microsoft Corporation | Accelerated face detection based on prior probability of a view |
Cited By (180)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20090116750A1 (en) * | 2006-05-30 | 2009-05-07 | Ho-Young Lee | Color interpolation method and device |
US9049411B2 (en) | 2008-05-20 | 2015-06-02 | Pelican Imaging Corporation | Camera arrays incorporating 3×3 imager configurations |
US11412158B2 (en) | 2008-05-20 | 2022-08-09 | Fotonation Limited | Capturing and processing of images including occlusions focused on an image sensor by a lens stack array |
US9712759B2 (en) | 2008-05-20 | 2017-07-18 | Fotonation Cayman Limited | Systems and methods for generating depth maps using a camera arrays incorporating monochrome and color cameras |
US9049391B2 (en) | 2008-05-20 | 2015-06-02 | Pelican Imaging Corporation | Capturing and processing of near-IR images including occlusions using camera arrays incorporating near-IR light sources |
US20140347509A1 (en) * | 2008-05-20 | 2014-11-27 | Pelican Imaging Corporation | Capturing and Processing of Images Including Occlusions Captured by Arrays of Luma and Chroma Cameras |
US9124815B2 (en) * | 2008-05-20 | 2015-09-01 | Pelican Imaging Corporation | Capturing and processing of images including occlusions captured by arrays of luma and chroma cameras |
US9576369B2 (en) | 2008-05-20 | 2017-02-21 | Fotonation Cayman Limited | Systems and methods for generating depth maps using images captured by camera arrays incorporating cameras having different fields of view |
US11792538B2 (en) | 2008-05-20 | 2023-10-17 | Adeia Imaging Llc | Capturing and processing of images including occlusions focused on an image sensor by a lens stack array |
US9485496B2 (en) | 2008-05-20 | 2016-11-01 | Pelican Imaging Corporation | Systems and methods for measuring depth using images captured by a camera array including cameras surrounding a central camera |
US10027901B2 (en) | 2008-05-20 | 2018-07-17 | Fotonation Cayman Limited | Systems and methods for generating depth maps using a camera arrays incorporating monochrome and color cameras |
US10142560B2 (en) | 2008-05-20 | 2018-11-27 | Fotonation Limited | Capturing and processing of images including occlusions focused on an image sensor by a lens stack array |
US9188765B2 (en) | 2008-05-20 | 2015-11-17 | Pelican Imaging Corporation | Capturing and processing of images including occlusions focused on an image sensor by a lens stack array |
US9041829B2 (en) | 2008-05-20 | 2015-05-26 | Pelican Imaging Corporation | Capturing and processing of high dynamic range images using camera arrays |
US9041823B2 (en) | 2008-05-20 | 2015-05-26 | Pelican Imaging Corporation | Systems and methods for performing post capture refocus using images captured by camera arrays |
US9191580B2 (en) | 2008-05-20 | 2015-11-17 | Pelican Imaging Corporation | Capturing and processing of images including occlusions captured by camera arrays |
US9235898B2 (en) | 2008-05-20 | 2016-01-12 | Pelican Imaging Corporation | Systems and methods for generating depth maps using light focused on an image sensor by a lens element array |
US9049390B2 (en) | 2008-05-20 | 2015-06-02 | Pelican Imaging Corporation | Capturing and processing of images captured by arrays including polychromatic cameras |
US9749547B2 (en) | 2008-05-20 | 2017-08-29 | Fotonation Cayman Limited | Capturing and processing of images using camera array incorperating Bayer cameras having different fields of view |
US12022207B2 (en) | 2008-05-20 | 2024-06-25 | Adeia Imaging Llc | Capturing and processing of images including occlusions focused on an image sensor by a lens stack array |
US12041360B2 (en) | 2008-05-20 | 2024-07-16 | Adeia Imaging Llc | Capturing and processing of images including occlusions focused on an image sensor by a lens stack array |
US9049367B2 (en) | 2008-05-20 | 2015-06-02 | Pelican Imaging Corporation | Systems and methods for synthesizing higher resolution images using images captured by camera arrays |
US9049381B2 (en) | 2008-05-20 | 2015-06-02 | Pelican Imaging Corporation | Systems and methods for normalizing image data captured by camera arrays |
US9055233B2 (en) | 2008-05-20 | 2015-06-09 | Pelican Imaging Corporation | Systems and methods for synthesizing higher resolution images using a set of images containing a baseline image |
US9055213B2 (en) | 2008-05-20 | 2015-06-09 | Pelican Imaging Corporation | Systems and methods for measuring depth using images captured by monolithic camera arrays including at least one bayer camera |
US9060124B2 (en) | 2008-05-20 | 2015-06-16 | Pelican Imaging Corporation | Capturing and processing of images using non-monolithic camera arrays |
US9060121B2 (en) | 2008-05-20 | 2015-06-16 | Pelican Imaging Corporation | Capturing and processing of images captured by camera arrays including cameras dedicated to sampling luma and cameras dedicated to sampling chroma |
US9060142B2 (en) | 2008-05-20 | 2015-06-16 | Pelican Imaging Corporation | Capturing and processing of images captured by camera arrays including heterogeneous optics |
US9060120B2 (en) | 2008-05-20 | 2015-06-16 | Pelican Imaging Corporation | Systems and methods for generating depth maps using images captured by camera arrays |
US9077893B2 (en) | 2008-05-20 | 2015-07-07 | Pelican Imaging Corporation | Capturing and processing of images captured by non-grid camera arrays |
US9094661B2 (en) | 2008-05-20 | 2015-07-28 | Pelican Imaging Corporation | Systems and methods for generating depth maps using a set of images containing a baseline image |
US8077253B2 (en) * | 2008-06-17 | 2011-12-13 | Kabushiki Kaisha Toshiba | Solid-state device having digital signal processing circuit |
US20090310000A1 (en) * | 2008-06-17 | 2009-12-17 | Junichi Hosokawa | Solid-state imaging device |
US9264610B2 (en) | 2009-11-20 | 2016-02-16 | Pelican Imaging Corporation | Capturing and processing of images including occlusions captured by heterogeneous camera arrays |
US10306120B2 (en) | 2009-11-20 | 2019-05-28 | Fotonation Limited | Capturing and processing of images captured by camera arrays incorporating cameras with telephoto and conventional lenses to generate depth maps |
US9936148B2 (en) | 2010-05-12 | 2018-04-03 | Fotonation Cayman Limited | Imager array interfaces |
US10455168B2 (en) | 2010-05-12 | 2019-10-22 | Fotonation Limited | Imager array interfaces |
US8929651B2 (en) | 2010-11-16 | 2015-01-06 | Thomson Licensing | System and method for the repair of anomalies in images |
US11423513B2 (en) | 2010-12-14 | 2022-08-23 | Fotonation Limited | Systems and methods for synthesizing high resolution images using images captured by an array of independently controllable imagers |
US10366472B2 (en) | 2010-12-14 | 2019-07-30 | Fotonation Limited | Systems and methods for synthesizing high resolution images using images captured by an array of independently controllable imagers |
US9361662B2 (en) | 2010-12-14 | 2016-06-07 | Pelican Imaging Corporation | Systems and methods for synthesizing high resolution images using images captured by an array of independently controllable imagers |
US11875475B2 (en) | 2010-12-14 | 2024-01-16 | Adeia Imaging Llc | Systems and methods for synthesizing high resolution images using images captured by an array of independently controllable imagers |
US9047684B2 (en) | 2010-12-14 | 2015-06-02 | Pelican Imaging Corporation | Systems and methods for synthesizing high resolution images using a set of geometrically registered images |
US9041824B2 (en) | 2010-12-14 | 2015-05-26 | Pelican Imaging Corporation | Systems and methods for dynamic refocusing of high resolution images generated using images captured by a plurality of imagers |
US8693775B2 (en) * | 2011-03-29 | 2014-04-08 | Sony Corporation | Image processing apparatus, method, recording medium, and program |
US10218889B2 (en) | 2011-05-11 | 2019-02-26 | Fotonation Limited | Systems and methods for transmitting and receiving array camera image data |
US10742861B2 (en) | 2011-05-11 | 2020-08-11 | Fotonation Limited | Systems and methods for transmitting and receiving array camera image data |
US9866739B2 (en) | 2011-05-11 | 2018-01-09 | Fotonation Cayman Limited | Systems and methods for transmitting and receiving array camera image data |
US9128228B2 (en) | 2011-06-28 | 2015-09-08 | Pelican Imaging Corporation | Optical arrangements for use with an array camera |
US9516222B2 (en) | 2011-06-28 | 2016-12-06 | Kip Peli P1 Lp | Array cameras incorporating monolithic array camera modules with high MTF lens stacks for capture of images used in super-resolution processing |
US9578237B2 (en) | 2011-06-28 | 2017-02-21 | Fotonation Cayman Limited | Array cameras incorporating optics with modulation transfer functions greater than sensor Nyquist frequency for capture of images used in super-resolution processing |
US9179113B2 (en) * | 2011-08-31 | 2015-11-03 | Sony Corporation | Image processing device, and image processing method, and program |
US20140253808A1 (en) * | 2011-08-31 | 2014-09-11 | Sony Corporation | Image processing device, and image processing method, and program |
US10375302B2 (en) | 2011-09-19 | 2019-08-06 | Fotonation Limited | Systems and methods for controlling aliasing in images captured by an array camera for use in super resolution processing using pixel apertures |
US9794476B2 (en) | 2011-09-19 | 2017-10-17 | Fotonation Cayman Limited | Systems and methods for controlling aliasing in images captured by an array camera for use in super resolution processing using pixel apertures |
US10275676B2 (en) | 2011-09-28 | 2019-04-30 | Fotonation Limited | Systems and methods for encoding image files containing depth maps stored as metadata |
US9025894B2 (en) | 2011-09-28 | 2015-05-05 | Pelican Imaging Corporation | Systems and methods for decoding light field image files having depth and confidence maps |
US9129183B2 (en) | 2011-09-28 | 2015-09-08 | Pelican Imaging Corporation | Systems and methods for encoding light field image files |
US11729365B2 (en) | 2011-09-28 | 2023-08-15 | Adela Imaging LLC | Systems and methods for encoding image files containing depth maps stored as metadata |
US9042667B2 (en) | 2011-09-28 | 2015-05-26 | Pelican Imaging Corporation | Systems and methods for decoding light field image files using a depth map |
US9031343B2 (en) | 2011-09-28 | 2015-05-12 | Pelican Imaging Corporation | Systems and methods for encoding light field image files having a depth map |
US9031335B2 (en) | 2011-09-28 | 2015-05-12 | Pelican Imaging Corporation | Systems and methods for encoding light field image files having depth and confidence maps |
US9025895B2 (en) | 2011-09-28 | 2015-05-05 | Pelican Imaging Corporation | Systems and methods for decoding refocusable light field image files |
US20180197035A1 (en) | 2011-09-28 | 2018-07-12 | Fotonation Cayman Limited | Systems and Methods for Encoding Image Files Containing Depth Maps Stored as Metadata |
US10019816B2 (en) | 2011-09-28 | 2018-07-10 | Fotonation Cayman Limited | Systems and methods for decoding image files containing depth maps stored as metadata |
US10984276B2 (en) | 2011-09-28 | 2021-04-20 | Fotonation Limited | Systems and methods for encoding image files containing depth maps stored as metadata |
US9036931B2 (en) | 2011-09-28 | 2015-05-19 | Pelican Imaging Corporation | Systems and methods for decoding structured light field image files |
US9864921B2 (en) | 2011-09-28 | 2018-01-09 | Fotonation Cayman Limited | Systems and methods for encoding image files containing depth maps stored as metadata |
US10430682B2 (en) | 2011-09-28 | 2019-10-01 | Fotonation Limited | Systems and methods for decoding image files containing depth maps stored as metadata |
US9536166B2 (en) | 2011-09-28 | 2017-01-03 | Kip Peli P1 Lp | Systems and methods for decoding image files containing depth maps stored as metadata |
US12052409B2 (en) | 2011-09-28 | 2024-07-30 | Adela Imaging LLC | Systems and methods for encoding image files containing depth maps stored as metadata |
US9811753B2 (en) | 2011-09-28 | 2017-11-07 | Fotonation Cayman Limited | Systems and methods for encoding light field image files |
US10311649B2 (en) | 2012-02-21 | 2019-06-04 | Fotonation Limited | Systems and method for performing depth based image editing |
US9412206B2 (en) | 2012-02-21 | 2016-08-09 | Pelican Imaging Corporation | Systems and methods for the manipulation of captured light field image data |
US9754422B2 (en) | 2012-02-21 | 2017-09-05 | Fotonation Cayman Limited | Systems and method for performing depth based image editing |
US9210392B2 (en) | 2012-05-01 | 2015-12-08 | Pelican Imaging Coporation | Camera modules patterned with pi filter groups |
US9706132B2 (en) | 2012-05-01 | 2017-07-11 | Fotonation Cayman Limited | Camera modules patterned with pi filter groups |
US9113118B2 (en) | 2012-05-31 | 2015-08-18 | Apple Inc. | Green non-uniformity correction |
US9807382B2 (en) | 2012-06-28 | 2017-10-31 | Fotonation Cayman Limited | Systems and methods for detecting defective camera arrays and optic arrays |
US9100635B2 (en) | 2012-06-28 | 2015-08-04 | Pelican Imaging Corporation | Systems and methods for detecting defective camera arrays and optic arrays |
US10334241B2 (en) | 2012-06-28 | 2019-06-25 | Fotonation Limited | Systems and methods for detecting defective camera arrays and optic arrays |
US11022725B2 (en) | 2012-06-30 | 2021-06-01 | Fotonation Limited | Systems and methods for manufacturing camera modules using active alignment of lens stack arrays and sensors |
US10261219B2 (en) | 2012-06-30 | 2019-04-16 | Fotonation Limited | Systems and methods for manufacturing camera modules using active alignment of lens stack arrays and sensors |
US9766380B2 (en) | 2012-06-30 | 2017-09-19 | Fotonation Cayman Limited | Systems and methods for manufacturing camera modules using active alignment of lens stack arrays and sensors |
US9858673B2 (en) | 2012-08-21 | 2018-01-02 | Fotonation Cayman Limited | Systems and methods for estimating depth and visibility from a reference viewpoint for pixels in a set of images captured from different viewpoints |
US12002233B2 (en) | 2012-08-21 | 2024-06-04 | Adeia Imaging Llc | Systems and methods for estimating depth and visibility from a reference viewpoint for pixels in a set of images captured from different viewpoints |
US9123117B2 (en) | 2012-08-21 | 2015-09-01 | Pelican Imaging Corporation | Systems and methods for generating depth maps and corresponding confidence maps indicating depth estimation reliability |
US9123118B2 (en) | 2012-08-21 | 2015-09-01 | Pelican Imaging Corporation | System and methods for measuring depth using an array camera employing a bayer filter |
US9129377B2 (en) | 2012-08-21 | 2015-09-08 | Pelican Imaging Corporation | Systems and methods for measuring depth based upon occlusion patterns in images |
US9240049B2 (en) | 2012-08-21 | 2016-01-19 | Pelican Imaging Corporation | Systems and methods for measuring depth using an array of independently controllable cameras |
US10380752B2 (en) | 2012-08-21 | 2019-08-13 | Fotonation Limited | Systems and methods for estimating depth and visibility from a reference viewpoint for pixels in a set of images captured from different viewpoints |
US9147254B2 (en) | 2012-08-21 | 2015-09-29 | Pelican Imaging Corporation | Systems and methods for measuring depth in the presence of occlusions using a subset of images |
US9235900B2 (en) | 2012-08-21 | 2016-01-12 | Pelican Imaging Corporation | Systems and methods for estimating depth and visibility from a reference viewpoint for pixels in a set of images captured from different viewpoints |
US9813616B2 (en) | 2012-08-23 | 2017-11-07 | Fotonation Cayman Limited | Feature based high resolution motion estimation from low resolution images captured using an array source |
US10462362B2 (en) | 2012-08-23 | 2019-10-29 | Fotonation Limited | Feature based high resolution motion estimation from low resolution images captured using an array source |
US9214013B2 (en) | 2012-09-14 | 2015-12-15 | Pelican Imaging Corporation | Systems and methods for correcting user identified artifacts in light field images |
US10390005B2 (en) | 2012-09-28 | 2019-08-20 | Fotonation Limited | Generating images from light fields utilizing virtual viewpoints |
US9143711B2 (en) | 2012-11-13 | 2015-09-22 | Pelican Imaging Corporation | Systems and methods for array camera focal plane control |
US9749568B2 (en) | 2012-11-13 | 2017-08-29 | Fotonation Cayman Limited | Systems and methods for array camera focal plane control |
US9462164B2 (en) | 2013-02-21 | 2016-10-04 | Pelican Imaging Corporation | Systems and methods for generating compressed light field representation data using captured light fields, array geometry, and parallax information |
US10009538B2 (en) | 2013-02-21 | 2018-06-26 | Fotonation Cayman Limited | Systems and methods for generating compressed light field representation data using captured light fields, array geometry, and parallax information |
US9774831B2 (en) | 2013-02-24 | 2017-09-26 | Fotonation Cayman Limited | Thin form factor computational array cameras and modular array cameras |
US9253380B2 (en) | 2013-02-24 | 2016-02-02 | Pelican Imaging Corporation | Thin form factor computational array cameras and modular array cameras |
US9374512B2 (en) | 2013-02-24 | 2016-06-21 | Pelican Imaging Corporation | Thin form factor computational array cameras and modular array cameras |
US9743051B2 (en) | 2013-02-24 | 2017-08-22 | Fotonation Cayman Limited | Thin form factor computational array cameras and modular array cameras |
US9917998B2 (en) | 2013-03-08 | 2018-03-13 | Fotonation Cayman Limited | Systems and methods for measuring scene information while capturing images using array cameras |
US9774789B2 (en) | 2013-03-08 | 2017-09-26 | Fotonation Cayman Limited | Systems and methods for high dynamic range imaging using array cameras |
US9986224B2 (en) | 2013-03-10 | 2018-05-29 | Fotonation Cayman Limited | System and methods for calibration of an array camera |
US10225543B2 (en) | 2013-03-10 | 2019-03-05 | Fotonation Limited | System and methods for calibration of an array camera |
US11570423B2 (en) | 2013-03-10 | 2023-01-31 | Adeia Imaging Llc | System and methods for calibration of an array camera |
US11272161B2 (en) | 2013-03-10 | 2022-03-08 | Fotonation Limited | System and methods for calibration of an array camera |
US10958892B2 (en) | 2013-03-10 | 2021-03-23 | Fotonation Limited | System and methods for calibration of an array camera |
US11985293B2 (en) | 2013-03-10 | 2024-05-14 | Adeia Imaging Llc | System and methods for calibration of an array camera |
US9521416B1 (en) | 2013-03-11 | 2016-12-13 | Kip Peli P1 Lp | Systems and methods for image data compression |
US9519972B2 (en) | 2013-03-13 | 2016-12-13 | Kip Peli P1 Lp | Systems and methods for synthesizing images from image data captured by an array camera using restricted depth of field depth maps in which depth estimation precision varies |
US10127682B2 (en) | 2013-03-13 | 2018-11-13 | Fotonation Limited | System and methods for calibration of an array camera |
US9800856B2 (en) | 2013-03-13 | 2017-10-24 | Fotonation Cayman Limited | Systems and methods for synthesizing images from image data captured by an array camera using restricted depth of field depth maps in which depth estimation precision varies |
US9106784B2 (en) | 2013-03-13 | 2015-08-11 | Pelican Imaging Corporation | Systems and methods for controlling aliasing in images captured by an array camera for use in super-resolution processing |
US9888194B2 (en) | 2013-03-13 | 2018-02-06 | Fotonation Cayman Limited | Array camera architecture implementing quantum film image sensors |
US9733486B2 (en) | 2013-03-13 | 2017-08-15 | Fotonation Cayman Limited | Systems and methods for controlling aliasing in images captured by an array camera for use in super-resolution processing |
US9741118B2 (en) | 2013-03-13 | 2017-08-22 | Fotonation Cayman Limited | System and methods for calibration of an array camera |
US9578259B2 (en) | 2013-03-14 | 2017-02-21 | Fotonation Cayman Limited | Systems and methods for reducing motion blur in images or video in ultra low light with array cameras |
US9787911B2 (en) | 2013-03-14 | 2017-10-10 | Fotonation Cayman Limited | Systems and methods for photometric normalization in array cameras |
US10547772B2 (en) | 2013-03-14 | 2020-01-28 | Fotonation Limited | Systems and methods for reducing motion blur in images or video in ultra low light with array cameras |
US9100586B2 (en) | 2013-03-14 | 2015-08-04 | Pelican Imaging Corporation | Systems and methods for photometric normalization in array cameras |
US10412314B2 (en) | 2013-03-14 | 2019-09-10 | Fotonation Limited | Systems and methods for photometric normalization in array cameras |
US10091405B2 (en) | 2013-03-14 | 2018-10-02 | Fotonation Cayman Limited | Systems and methods for reducing motion blur in images or video in ultra low light with array cameras |
US10182216B2 (en) | 2013-03-15 | 2019-01-15 | Fotonation Limited | Extended color processing on pelican array cameras |
US9800859B2 (en) | 2013-03-15 | 2017-10-24 | Fotonation Cayman Limited | Systems and methods for estimating depth using stereo array cameras |
US9955070B2 (en) | 2013-03-15 | 2018-04-24 | Fotonation Cayman Limited | Systems and methods for synthesizing high resolution images using image deconvolution based on motion and depth information |
US9497429B2 (en) | 2013-03-15 | 2016-11-15 | Pelican Imaging Corporation | Extended color processing on pelican array cameras |
US10455218B2 (en) | 2013-03-15 | 2019-10-22 | Fotonation Limited | Systems and methods for estimating depth using stereo array cameras |
US9633442B2 (en) | 2013-03-15 | 2017-04-25 | Fotonation Cayman Limited | Array cameras including an array camera module augmented with a separate camera |
US9497370B2 (en) | 2013-03-15 | 2016-11-15 | Pelican Imaging Corporation | Array camera architecture implementing quantum dot color filters |
US10542208B2 (en) | 2013-03-15 | 2020-01-21 | Fotonation Limited | Systems and methods for synthesizing high resolution images using image deconvolution based on motion and depth information |
US10674138B2 (en) | 2013-03-15 | 2020-06-02 | Fotonation Limited | Autofocus system for a conventional camera that uses depth information from an array camera |
US9438888B2 (en) | 2013-03-15 | 2016-09-06 | Pelican Imaging Corporation | Systems and methods for stereo imaging with camera arrays |
US10122993B2 (en) | 2013-03-15 | 2018-11-06 | Fotonation Limited | Autofocus system for a conventional camera that uses depth information from an array camera |
US10638099B2 (en) | 2013-03-15 | 2020-04-28 | Fotonation Limited | Extended color processing on pelican array cameras |
US9898856B2 (en) | 2013-09-27 | 2018-02-20 | Fotonation Cayman Limited | Systems and methods for depth-assisted perspective distortion correction |
US8995766B1 (en) * | 2013-09-27 | 2015-03-31 | Himax Imaging Limited | Image processing method and image processing device |
US10540806B2 (en) | 2013-09-27 | 2020-01-21 | Fotonation Limited | Systems and methods for depth-assisted perspective distortion correction |
US9264592B2 (en) | 2013-11-07 | 2016-02-16 | Pelican Imaging Corporation | Array camera modules incorporating independently aligned lens stacks |
US9185276B2 (en) | 2013-11-07 | 2015-11-10 | Pelican Imaging Corporation | Methods of manufacturing array camera modules incorporating independently aligned lens stacks |
US9924092B2 (en) | 2013-11-07 | 2018-03-20 | Fotonation Cayman Limited | Array cameras incorporating independently aligned lens stacks |
US10767981B2 (en) | 2013-11-18 | 2020-09-08 | Fotonation Limited | Systems and methods for estimating depth from projected texture using camera arrays |
US10119808B2 (en) | 2013-11-18 | 2018-11-06 | Fotonation Limited | Systems and methods for estimating depth from projected texture using camera arrays |
US11486698B2 (en) | 2013-11-18 | 2022-11-01 | Fotonation Limited | Systems and methods for estimating depth from projected texture using camera arrays |
US10708492B2 (en) | 2013-11-26 | 2020-07-07 | Fotonation Limited | Array camera configurations incorporating constituent array cameras and constituent cameras |
US9426361B2 (en) | 2013-11-26 | 2016-08-23 | Pelican Imaging Corporation | Array camera configurations incorporating multiple constituent array cameras |
US9813617B2 (en) | 2013-11-26 | 2017-11-07 | Fotonation Cayman Limited | Array camera configurations incorporating constituent array cameras and constituent cameras |
US10574905B2 (en) | 2014-03-07 | 2020-02-25 | Fotonation Limited | System and methods for depth regularization and semiautomatic interactive matting using RGB-D images |
US10089740B2 (en) | 2014-03-07 | 2018-10-02 | Fotonation Limited | System and methods for depth regularization and semiautomatic interactive matting using RGB-D images |
US9247117B2 (en) | 2014-04-07 | 2016-01-26 | Pelican Imaging Corporation | Systems and methods for correcting for warpage of a sensor array in an array camera module by introducing warpage into a focal plane of a lens stack array |
US9521319B2 (en) | 2014-06-18 | 2016-12-13 | Pelican Imaging Corporation | Array cameras and array camera modules including spectral filters disposed outside of a constituent image sensor |
US11546576B2 (en) | 2014-09-29 | 2023-01-03 | Adeia Imaging Llc | Systems and methods for dynamic calibration of array cameras |
US10250871B2 (en) | 2014-09-29 | 2019-04-02 | Fotonation Limited | Systems and methods for dynamic calibration of array cameras |
US9942474B2 (en) | 2015-04-17 | 2018-04-10 | Fotonation Cayman Limited | Systems and methods for performing high speed video capture and depth estimation using array cameras |
US10482618B2 (en) | 2017-08-21 | 2019-11-19 | Fotonation Limited | Systems and methods for hybrid depth regularization |
US11562498B2 (en) | 2017-08-21 | 2023-01-24 | Adela Imaging LLC | Systems and methods for hybrid depth regularization |
US11983893B2 (en) | 2017-08-21 | 2024-05-14 | Adeia Imaging Llc | Systems and methods for hybrid depth regularization |
US10818026B2 (en) | 2017-08-21 | 2020-10-27 | Fotonation Limited | Systems and methods for hybrid depth regularization |
TWI638339B (en) * | 2017-11-14 | 2018-10-11 | 瑞昱半導體股份有限公司 | False color removal method |
US11699273B2 (en) | 2019-09-17 | 2023-07-11 | Intrinsic Innovation Llc | Systems and methods for surface modeling using polarization cues |
US11270110B2 (en) | 2019-09-17 | 2022-03-08 | Boston Polarimetrics, Inc. | Systems and methods for surface modeling using polarization cues |
US11525906B2 (en) | 2019-10-07 | 2022-12-13 | Intrinsic Innovation Llc | Systems and methods for augmentation of sensor systems and imaging systems with polarization |
US11982775B2 (en) | 2019-10-07 | 2024-05-14 | Intrinsic Innovation Llc | Systems and methods for augmentation of sensor systems and imaging systems with polarization |
US12099148B2 (en) | 2019-10-07 | 2024-09-24 | Intrinsic Innovation Llc | Systems and methods for surface normals sensing with polarization |
US11302012B2 (en) | 2019-11-30 | 2022-04-12 | Boston Polarimetrics, Inc. | Systems and methods for transparent object segmentation using polarization cues |
US11842495B2 (en) | 2019-11-30 | 2023-12-12 | Intrinsic Innovation Llc | Systems and methods for transparent object segmentation using polarization cues |
US11580667B2 (en) | 2020-01-29 | 2023-02-14 | Intrinsic Innovation Llc | Systems and methods for characterizing object pose detection and measurement systems |
US11797863B2 (en) | 2020-01-30 | 2023-10-24 | Intrinsic Innovation Llc | Systems and methods for synthesizing data for training statistical models on different imaging modalities including polarized images |
US11953700B2 (en) | 2020-05-27 | 2024-04-09 | Intrinsic Innovation Llc | Multi-aperture polarization optical systems using beam splitters |
US12020455B2 (en) | 2021-03-10 | 2024-06-25 | Intrinsic Innovation Llc | Systems and methods for high dynamic range image reconstruction |
US12069227B2 (en) | 2021-03-10 | 2024-08-20 | Intrinsic Innovation Llc | Multi-modal and multi-spectral stereo camera arrays |
US11954886B2 (en) | 2021-04-15 | 2024-04-09 | Intrinsic Innovation Llc | Systems and methods for six-degree of freedom pose estimation of deformable objects |
US11683594B2 (en) | 2021-04-15 | 2023-06-20 | Intrinsic Innovation Llc | Systems and methods for camera exposure control |
US11290658B1 (en) | 2021-04-15 | 2022-03-29 | Boston Polarimetrics, Inc. | Systems and methods for camera exposure control |
US12067746B2 (en) | 2021-05-07 | 2024-08-20 | Intrinsic Innovation Llc | Systems and methods for using computer vision to pick up small objects |
US11689813B2 (en) | 2021-07-01 | 2023-06-27 | Intrinsic Innovation Llc | Systems and methods for high dynamic range imaging using crossed polarizers |
Also Published As
Publication number | Publication date |
---|---|
US7477781B1 (en) | 2009-01-13 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US20090263017A1 (en) | Method for reconstruction of pixel color values | |
US6806902B1 (en) | System and method for correcting bad pixel data in a digital camera | |
US6625325B2 (en) | Noise cleaning and interpolating sparsely populated color digital image using a variable noise cleaning kernel | |
EP0304643B1 (en) | Improved method and apparatus for reconstructing missing color | |
US7486844B2 (en) | Color interpolation apparatus and color interpolation method utilizing edge indicators adjusted by stochastic adjustment factors to reconstruct missing colors for image pixels | |
EP1641283B1 (en) | Image processing method, image processing program, image processor | |
US7728882B2 (en) | Green reconstruction for image sensors | |
US8270774B2 (en) | Image processing device for performing interpolation | |
US20150049223A1 (en) | Image processing apparatus, image processing method, and program | |
WO1999046730A1 (en) | Edge-dependent interpolation method for color reconstruction in image processing devices | |
CN102907103A (en) | Image processing device, image processing method and program | |
Chen et al. | Effective demosaicking algorithm based on edge property for color filter arrays | |
CN113454687A (en) | Image processing method, apparatus and system, computer readable storage medium | |
US8233733B2 (en) | Image processing device | |
US20100134661A1 (en) | Image processing apparatus, image processing method and program | |
JPH0646817B2 (en) | Video signal processor | |
US8363135B2 (en) | Method and device for reconstructing a color image | |
US7573515B2 (en) | Method and apparatus for processing a sensor signal having a plurality of pixels from an image sensor, computer program product, computing system and camera | |
CN104427321B (en) | Image processing apparatus and its control method | |
Wang et al. | Demosaicing with improved edge direction detection | |
JP4178919B2 (en) | Noise removal method, imaging apparatus, and noise removal program | |
JP3115123B2 (en) | Image signal processing device | |
EP1522047B1 (en) | Method and apparatus for signal processing, computer program product, computing system and camera | |
JP4666786B2 (en) | Image interpolation device | |
JP4495355B2 (en) | Image interpolation device |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |