[go: up one dir, main page]

WO2006112814A1 - Edge-sensitive denoising and color interpolation of digital images - Google Patents

Edge-sensitive denoising and color interpolation of digital images Download PDF

Info

Publication number
WO2006112814A1
WO2006112814A1 PCT/US2005/012363 US2005012363W WO2006112814A1 WO 2006112814 A1 WO2006112814 A1 WO 2006112814A1 US 2005012363 W US2005012363 W US 2005012363W WO 2006112814 A1 WO2006112814 A1 WO 2006112814A1
Authority
WO
WIPO (PCT)
Prior art keywords
color
pixel
information
denoising
interpolation
Prior art date
Application number
PCT/US2005/012363
Other languages
French (fr)
Inventor
Renato Keshet
Casey L Miller
Benjamin D. Kimbell
Original Assignee
Hewlett-Packard Development Company L.P.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Hewlett-Packard Development Company L.P. filed Critical Hewlett-Packard Development Company L.P.
Priority to PCT/US2005/012363 priority Critical patent/WO2006112814A1/en
Publication of WO2006112814A1 publication Critical patent/WO2006112814A1/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformations in the plane of the image
    • G06T3/40Scaling of whole images or parts thereof, e.g. expanding or contracting
    • G06T3/403Edge-driven scaling; Edge-based scaling
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/70Denoising; Smoothing

Definitions

  • a typical digital camera includes an array of photosensors, with each photosensor sensitive to only a single color of light. For example, each photosensor is sensitive to one of red, green and blue light.
  • each photosensor is sensitive to one of red, green and blue light.
  • an image is focused on the photosensor array, and each photosensor measures or "samples" a single color of the image. If red-sensitive, green- sensitive and blue-sensitive photosensors can be located at each pixel, the photosensor array can acquire an image having "full color" at each pixel.
  • the photosensor arrays of certain digital cameras have only a single photosensor at each pixel location. These cameras produce digital images that do not have full color information at each pixel. Since each photosensor is sensitive to only a single color, the photosensor array produces a digital image having only a single color sample at each pixel. Consider a digital camera that produces a digital image having one of red, green and blue sampled information at each pixel. Information about the other two colors at each pixel is missing. This undersampled digital image is referred to as a "mosaic" image.
  • a demosaicing algorithm may be used to transform an undersampled digital image into a digital image having full color information at each pixel value.
  • a typical demosaicing algorithm interpolates the missing pixel information from the sampled pixel values in the mosaic image.
  • a simple demosaicing algorithm such as bilinear interpolation fills in a missing pixel value by taking an average of sampled values from neighboring pixels. However, some of those neighbors might lie on opposite sides of an edge. Some of the neighbors lying on one side of the edge might belong to one object, while the other neighbors lying on the other side might belong to a different object. Consequently, the interpolated pixel information might not describe either object. Since traditional bilinear interpolation does not account for edges, color information at edges in demosaiced images can be distorted. [0005] More complex demosaicing algorithms try to account for edges. Still, even the more complex demosaicing algorithm can introduce artifacts into demosaiced images. Zippering and fringing are typical artifacts at edges in the demosaiced image. These artifacts can degrade image quality.
  • a digital camera can perform the demosaicing.
  • On-board demosaicing allows a digital camera to display and upload images having full color at each pixel.
  • memory and processing power of a typical digital camera are limited. The limited memory and processing power can constrain the complexity of the demosaicing algorithm and hamper the ability to account for edges and reduce artifacts at the edges during demosaicing.
  • a demosaicing algorithm that is simple and fast, and that reduces edge blurring and the visibility of certain artifacts at edges, is desirable. Such a demosaicing algorithm is especially desirable for digital cameras.
  • demosaiced images there is also a need to reduce noise in the demosaiced images. If the mosaic images are noisy, the demosaiced images produced by the demosaicing algorithm could be noisy too. Further, if post-processing such as image sharpening is performed on the demosaiced images, the post-processing could significantly increase structured or unstructured noise. Such noise might not be perceptible in unsharpened images, but it might become very visible after sharpening. Still further, the demosaicing could create cross-hatching artifacts on critically sampled regions (that is, regions containing components with spatial frequency close to the sampling frequency) of the sharpened images.
  • a method of processing a mosaic digital image includes simultaneously performing edge- sensitive denoising and color interpolation on a first color plane of the digital image.
  • the interpolation fills in missing pixel information in the first color plane.
  • Figure 1 is an illustration of a photosensor arrangement for a typical digital camera, and corresponding pixels of a mosaic image.
  • Figure 2 is an illustration of a method of performing image processing in accordance with an embodiment of the present invention.
  • Figure 3 is an illustration of a method of performing edge- sensitive denoising of sampled values in accordance with an embodiment of the present invention.
  • Figure 4 is an illustration of a method of performing denoising and color interpolation in accordance with an embodiment of the present invention.
  • Figure 5 is an illustration of a pixel to be processed and directions of neighboring pixels in a mosaic image.
  • Figure 6 is an illustration of an apparatus for performing image processing in accordance with an embodiment of the present invention.
  • Figure 7 is an illustration of an apparatus for performing image processing in accordance with an embodiment of the present invention.
  • the present invention is embodied in an algorithm for processing of a mosaic image.
  • Each pixel of the mosaic image has less than full information.
  • the algorithm includes performing edge-sensitive denoising and color interpolation simultaneously. Resulting is an output image having full color information at each pixel.
  • the algorithm is simple and fast, and it reduces edge blurring and the visibility of certain artifacts at edges.
  • the algorithm also reduces structured noise on smooth regions of sharpened outputs, as well as cross-hatching artifacts.
  • FIG. 1 illustrates a photosensor array 1 10 having photosensors 112 arranged in a Bayer CFA.
  • the photosensors 112 are arranged in 2x2 cells 114.
  • Each cell 114 consists of two photosensors sensitive to green (G) light only, one photosensor sensitive to red (R) light only, and one photosensor sensitive to blue (B) light only.
  • the cells 1 14 are repeated (tiled) across the photosensor array 110.
  • Figure 1 also illustrates a mosaic digital image 150.
  • Each block 154 of the mosaic image 150 corresponds to a cell 1 14 of the photosensor array 110.
  • Each pixel 152 of the mosaic image 150 is described by an n-bit word, and each n-bit word provides one of red, green and blue color sampled information.
  • green information is sampled at two pixels
  • red information is sampled at one pixel
  • blue information is sampled at one pixel.
  • FIG. 2 illustrates a method of processing a mosaic image into an output image having full color information at each pixel.
  • Each pixel of the output image may be represented by a digital word including n-bits for red information, n-bits for green information, and n-bits for blue information.
  • the sampled values may be pre-processed (210).
  • the preprocessing is application-specific. In a digital camera for example, preprocessing between image acquisition and demosaicing may include flare reduction, data linearization, and color balance.
  • the sampled values of the mosaic image are copied to their corresponding positions in the output image (220).
  • a mosaic image with a Bayer pattern has pixel inputs m(i,j) , where / and j are indices to the rows and columns of the mosaic image.
  • the sampled values of the mosaic image may be copied to the output image as follows:
  • I R (i,j) m(i,j) for all red samples (/ and j even)
  • I 0 (JJ) m(ij) for all green samples (/ + y odd)
  • I R ⁇ i,j) represents red information at pixel (ij) of the output image
  • I G (i,j) represents green information at pixel (ij) of the output image
  • I B ⁇ i,j) represents blue information at pixel (ij) of the output image.
  • a smooth estimation of the first color plane is computed (230).
  • the smooth estimation may be a bilinear interpolation of the samples in the first color plane.
  • the G1 samples have the same spatial characteristic as the red and blue samples, and the G2 samples have the same spatial characteristic as the red and blue samples, and thus the bilinear interpolation of each of the G1 and G2 samples yields a smooth estimation. Therefore, the sum of the bilinear interpolation of the G1 samples with the sum of the bilinear interpolation of the G2 samples also yields a smooth estimation.
  • Such a smooth estimation may be computed as
  • This smooth estimation does not contain artifacts that would otherwise occur in bilinear interpolation of the green plane. Moreover, the smooth estimation is blurrier than a bilinear interpolation (which also can be advantageous).
  • the red or blue plane may be used instead of the green plane.
  • the smooth estimations of the red or blue plane may be a bilinear interpolation of the samples. Smooth estimations of the red or blue plane may be computed as
  • the green color plane is preferred because it corresponds to the largest component of the luminance, and because it has twice the amount of information in a Bayer cell compared to the other color planes (the green plane yields a stronger response by the human visual system than either the red plane or the blue plane).
  • the sampled values of the first color plane may be denoised (240). Denoising may follow the general approach illustrated in Figure 3: an error between a sampled value and the value of a corresponding pixel in the smooth estimation are computed (310), the error is modified according to local noise level (320), and the modified error is used to modify the corresponding pixel in the smooth estimation (330).
  • the value ⁇ ⁇ (x) should be close to zero if
  • 1 ⁇ J' should be close to m(i,j) if the absolute difference of m GvJ) -° G (hj ) i - S
  • Such filtering is edge-sensitive because the "error" grows significantly when one is close to an edge. In a smooth region, far from an edge, the region is smoothed. In a region near an edge, the region is not smoothed.
  • the denoising preserves strong features such as edges (i.e., it doesn't blur edges), yet removes small features such as noise.
  • Missing information in the first color plane of the output image is added to the output image by simultaneously performing edge-sensitive denoising and color interpolation on the first color (250).
  • the edge-sensitive denoising and color interpolation of a given pixel includes determining a likelihood of whether the given pixel belongs to the same region as each of at least two neighboring denoised samples of the first color (410), the neighboring samples in different directions relative to the given pixel; and using the likelihoods and the denoised neighboring samples to interpolate the first color at the given pixel (420). A sample having a higher likelihood of belonging to the same region as the given pixel is given a higher weight in the interpolation.
  • Different approaches for performing the simultaneous denoising and color interpolation will be described below. All of the approaches are sensitive to edges in the smooth estimation.
  • Missing information in the other color planes is interpolated and added to the output image (260).
  • Blurring kernels may be used to populate the red and blue color planes as follows:
  • the blurring kernel h may be
  • This blurring kernel can reduce color artifacts.
  • a larger blurring kernel h e.g., a 5x5 kernel
  • Another method of interpolating the missing information in each of the red and blue color planes includes performing the simultaneous edge- sensitive denoising and color interpolation. Still other methods of interpolating the missing information in each of the red and blue color planes include bilinear interpolation, affine interpolation, and a canonical correlation method. These other methods are described in detail in assignee's U.S. Serial No. 10/743,623 filed December 22, 2003 and incorporated herein by reference.
  • each pixel /(/,_/) in the output image has red information I R (i,j) , green information I G (i,j) , and blue information I B (i, j) .
  • Post processing may be performed on the output (demosaiced) image (270).
  • the post-processing is application-specific. In a digital camera for example, the post-processing may include color transformation, tone reproduction, sharpening, compression, and adaptive lighting.
  • Figure 5 shows raw samples of a mosaic image 150.
  • the given pixel i.e., the pixel being processed
  • the given pixel in Figure 5 has a sampled red value, but no sampled values for green or blue.
  • the first approach involves determining a likelihood of whether the given pixel belongs to the same region as each of at least two other pixels having smooth estimates of the first color.
  • the other pixels are in different directions relative to the given pixel.
  • the samples and the likelihoods and the smooth estimates of the other pixels are used to interpolate the green information at the given pixel. This first approach may be performed according to equation (1).
  • Io (i, J) oc (i, j) + ⁇ v , . ⁇ ⁇ (1 )
  • the term N 4 represents a neighborhood consisting of four pixels.
  • the index (a,b) indicates the direction and number of pixels relative to the given pixel.
  • the neighborhood N 4 may consist of ⁇ (-1 ,0), (1 ,0), (0,- 1) and (0,1) ⁇ .
  • the green samples are due west (-1 ,0) , due east (1 ,0), due south (0,-1) and due north (0,1) of the given pixel. Equation (1 ) holds only for those pixels (i,j) without green samples in the mosaic image 150.
  • Weight term w a ,b may be computed as w ' tone*(* > j) + ⁇ o G (i + 2aJ + 2b) - o ( ; (i + aJ + ⁇ .
  • the first absolute difference o G (i + a,j + b) - o G (i,j) in the weight term w a ,b represents the intensity of the difference between the smooth green estimate of the given pixel and that of the nearest green pixel in the direction determined by (a,b); and the second absolute difference in the weighting term o G (i + 2a,j + 2b) - oa(i + a,j + b) ⁇ represents the intensity of the difference between the smooth green estimate of the nearest green pixel in the direction determined by (a,b) and that of the next pixel in the same direction.
  • the weight terms w a , b are based on smooth estimates, not raw samples. Use of the smooth estimates instead of raw samples in the second absolute difference allows differences between adjacent neighboring pixels to be considered. If raw samples are used, in contrast, the differences are between neighbors that are two pixels apart.
  • the function p() is referred to as a similarity function.
  • the similarity function is used to determine the likelihood of a given pixel belonging to the same region as a neighboring pixel as a function of the sum of absolute differences corresponding to that direction.
  • a large absolute difference in sampled values over a short distance that is, an abrupt transition
  • a large sum of absolute differences in a given direction implies that the given pixel lies on one side of an edge (and belongs to one object), and the neighbor at that direction lies on the other side of the edge (and belongs to a different object). Therefore, a low likelihood exists that the green values of the given pixel and neighbor are similar.
  • the quantity [m(i,j) - m(i + 2a,j + 2b)] represents the difference between the red sample on the given pixel and the nearest red samples in the direction determined by (a,b) relative to the given pixel.
  • the parameter A is a tunable parameter that corrects the interpolated value for camera-specific items, such as image resolution, camera optics, and scene distance (and, therefore, is used in ⁇ Oii (/,y) but not the weight terms w a,b ).
  • the parameter ⁇ may have a value between 0.5 and 1.
  • the parameter ⁇ may have a value of 0.5.
  • the parameter ⁇ may have a value of less than 0.5.
  • the high frequency correction term ⁇ a , b is not used, whereby equation (1) is reduced to the following:
  • a third approach involves taking the psi-sigma function ( ⁇ ⁇ ) and the subtraction by o G out of the loop of equation (1). This leads to the equation (3).
  • This third approach (referred to as post-selective filtering) is faster than the first approach, since the psi-sigma function is applied only once for each input pixel (i,j) .
  • the first approach (referred to as pre-selective filtering) applies the psi-sigma function to each pixel (a,b) in a local neighborhood of each input pixel (i,j).
  • the pre-selective filtering tends to produce less noise around edges in the output image.
  • I G (iJ) o G (iJ) + ⁇ c -o G (i,j) (4)
  • a lookup table may be used instead of applying the psi-sigma function.
  • the lookup table stores samples (that is, pre-computed values) of the psi-sigma function.
  • the method is not limited to the N, S, E and W directions, or the four closest neighbors. Larger neighborhoods may be used, and pixels of different distances may be used. Weights may be reduced as distance increases from the given pixel.
  • a digital camera 610 includes a photosensor array 612 such as a CCD.
  • the photosensors of the array 612 may be arranged in a Bayer CFA.
  • the array 612 produces mosaic digital images.
  • the digital camera 610 also includes a digital signal processor 614 and memory 616 (e.g., EPROM).
  • a program 618 encoded in the memory 616 instructs the digital signal processor 614 to transform mosaic images into digital images having full color information at each pixel in accordance with an embodiment of the present invention.
  • the program 618 may include a lookup table for different values of the similarity function. Weights are computed on the fly, and applied to the pixels using an algorithm.
  • ASIC application-specific integrated circuit
  • the ASIC includes a processor for transforming mosaic images into digital images having full color information at each pixel in accordance with an embodiment of the present invention.
  • the processor may include a state machine, or it may include an embedded processor and memory, or it may have some other design.
  • a digital imaging system 710 includes a general purpose processor 712 and memory 714 for instructing the processor 712 to process mosaic images in accordance with an embodiment of the present invention.
  • the system 710 may be, for example, a personal computer.
  • Mosaic images may be supplied to the system 710 directly by a capture device (e.g., a scanner, a digital camera) or indirectly (via the Internet, accessed from remote or local storage, etc.).
  • a machine for generating the program may be a personal computer such as the one shown in Figure 7.
  • Source code is generated, compiled into executable code, and distributed.
  • the executable code may be distributed in variety of ways.
  • the executable code may be distributed to a digital camera by burning the code into a memory 616 (e.g., the EPROM) of the digital camera 610.
  • the executable code can be distributed to another computer via a network transmission, removable storage medium 718 (e.g., CD), etc.

Landscapes

  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Color Television Image Signal Generators (AREA)
  • Image Processing (AREA)

Abstract

A method of processing a mosaic digital image includes simultaneously performing edge-sensitive denoising and color interpolation on a first color plane of the digital image (250). The interpolation fills in missing pixel information in the first color plane.

Description

EDGE-SENSITIVE DENOISING AND COLOR INTERPOLATION OF
DIGITAL IMAGES
BACKGROUND
[0001] A typical digital camera includes an array of photosensors, with each photosensor sensitive to only a single color of light. For example, each photosensor is sensitive to one of red, green and blue light. During image acquisition, an image is focused on the photosensor array, and each photosensor measures or "samples" a single color of the image. If red-sensitive, green- sensitive and blue-sensitive photosensors can be located at each pixel, the photosensor array can acquire an image having "full color" at each pixel.
[0002] The photosensor arrays of certain digital cameras have only a single photosensor at each pixel location. These cameras produce digital images that do not have full color information at each pixel. Since each photosensor is sensitive to only a single color, the photosensor array produces a digital image having only a single color sample at each pixel. Consider a digital camera that produces a digital image having one of red, green and blue sampled information at each pixel. Information about the other two colors at each pixel is missing. This undersampled digital image is referred to as a "mosaic" image.
[0003] A demosaicing algorithm may be used to transform an undersampled digital image into a digital image having full color information at each pixel value. A typical demosaicing algorithm interpolates the missing pixel information from the sampled pixel values in the mosaic image.
[0004] Edges and other abrupt photometric transitions present a particular problem to demosaicing. A simple demosaicing algorithm such as bilinear interpolation fills in a missing pixel value by taking an average of sampled values from neighboring pixels. However, some of those neighbors might lie on opposite sides of an edge. Some of the neighbors lying on one side of the edge might belong to one object, while the other neighbors lying on the other side might belong to a different object. Consequently, the interpolated pixel information might not describe either object. Since traditional bilinear interpolation does not account for edges, color information at edges in demosaiced images can be distorted. [0005] More complex demosaicing algorithms try to account for edges. Still, even the more complex demosaicing algorithm can introduce artifacts into demosaiced images. Zippering and fringing are typical artifacts at edges in the demosaiced image. These artifacts can degrade image quality.
[0006] A digital camera can perform the demosaicing. On-board demosaicing allows a digital camera to display and upload images having full color at each pixel. However, memory and processing power of a typical digital camera are limited. The limited memory and processing power can constrain the complexity of the demosaicing algorithm and hamper the ability to account for edges and reduce artifacts at the edges during demosaicing.
[0007] A demosaicing algorithm that is simple and fast, and that reduces edge blurring and the visibility of certain artifacts at edges, is desirable. Such a demosaicing algorithm is especially desirable for digital cameras.
[0008] There is also a need to reduce noise in the demosaiced images. If the mosaic images are noisy, the demosaiced images produced by the demosaicing algorithm could be noisy too. Further, if post-processing such as image sharpening is performed on the demosaiced images, the post-processing could significantly increase structured or unstructured noise. Such noise might not be perceptible in unsharpened images, but it might become very visible after sharpening. Still further, the demosaicing could create cross-hatching artifacts on critically sampled regions (that is, regions containing components with spatial frequency close to the sampling frequency) of the sharpened images.
[0009] A demosaicing algorithm with embedded noise reduction capability is desirable.
SUMMARY
[0010] According to one aspect of the present invention, a method of processing a mosaic digital image includes simultaneously performing edge- sensitive denoising and color interpolation on a first color plane of the digital image. The interpolation fills in missing pixel information in the first color plane.
[0011] Other aspects and advantages of the present invention will become apparent from the following detailed description, taken in conjunction with the accompanying drawings, illustrating by way of example the principles of the present invention.
BRIEF DESCRIPTION OF THE DRAWINGS
[0012] Figure 1 is an illustration of a photosensor arrangement for a typical digital camera, and corresponding pixels of a mosaic image.
[0013] Figure 2 is an illustration of a method of performing image processing in accordance with an embodiment of the present invention.
[0014] Figure 3 is an illustration of a method of performing edge- sensitive denoising of sampled values in accordance with an embodiment of the present invention.
[0015] Figure 4 is an illustration of a method of performing denoising and color interpolation in accordance with an embodiment of the present invention.
[0016] Figure 5 is an illustration of a pixel to be processed and directions of neighboring pixels in a mosaic image.
[0017] Figure 6 is an illustration of an apparatus for performing image processing in accordance with an embodiment of the present invention.
[0018] Figure 7 is an illustration of an apparatus for performing image processing in accordance with an embodiment of the present invention.
DETAILED DESCRPTION
[0019] As shown in the drawings for purposes of illustration, the present invention is embodied in an algorithm for processing of a mosaic image. Each pixel of the mosaic image has less than full information. The algorithm includes performing edge-sensitive denoising and color interpolation simultaneously. Resulting is an output image having full color information at each pixel. The algorithm is simple and fast, and it reduces edge blurring and the visibility of certain artifacts at edges. The algorithm also reduces structured noise on smooth regions of sharpened outputs, as well as cross-hatching artifacts.
[0020] Also for the purposes of illustration, the algorithm will be described in connection with a photosensor arrangement known as a Bayer color filter array (CFA). Such a CFA is used in certain digital cameras. [0021] Reference is made to Figure 1 , which illustrates a photosensor array 1 10 having photosensors 112 arranged in a Bayer CFA. The photosensors 112 are arranged in 2x2 cells 114. Each cell 114 consists of two photosensors sensitive to green (G) light only, one photosensor sensitive to red (R) light only, and one photosensor sensitive to blue (B) light only. The cells 1 14 are repeated (tiled) across the photosensor array 110.
[0022] Figure 1 also illustrates a mosaic digital image 150. Each block 154 of the mosaic image 150 corresponds to a cell 1 14 of the photosensor array 110. Each pixel 152 of the mosaic image 150 is described by an n-bit word, and each n-bit word provides one of red, green and blue color sampled information. In each 2x2 block 154 of the mosaic image 150, green information is sampled at two pixels, red information is sampled at one pixel, and blue information is sampled at one pixel.
[0023] Reference is made to Figure 2, which illustrates a method of processing a mosaic image into an output image having full color information at each pixel. Each pixel of the output image may be represented by a digital word including n-bits for red information, n-bits for green information, and n-bits for blue information.
[0024] The sampled values may be pre-processed (210). The preprocessing is application-specific. In a digital camera for example, preprocessing between image acquisition and demosaicing may include flare reduction, data linearization, and color balance.
[0025] The sampled values of the mosaic image are copied to their corresponding positions in the output image (220). For example, a mosaic image with a Bayer pattern has pixel inputs m(i,j) , where / and j are indices to the rows and columns of the mosaic image. The sampled values of the mosaic image may be copied to the output image as follows:
IR(i,j) = m(i,j) for all red samples (/ and j even), I0(JJ) = m(ij) for all green samples (/ + y odd), and IR QJ) - m(ij) for all blue samples (/ and j odd), where IR{i,j) represents red information at pixel (ij) of the output image, IG(i,j) represents green information at pixel (ij) of the output image, and IB{i,j) represents blue information at pixel (ij) of the output image.
[0026] A smooth estimation of the first color plane is computed (230). The smooth estimation may be a bilinear interpolation of the samples in the first color plane. Bilinear interpolation on the green plane may be computed on the green plane of a mosaic image with a Bayer pattern as oG = mG * k2 , where * denotes linear convolution, mG is the sampled values in the green plane, and
Figure imgf000007_0001
[0027] If bilinear interpolation is performed on the green plane of a mosaic image with a Bayer pattern, however, zippering artifacts could occur on edges. The zippering artifacts may be avoided by considering the two green samples of each cell separately. Those green samples having red samples as North-South neighbors (e.g., the green sample in the lower left corner of the cell 114 in Figure 1) are referred to as G1 samples; and those green samples having red samples as East-West neighbors (e.g., the green sample in the upper right corner of the cell 114 in Figure 1) are referred to as G2 samples. The G1 samples have the same spatial characteristic as the red and blue samples, and the G2 samples have the same spatial characteristic as the red and blue samples, and thus the bilinear interpolation of each of the G1 and G2 samples yields a smooth estimation. Therefore, the sum of the bilinear interpolation of the G1 samples with the sum of the bilinear interpolation of the G2 samples also yields a smooth estimation. Such a smooth estimation may be computed as
OG = mG * k] /2 , where * denotes linear convolution, mc is the sampled values in the green plane, and ki is
Figure imgf000007_0002
This smooth estimation does not contain artifacts that would otherwise occur in bilinear interpolation of the green plane. Moreover, the smooth estimation is blurrier than a bilinear interpolation (which also can be advantageous).
[0028] The red or blue plane may be used instead of the green plane. The smooth estimations of the red or blue plane may be a bilinear interpolation of the samples. Smooth estimations of the red or blue plane may be computed as
OR - mR * kx , or OB = mB * kl , where mR andmB are the sampled values in the red and blue planes. However, the green color plane is preferred because it corresponds to the largest component of the luminance, and because it has twice the amount of information in a Bayer cell compared to the other color planes (the green plane yields a stronger response by the human visual system than either the red plane or the blue plane).
[0029] The sampled values of the first color plane may be denoised (240). Denoising may follow the general approach illustrated in Figure 3: an error between a sampled value and the value of a corresponding pixel in the smooth estimation are computed (310), the error is modified according to local noise level (320), and the modified error is used to modify the corresponding pixel in the smooth estimation (330).
[0030] For example, the sampled values in the green plane may be denoised by computing I1. (/, j) = oG (/, j) + ψσ [mG (/', j) -oG (i, j)], where ψσ is an influence function that depends upon the noise level σ . The value ψσ(x) should be close to zero if |x| is smaller than σ and close to x otherwise. Thus 1^J' should be close to m(i,j) if the absolute difference of mGvJ) -°G(hj) i-S |arger than σ and close to oG(i,j) otherwise. Such filtering is edge-sensitive because the "error" grows significantly when one is close to an edge. In a smooth region, far from an edge, the region is smoothed. In a region near an edge, the region is not smoothed. Thus the denoising preserves strong features such as edges (i.e., it doesn't blur edges), yet removes small features such as noise.
[0031] Missing information in the first color plane of the output image is added to the output image by simultaneously performing edge-sensitive denoising and color interpolation on the first color (250). As illustrated in Figure 4, the edge-sensitive denoising and color interpolation of a given pixel includes determining a likelihood of whether the given pixel belongs to the same region as each of at least two neighboring denoised samples of the first color (410), the neighboring samples in different directions relative to the given pixel; and using the likelihoods and the denoised neighboring samples to interpolate the first color at the given pixel (420). A sample having a higher likelihood of belonging to the same region as the given pixel is given a higher weight in the interpolation. Different approaches for performing the simultaneous denoising and color interpolation will be described below. All of the approaches are sensitive to edges in the smooth estimation.
[0032] Missing information in the other color planes is interpolated and added to the output image (260). Blurring kernels may be used to populate the red and blue color planes as follows:
IR = jG + (pR -oa) * h and h = IG + (OB -oc) * h where h is a blurring kernel. The blurring kernel h may be
Figure imgf000009_0001
This blurring kernel can reduce color artifacts. A larger blurring kernel h (e.g., a 5x5 kernel) can further reduce the color artifacts.
[0033] Another method of interpolating the missing information in each of the red and blue color planes includes performing the simultaneous edge- sensitive denoising and color interpolation. Still other methods of interpolating the missing information in each of the red and blue color planes include bilinear interpolation, affine interpolation, and a canonical correlation method. These other methods are described in detail in assignee's U.S. Serial No. 10/743,623 filed December 22, 2003 and incorporated herein by reference. At the completion of step 260, each pixel /(/,_/) in the output image has red information IR(i,j) , green information IG(i,j) , and blue information IB(i, j) .
[0034] Post processing may be performed on the output (demosaiced) image (270). The post-processing is application-specific. In a digital camera for example, the post-processing may include color transformation, tone reproduction, sharpening, compression, and adaptive lighting.
[0035] Different approaches for performing the simultaneous denoising and color interpolation will now be described. The description of a first approach will refer to Figure 5. Figure 5 shows raw samples of a mosaic image 150. The given pixel (i.e., the pixel being processed) is circled. The given pixel in Figure 5 has a sampled red value, but no sampled values for green or blue.
[0036] The first approach involves determining a likelihood of whether the given pixel belongs to the same region as each of at least two other pixels having smooth estimates of the first color. The other pixels are in different directions relative to the given pixel. The samples and the likelihoods and the smooth estimates of the other pixels are used to interpolate the green information at the given pixel. This first approach may be performed according to equation (1).
∑ ψa [m(i + a,j + b) + Δα „ (ι, j) - oυ (i, jψa,b (ι, j)
Io (i, J) = oc (i, j) + ^^ v , . Λ ■ (1 )
(a,h)eNt
[0037] The term N4 represents a neighborhood consisting of four pixels. The index (a,b) indicates the direction and number of pixels relative to the given pixel. For example, the neighborhood N4 may consist of {(-1 ,0), (1 ,0), (0,- 1) and (0,1)}. The green samples are due west (-1 ,0) , due east (1 ,0), due south (0,-1) and due north (0,1) of the given pixel. Equation (1 ) holds only for those pixels (i,j) without green samples in the mosaic image 150.
[0038] Weight term wa,b may be computed as w '„*(*> j)
Figure imgf000010_0001
+ \ oG(i + 2aJ + 2b) - o(; (i + aJ +
Figure imgf000010_0002
\.
The first absolute difference oG(i + a,j + b) - oG (i,j) in the weight term wa,b represents the intensity of the difference between the smooth green estimate of the given pixel and that of the nearest green pixel in the direction determined by (a,b); and the second absolute difference in the weighting term o G (i + 2a,j + 2b) - oa(i + a,j + b)\ represents the intensity of the difference between the smooth green estimate of the nearest green pixel in the direction determined by (a,b) and that of the next pixel in the same direction. Thus the weight terms wa,b are based on smooth estimates, not raw samples. Use of the smooth estimates instead of raw samples in the second absolute difference allows differences between adjacent neighboring pixels to be considered. If raw samples are used, in contrast, the differences are between neighbors that are two pixels apart.
[0039] The function p() is referred to as a similarity function. The similarity function is used to determine the likelihood of a given pixel belonging to the same region as a neighboring pixel as a function of the sum of absolute differences corresponding to that direction. A large absolute difference in sampled values over a short distance (that is, an abrupt transition) is assumed to result from an edge. Thus a large sum of absolute differences in a given direction implies that the given pixel lies on one side of an edge (and belongs to one object), and the neighbor at that direction lies on the other side of the edge (and belongs to a different object). Therefore, a low likelihood exists that the green values of the given pixel and neighbor are similar. Exemplary similarity functions ρ(x) include, but are not limited to, the following: c2 p(x) = exp( — -)
p(x) = x + 1
T p(x) = min(|—,l)
|x| p(;t) = l/maxφc|,l) where T and σ are scaling parameters that account for the expected. noise level in the raw samples. The input x corresponds to a sum of absolute differences.
[0040] The term Δa,b represents high frequency correction terms such as Aa b(i,j) = λ[m(iJ) - m(i + 2aJ + 2b)] . The quantity [m(i,j) - m(i + 2a,j + 2b)] represents the difference between the red sample on the given pixel and the nearest red samples in the direction determined by (a,b) relative to the given pixel. [0041] The parameter A is a tunable parameter that corrects the interpolated value for camera-specific items, such as image resolution, camera optics, and scene distance (and, therefore, is used in ΔOii(/,y) but not the weight terms wa,b). If device resolution is small or edges are too abrupt, the parameter λ may have a value between 0.5 and 1. For mid-resolution cameras, the parameter λ may have a value of 0.5. For high resolution cameras, the parameter λ may have a value of less than 0.5.
[0042] A second approach is similar to the first approach, except that λ=0. By setting λ=0, the high frequency correction term Δa,b is not used, whereby equation (1) is reduced to the following:
Figure imgf000012_0001
. Setting λ=0 can cause a posterization
Figure imgf000012_0002
effect on the output image, but it can significantly reduce the computation of the output image.
[0043] A third approach involves taking the psi-sigma function (ψσ) and the subtraction by oG out of the loop of equation (1). This leads to the equation (3).
IcλiJ) - oa {i, j) (3)
Figure imgf000012_0003
This third approach (referred to as post-selective filtering) is faster than the first approach, since the psi-sigma function is applied only once for each input pixel (i,j) . In contrast, the first approach (referred to as pre-selective filtering) applies the psi-sigma function to each pixel (a,b) in a local neighborhood of each input pixel (i,j). The pre-selective filtering, however, tends to produce less noise around edges in the output image. [0044] A fourth approach involves combining the simplifications in the second approach (setting λ=0) and the third approach (post-selective filtering). The fourth approach may be performed according to equation (4).
∑mQ + aJ + fyw^faj)
IG(iJ) = oG (iJ) + ψc -oG(i,j) (4)
Figure imgf000013_0001
(a,b)≡Nt
[0045] To further improve processing speed, a lookup table may be used instead of applying the psi-sigma function. The lookup table stores samples (that is, pre-computed values) of the psi-sigma function.
[0046] Yet another approach may involve making hard decisions about interpolated values instead of using likelihoods. For example, define a = \2oG (/, j)- oG (/ + 2, j) - oG (i - 2, j) + \oG (i + 1, j) - oG (i - 1, j] and β = \2oG (/, j) - oG (i, j + 2)- dG {i, j - 2) + \oG (i, j + l)- oG (i, j - l] . The weighting terms wa h(i,j) may be computed as a < β => = W-l,0 = 1
W0,l = W0,-l = 0 a > β ^ Wl,0 = W-l,0 = 0 wo,ι = W0 _, - 1 a = β => = W-l,0 = wo,i = wo,-i = 1 -
[0047] The algorithms above are not limited to a Bayer CFA. The method may be applied to other color filter arrays.
[0048] The method is not limited to the N, S, E and W directions, or the four closest neighbors. Larger neighborhoods may be used, and pixels of different distances may be used. Weights may be reduced as distance increases from the given pixel.
[0049] Other methods of interpolating the missing information in each of the red and blue color planes include bilinear interpolation, affine interpolation, and a canonical correlation method.
[0050] Reference is made to Figure 6, which illustrates an exemplary hardware implementation. A digital camera 610 includes a photosensor array 612 such as a CCD. The photosensors of the array 612 may be arranged in a Bayer CFA. The array 612 produces mosaic digital images.
[0051] The digital camera 610 also includes a digital signal processor 614 and memory 616 (e.g., EPROM). A program 618 encoded in the memory 616 instructs the digital signal processor 614 to transform mosaic images into digital images having full color information at each pixel in accordance with an embodiment of the present invention. The program 618 may include a lookup table for different values of the similarity function. Weights are computed on the fly, and applied to the pixels using an algorithm.
[0052] Another implementation (not shown) may include an application- specific integrated circuit (ASIC). The ASIC includes a processor for transforming mosaic images into digital images having full color information at each pixel in accordance with an embodiment of the present invention. The processor may include a state machine, or it may include an embedded processor and memory, or it may have some other design.
[0053] Reference is now made to Figure 7, which illustrates another hardware implementation. A digital imaging system 710 includes a general purpose processor 712 and memory 714 for instructing the processor 712 to process mosaic images in accordance with an embodiment of the present invention. The system 710 may be, for example, a personal computer. Mosaic images may be supplied to the system 710 directly by a capture device (e.g., a scanner, a digital camera) or indirectly (via the Internet, accessed from remote or local storage, etc.).
[0054] A machine for generating the program may be a personal computer such as the one shown in Figure 7. Source code is generated, compiled into executable code, and distributed.
[0055] The executable code may be distributed in variety of ways. For example, the executable code may be distributed to a digital camera by burning the code into a memory 616 (e.g., the EPROM) of the digital camera 610. The executable code can be distributed to another computer via a network transmission, removable storage medium 718 (e.g., CD), etc.
[0056] Although specific embodiments of the present invention have been described and illustrated, the present invention is not limited to the specific forms or arrangements of parts so described and illustrated. Instead, the present invention is construed according to the following claims.

Claims

EDGE-SENSITIVE DENOISING AND COLOR INTERPOLATION OFDIGITAL IMAGESTHE CL-AIMS
1. Apparatus (610, 710 ) for processing a digital image, each pixel of the digital image having less than full information, the apparatus comprising a processor (614, 712) for simultaneously performing edge-sensitive denoising and color interpolation on a first color plane of the digital image (250), whereby the interpolation fills in missing pixel information in the first color plane.
2. The apparatus of claim 1 , wherein simultaneously performing the denoising and color interpolation on a pixel missing the first color information includes using smooth estimates and sampled values of neighboring pixels lying in different directions to generate the missing first color information.
3. The apparatus of claim 2, wherein simultaneously performing the denoising and color interpolation on a pixel missing the first color information further includes using likelihoods of the pixel belonging to the same region as its neighboring pixels (410-420).
4. The apparatus of claim 3, wherein the likelihoods are based on the smooth estimates.
5. The apparatus of claim 2, wherein simultaneously performing the denoising and color interpolation on a pixel missing the first color information further includes using high frequency correction information along with the sampled values and smooth estimates.
6. The apparatus of claim 2, wherein post-selective processing is performed.
7. The apparatus of claim 1 , wherein the simultaneous denoising and color interpolation on each pixel missing green information is performed
∑ Ψσ H' + aJ + b) + AaJ> (i, J) - OG (i, j)]vaJ> (ι, j) according to I0 (i, j) = oQ (i, j) + ^^
(.c ∑bytNW,_,»('»Λ
and AaJ)(iJ)=λ[m{i,j) - m(J + 2a,j + 2b)] and
w.fc0'.y) = pl oa(i + a,j + b)-θσ(i,j) + oG{i + 2a,j + 2b)-oo(i + a,j + b 4
8. The apparatus of claim 1 , wherein the simultaneous denoising and color interpolation on each pixel missing green information is performed according to
Figure imgf000017_0001
where
Figure imgf000017_0002
oσ(i +
Figure imgf000017_0003
\.
9. The apparatus of claim 1 , wherein the simultaneous denoising and color interpolation on each pixel missing green information is performed according to
Ia(i,J) = θG(i,j) + ψc Oc (j, j)
Figure imgf000017_0004
10. The apparatus of claim 1 , wherein the simultaneous denoising and color interpolation on each pixel missing green information is performed according to
IG(1J)
Figure imgf000017_0005
PCT/US2005/012363 2005-04-13 2005-04-13 Edge-sensitive denoising and color interpolation of digital images WO2006112814A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
PCT/US2005/012363 WO2006112814A1 (en) 2005-04-13 2005-04-13 Edge-sensitive denoising and color interpolation of digital images

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/US2005/012363 WO2006112814A1 (en) 2005-04-13 2005-04-13 Edge-sensitive denoising and color interpolation of digital images

Publications (1)

Publication Number Publication Date
WO2006112814A1 true WO2006112814A1 (en) 2006-10-26

Family

ID=34964956

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2005/012363 WO2006112814A1 (en) 2005-04-13 2005-04-13 Edge-sensitive denoising and color interpolation of digital images

Country Status (1)

Country Link
WO (1) WO2006112814A1 (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8135237B2 (en) 2008-02-25 2012-03-13 Aptina Imaging Corporation Apparatuses and methods for noise reduction
WO2013029231A1 (en) * 2011-08-30 2013-03-07 Technicolor (China) Technology Co., Ltd. Anisotropic gradient regularization for image denoising, compression, and interpolation
CN112349735A (en) * 2019-08-08 2021-02-09 爱思开海力士有限公司 Image sensor, image signal processor and image processing system including the same

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20030169353A1 (en) * 2002-03-11 2003-09-11 Renato Keshet Method and apparatus for processing sensor images
US20040028271A1 (en) * 2001-07-27 2004-02-12 Pollard Stephen Bernard Colour correction of images

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20040028271A1 (en) * 2001-07-27 2004-02-12 Pollard Stephen Bernard Colour correction of images
US20030169353A1 (en) * 2002-03-11 2003-09-11 Renato Keshet Method and apparatus for processing sensor images

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
LUKAC R ET AL: "NORMALIZED COLOR-RATIO MODELING FOR CFA INTERPOLATION", IEEE TRANSACTIONS ON CONSUMER ELECTRONICS, IEEE INC. NEW YORK, US, vol. 50, no. 2, May 2004 (2004-05-01), pages 737 - 745, XP001224994, ISSN: 0098-3063 *
RAMANATH R ET AL: "DEMOSAICKING AS A BILATERAL FILTERING PROCESS", PROCEEDINGS OF THE SPIE, SPIE, BELLINGHAM, VA, US, vol. 4667, 23 January 2002 (2002-01-23), pages 236 - 244, XP009004111, ISSN: 0277-786X *

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8135237B2 (en) 2008-02-25 2012-03-13 Aptina Imaging Corporation Apparatuses and methods for noise reduction
WO2013029231A1 (en) * 2011-08-30 2013-03-07 Technicolor (China) Technology Co., Ltd. Anisotropic gradient regularization for image denoising, compression, and interpolation
CN112349735A (en) * 2019-08-08 2021-02-09 爱思开海力士有限公司 Image sensor, image signal processor and image processing system including the same
CN112349735B (en) * 2019-08-08 2024-04-09 爱思开海力士有限公司 Image sensor, image signal processor and image processing system including the same

Similar Documents

Publication Publication Date Title
US7418130B2 (en) Edge-sensitive denoising and color interpolation of digital images
Delbracio et al. Mobile computational photography: A tour
US11625815B2 (en) Image processor and method
US7916940B2 (en) Processing of mosaic digital images
US8594451B2 (en) Edge mapping incorporating panchromatic pixels
US9747514B2 (en) Noise filtering and image sharpening utilizing common spatial support
JP5688090B2 (en) System and method for demosaic processing of image data using weighted gradient
CN102640489B (en) System and method for detecting and correcting defective pixels in an image sensor
CN111127336A (en) An Image Signal Processing Method Based on Adaptive Selection Module
Ratnasingam Deep camera: A fully convolutional neural network for image signal processing
EP1577833A1 (en) High-quality gradient-corrected linear interpolation for demosaicing of color images
US20170070689A1 (en) Automatic compensation of lens flare
KR101225056B1 (en) Apparatus and method for reducing noise from image sensor
US8238685B2 (en) Image noise reduction method and image processing apparatus using the same
CN101232574A (en) Imaging device, noise removal device, and noise removal method
CN111510691A (en) Color interpolation method and device, equipment and storage medium
US20240362745A1 (en) Methods and apparatus for blending unknown pixels in overlapping images
US7430334B2 (en) Digital imaging systems, articles of manufacture, and digital image processing methods
US7269295B2 (en) Digital image processing methods, digital image devices, and articles of manufacture
Buades et al. CFA video denoising and demosaicking chain via spatio-temporal patch-based filtering
Raimundo et al. Lan: Lightweight attention-based network for raw-to-rgb smartphone image processing
US7440016B2 (en) Method of processing a digital image
US20050123210A1 (en) Print processing of compressed noisy images
Adams Jr et al. Digital camera image processing chain design
WO2006112814A1 (en) Edge-sensitive denoising and color interpolation of digital images

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application
NENP Non-entry into the national phase

Ref country code: DE

NENP Non-entry into the national phase

Ref country code: RU

122 Ep: pct application non-entry in european phase

Ref document number: 05732556

Country of ref document: EP

Kind code of ref document: A1