US20050093994A1 - Method and apparatus for the removal of flash artifacts - Google Patents

Method and apparatus for the removal of flash artifacts Download PDF

Info

Publication number
US20050093994A1
US20050093994A1 US11/011,407 US1140704A US2005093994A1 US 20050093994 A1 US20050093994 A1 US 20050093994A1 US 1140704 A US1140704 A US 1140704A US 2005093994 A1 US2005093994 A1 US 2005093994A1
Authority
US
United States
Prior art keywords
flash
data
threshold
block
image
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US11/011,407
Inventor
John Baron
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Individual
Original Assignee
Individual
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Individual filed Critical Individual
Priority to US11/011,407 priority Critical patent/US20050093994A1/en
Publication of US20050093994A1 publication Critical patent/US20050093994A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/50Image enhancement or restoration by the use of more than one image, e.g. averaging, subtraction
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/70Circuitry for compensating brightness variation in the scene
    • H04N23/74Circuitry for compensating brightness variation in the scene by influencing the scene brightness using illuminating means

Definitions

  • the present invention relates to the field of photography, and more particularly, to the field of flash photography.
  • An image without use of a flash is taken, along with an image using a flash.
  • a difference image is generated by subtracting the flash-less image from the flashed image.
  • a threshold is applied to the difference image such that only large differences in intensity remain in the difference image.
  • This artifact image is then subtracted from the flash image, thereby removing flash artifacts such as specular reflections and red-eye.
  • the threshold used may be automatically calculated or may be set by the user. For some applications it may be desirable to set separate thresholds for each dimension of the color space (such as red, green, and blue) used. Once again these separate thresholds may be automatically calculated or may be set by the user.
  • FIG. 1 is a flowchart of a process for removing flash artifacts from a pair of digital images.
  • FIG. 2 is an example intensity histogram of an example difference image.
  • FIG. 3 is a flowchart of a process for removing flash artifacts from a pair of photographic slides, prints, or negatives.
  • FIG. 4 is a flowchart of a process for removing flash artifacts from a pair of digital images.
  • FIG. 5 is a block diagram of a device for removing flash artifacts from a pair of digital images.
  • FIG. 6 is a block diagram of a device similar to that of FIG. 5 where each primary color is treated independently.
  • FIG. 7 is a block diagram of a device similar to that of FIG. 6 wherein the artifact image for each primary color is subtracted from the flash image for each primary color before the primary colors are combined to produce a full-color image.
  • FIG. 8 is a flow chart of a method for removing flash artifacts from a pair of digital images on a pixel-by-pixel basis.
  • FIG. 1 is a flowchart of a process for removing flash artifacts from a pair of digital images.
  • a digital image is captured without the use of a flash.
  • another digital image of the same subject is captured with the use of a flash. Note that it makes no difference in which order these two digital images are captured. However, it is desirable that the subject does not move between the capture of these two digital images. This may be accomplished by taking the two images in rapid succession.
  • the flash-less digital image is subtracted from the flash digital image producing a difference image.
  • the digital images contain intensity data alone, and the difference image is a representation of the intensity changes in the image brought about by the use of the flash.
  • an intensity histogram is generated from the difference image. This intensity histogram may have a shape similar to the example histogram of FIG. 2 .
  • an intensity threshold is calculated from this intensity histogram. This intensity threshold is selected such that when applied to the difference image, in a step 112 , only the overly bright reflections will be stored in the artifact image, and the desirable changes in intensity from the flash will no longer exist in the artifact image.
  • this threshold step (or a similar process) were not done, subtracting the artifact (equal to the difference image in this case) image from the flash image would simply generate the flash-less image.
  • the intensity threshold is applied to the difference image. All portions of the difference image with an intensity greater than the intensity threshold are stored in an artifact image. This artifact image now contains intensity differences for just those areas of the image where the intensity difference due to the use of the flash are greater than what is desired.
  • some processing of the artifact image may be done to modify the effect of the artifact removal. For example, it may not be desirable to completely remove the strobe artifact, but to remove a fraction of the brightness difference instead of the entire difference.
  • the artifact image may be reduced by a factor, or it may be offset by a factor, either of which may be controlled by a user. Note that when no modification of the artifact image is desired, the multiplicative factor may be set to 1.0 or the additive factor may be set to 0.0 resulting in no change to the artifact image.
  • This processing of the artifact image as shown in step 114 , may be included in any or all of the implementations of the present invention described herein.
  • the modified artifact image is subtracted from the flash image resulting in a final image with the flash artifacts removed or modified. Also note that this algorithm helps correct for artifacts caused by portions of the image being too close to the flash. If there are parts of the subject much closer to the flash than other parts of the subject, the close parts may be unreasonably brightly lit by the flash and this method may replace those washed out parts of the image with the non-flashed image data.
  • FIG. 2 is an example intensity histogram of an example difference image.
  • the x-axis 200 represents intensity and the y-axis 202 represents frequency measured in number of pixels.
  • the possible intensities are quantized, so it is relatively easy to calculate how many pixels are represented by each possible intensity and produce a histogram similar to that shown in FIG. 2 .
  • the gray scale image data simply contains the intensity of each pixel and a histogram may be made directly from this data.
  • one axis of the color model may contain the intensity data, and this component alone of the color image data may be used to create the histogram.
  • the Y-axis is luminance and this component of the color image data is used to create the histogram.
  • the YIQ color model is used in television images transmitted according to the National Television System Committee (NTSC) standard.
  • This calculated luminance (or intensity) can then be used by the present invention to detect and correct intensity variations due to flashes in color images.
  • the example histogram of FIG. 2 shows two distinct lobes of intensity. First, there is a lower lobe 204 , that represents all the pixels where the intensity difference between the flash image and the flash-less image are present, but not extreme. However, there is also an upper lobe 206 , where the intensity difference between the flash image and the flash-less image are extreme. This upper lobe will contain pixels with specular reflection of the flash and possibly other flash artifacts.
  • an intensity threshold 208 may easily be selected at the frequency minimum between the two lobes.
  • all of the pixels with intensities less than the intensity threshold 208 have their intensity difference set to 0.
  • the pixels with intensities greater than the intensity threshold 208 have their intensity differences left as they were in the difference image.
  • the intensity threshold will have to be chosen by a method other than looking for a local minimum of the data. For example, the method may default to a set threshold when no local minimum exists, or it may be set at a percentage of the maximum intensity difference, or it may be set such that a certain percentage of the pixels are above the threshold.
  • FIG. 3 is a flowchart of a process for removing flash artifacts from a pair of photographic slides, prints, or negatives.
  • the process of removing flash artifacts starts with two non-digital images; one taken with a flash, and the other taken without a flash. Then, digital images are created of both non-digital images, for example, through the use of a scanner. This process is very similar to that described in FIG. 1 with two exceptions.
  • the flash-less digital image is captured from a flash-less slide, print, or negative, instead of directly from a subject.
  • the flash digital image is captured from a slide, print, or negative taken with the use of a flash, instead of directly from a subject.
  • the rest of the process follows the steps of the process described in FIG. 1 .
  • the flash-less digital image is subtracted from the flash digital image producing a difference image.
  • the digital images contain intensity data alone, and the difference image is a representation of the intensity changes in the image brought about by the use of the flash.
  • an intensity histogram is generated from the difference image. This intensity histogram may have a shape similar to the example histogram of FIG. 2 .
  • an intensity threshold is calculated. This intensity threshold is selected such that when applied to the difference image, in a step 112 , only the overly bright reflections will be stored in the artifact image, and the desirable changes in intensity from the flash will no longer exist in the artifact image. If this threshold step (or a similar process) were not done, subtracting the artifact image from the flash image would simply generate the flash-less image. In some implementations of the present invention, it may be desirable to allow the user to set the intensity threshold, or it may be permanently configured within the digital image capture device, or it may be selected through the use of methods other than a histogram that produce similar intensity threshold results. In a step 112 , the intensity threshold is applied to the difference image.
  • an artifact image All portions of the difference image with an intensity greater than the intensity threshold are stored in an artifact image.
  • This artifact image now contains intensity differences for just those areas of the image where the intensity difference due to the use of the flash are greater than what is desired.
  • some processing of the artifact image may be done to modify the effect of the artifact removal. For example, it may not be desirable to completely remove the strobe artifact, but to remove a fraction of the brightness difference instead of the entire difference.
  • the artifact image may be reduced by a factor, or it may be offset by a factor, either of which may be controlled by a user.
  • the multiplicative factor may be set to 1.0 or the additive factor may be set to 0.0 resulting in no change to the artifact image.
  • This processing of the artifact image may be included in any or all of the implementations of the present invention described herein.
  • the modified artifact image is subtracted from the flash image resulting in a final image with the flash artifacts removed or modified.
  • FIG. 4 is a flowchart of a process for removing flash artifacts from a pair of digital images.
  • color digital images are processed, and separate intensity thresholds 208 are calculated for each of the three primary colors within a color space.
  • many digital image capture devices use the red/green/blue (RGB) color space model comprising the three primary colors of red, green, and blue.
  • RGB red/green/blue
  • separate difference images would be created for each of the primary colors, and separate thresholds would be calculated for each of the primary colors.
  • the method is applied to each of the primary colors and will remove artifacts such as red eye where the subject's eyes reflect back as a bright red color from the flash.
  • each of the three primary colors red, green, and blue
  • HSV hue/saturation/value
  • the hue and saturation data define the shade of color and the value represents the intensity of the color.
  • the value data (the V in HSV) is used directly to determine the intensity threshold, whether by use of a histogram or another method. This process starts similar to the process described in FIG. 1 .
  • a digital image is captured without the use of a flash.
  • another digital image of the same subject is captured with the use of a flash. Note that it makes no difference in which order these two digital images are captured. However, it is desirable that the subject does not move between the capture of these two digital images. This may be accomplished by taking the two images in rapid succession.
  • separate difference images for each primary color are calculated by subtracting the flash-less digital images from the flash digital image. For example, in an RGB color space, the two digital images are stored as separate red, green, and blue digital images. The red digital image contains data for the primary color red.
  • the green and blue digital images contain data for the primary colors green and blue.
  • the red flash-less digital image is subtracted from the red flash image producing a red difference image.
  • the green and blue flash-less images are subtracted from the green and blue flash images producing green and blue difference images.
  • histograms are created for each of these red, green and blue difference images. Note that in some color space models, such as the HSV model discussed above, there may only be one histogram generated. In these color space models where only one histogram is generated, the present method will serve to correct variations in intensity, but will not detect color shifts unless the image data is converted to a color space model, such as RGB, where data for each primary color is available.
  • a step 404 three thresholds 208 are calculated for each of the three difference images using a process similar to that from the detailed description of FIG. 2 .
  • each threshold 208 is applied to its corresponding difference image, producing three artifact images for the primary colors of the color space. Note that this step results in an additional function of the present invention, the correction of color shifts due to a flash. For example, if the red channel shows a large increase in magnitude, but the blue and green remain relatively unchanged, this represents a red color shift in that portion of the image. In embodiments of the present invention using a single intensity histogram, this color shift may remain uncorrected. By processing each channel separately, color shifts without an accompanying intensity shift may be detected and corrected.
  • the three primary color artifact images are combined into a full color artifact image.
  • some processing of the artifact image may be done to modify the effect of the artifact removal. For example, it may not be desirable to completely remove the strobe artifact, but to remove a fraction of the brightness difference instead of the entire difference.
  • the artifact image may be reduced by a factor, or it may be offset by a factor, either of which may be controlled by a user. Note that when no modification of the artifact image is desired, the multiplicative factor may be set to 1.0 or the additive factor may be set to 0.0 resulting in no change to the artifact image.
  • step 410 This processing of the artifact image, as shown in step 410 , may be included in any or all of the implementations of the present invention described herein.
  • step 412 the full color artifact image is subtracted from the full color flash image producing a full color final image.
  • FIG. 5 is a block diagram of a device for removing flash artifacts from a pair of digital images.
  • This image data may be captured by an image capture device as described in the detailed descriptions for FIG. 1 and FIG. 4 , or alternatively may come from existing slides, prints, or negatives as described in the detailed description for FIG. 3 .
  • Flash-less digital image data 500 is subtracted from flash digital image data 502 , in a first subtract block 504 resulting in difference digital image data 506 .
  • This difference digital image data 506 is then used to calculate an intensity threshold 510 in a threshold block 508 .
  • the intensity threshold 510 may be calculated through the use of a histogram as described in the detailed description for FIG. 2 .
  • the gray scale data is intensity data and the difference image data is the used directly in calculating an intensity threshold 510 .
  • Alternative methods for determining the intensity threshold 510 such as a pre-set value, or user input, or other similar methods that are within the scope of the present invention.
  • a clipping block 512 all of the pixels in the difference digital image data 506 having an intensity less than the intensity threshold 510 are set to an intensity level of zero. This clipping block 512 produces the artifact image 514 .
  • some processing of the artifact image may be done to modify the effect of the artifact removal.
  • the artifact image may be reduced by a factor, or it may be offset by a factor, either of which may be controlled by a user.
  • the multiplicative factor may be set to 1.0 or the additive factor may be set to 0.0 resulting in no change to the artifact image.
  • This processing of the artifact image, as performed in the modification block 516 may be included in any or all of the implementations of the present invention described herein.
  • the modified artifact image 518 is then subtracted from the flash digital image 502 in a second subtract block 520 , producing the final flash-artifact-free image 522 .
  • this process shown in FIG. 5 may be applied on a pixel-by-pixel basis. This type of embodiment is shown in FIG. 7 and described below.
  • FIG. 6 is a block diagram of a device similar to that of FIG. 5 where each primary color is treated independently.
  • This image data may be captured by an image capture device as described in the detailed descriptions for FIG. 1 and FIG. 4 , or alternatively may come from existing slides, prints, or negatives as described in the detailed description for FIG. 3 .
  • Flash-less digital image data 500 enters the device at a flash-less split block 600 where the image data is separated into color data for each primary color in the current color space.
  • Flash digital image data 502 enters the device at a flash split block 602 where the image data is separated into color data for each primary color in the current color space.
  • RGB color space is used and each primary color has a magnitude (instead of an intensity value).
  • other color spaces such as Cyan/Magenta/Yellow (CMY) may be used instead of the RGB color space shown here.
  • CMY Cyan/Magenta/Yellow
  • some other color spaces may contain intensity data instead of saturation or magnitude data and the present invention is designed to work in either manner.
  • Color spaces, such as HSV, where there are no separate primary colors, will allow use of the present method to detect and correct variations in intensity due to a flash. However, unless the data is converted to a color space model with primary colors (such as RGB or CMY), shifts in color due to the flash will not be detected.
  • the flash-less split block 600 separates the flash-less digital image data into flash-less red data 603 , flash-less green data 604 , and flash-less blue data 605 .
  • the flash split block 602 separates the flash digital image data into flash red data 606 , flash green data 607 , and flash blue data 608 .
  • a first red subtraction block 610 the flash-less red data 603 is subtracted from the flash red data 606 , producing red difference data 620 .
  • the flash- less green data 604 is subtracted from the flash green data 607 , producing green difference data 622 .
  • a red threshold block 630 a red threshold 640 is determined using a histogram or other method.
  • a green threshold block 632 a green threshold 642 is determined using a histogram or other method.
  • a blue threshold block 634 a blue threshold 644 is determined using a histogram or other method.
  • a red clipping block 650 all of the pixels in the red difference data 620 having a value less than the red threshold 640 are set to a red saturation level of zero.
  • This red clipping block 650 produces the red artifact image 660 that contains red difference data for all of the pixels within the image where the red difference due to the use of the flash is greater than desired.
  • a green clipping block 652 all of the pixels in the green difference data 622 having a value less than the green threshold 642 are set to a green level of zero.
  • This green clipping block 652 produces the green artifact image 662 that contains green difference data for all of the pixels within the image where the green difference due to the use of the flash is greater than desired.
  • a blue clipping block 654 all of the pixels in the blue difference data 624 having a value less than the blue threshold 644 are set to a blue level of zero.
  • This blue clipping block 654 produces the blue artifact image 664 that contains blue difference data for all of the pixels within the image where the blue difference due to the use of the flash is greater than desired.
  • an artifact combination block 666 the red artifact image 660 , green artifact image 662 , and blue artifact image 664 are combined to produce a full-color artifact image 514 .
  • some processing of the artifact image may be included within the combination block 666 to modify the effect of the artifact removal. For example, it may not be desirable to completely remove the strobe artifact, but to remove a fraction of the brightness difference instead of the entire difference.
  • the artifact image may be reduced by a factor, or it may be offset by a factor, either of which may be controlled by a user.
  • the multiplicative factor may be set to 1.0 or the additive factor may be set to 0.0 resulting in no change to the artifact image.
  • This full-color artifact image 518 is then subtracted from the flash digital image 502 in a second subtract block 520 , producing the final flash-artifact-free image 522 . It is also possible to create an embodiment of the present invention in a system that uses more than three primary colors.
  • the additional primary colors may act to expand the range of the color space beyond the area covered by, for example, a red/green/blue color space and may be treated within the present invention as another primary color. While the additional primary colors may not be necessary in many implementations, the present invention still allows their use in removal of flash artifacts. Thus, while three (red, green, and blue) color pipelines are shown in FIG. 6 and FIG. 7 there may be additional pipelines present as needed by some implementations of the present invention.
  • FIG. 7 is a block diagram of a device similar to that of FIG. 6 wherein the artifact image for each primary color is subtracted from the flash image for each primary color before the primary colors are combined to produce a full-color image.
  • This image data may be captured by an image capture device as described in the detailed descriptions for FIG. 1 and FIG. 4 , or alternatively may come from existing slides, prints, or negatives as described in the detailed description for FIG. 3 .
  • Flash-less digital image data 500 enters the device at a flash-less split block 600 where the image data is separated into color data for each primary color in the current color space.
  • Flash digital image data 502 enters the device at a flash split block 602 where the image data is separated into color data for each primary color in the current color space.
  • an RGB color space is used and each primary color has a color value (instead of an intensity value).
  • other color spaces such as Cyan/Magenta/Yellow (CMY) may be used instead of the RGB color space shown here.
  • CML Cyan/Magenta/Yellow
  • some other color spaces may contain intensity data instead of saturation or magnitude data and the present invention is designed to work in either manner.
  • Color spaces, such as HSV, where there are no separate primary colors, will allow use of the present method to detect and correct variations in intensity due to a flash.
  • the flash-less split block 600 separates the flash-less digital image data into flash-less red data 603 , flash-less green data 604 , and flash-less blue data 605 .
  • the flash split block 602 separates the flash digital image data into flash red data 606 , flash green data 607 , and flash blue data 608 .
  • a first red subtraction block 610 the flash-less red data 603 is subtracted from the flash red data 606 , producing red difference data 620 .
  • a red threshold block 630 a red threshold 640 is determined using a histogram or other method.
  • a green threshold 642 is determined using a histogram or other method.
  • a blue threshold block 634 a blue threshold 644 is determined using a histogram or other method.
  • red clipping block 650 all of the pixels in the red difference data 620 having a value less than the red threshold 640 are set to a red level of zero.
  • This red clipping block 650 produces the red artifact image 660 that contains red difference data for all of the pixels within the image where the red difference due to the use of the flash is greater than desired.
  • green clipping block 652 all of the pixels in the green difference data 622 having a value less than the green threshold 642 are set to a green level of zero.
  • This green clipping block 652 produces the green artifact image 662 that contains green difference data for all of the pixels within the image where the green difference due to the use of the flash is greater than desired.
  • a blue clipping block 654 all of the pixels in the blue difference data 624 having a value less than the blue threshold 644 are set to a blue level of zero.
  • This blue clipping block 654 produces the blue artifact image 664 that contains blue difference data for all of the pixels within the image where the blue difference due to the use of the flash is greater than desired.
  • some processing of the artifact images may be included within the clipping blocks 650 , 652 , and 654 , or the second subtract blocks 700 , 702 , and 704 , to modify the effect of the artifact removal.
  • separate modification blocks may be included between the clipping blocks 650 , 652 , and 654 , and the second subtract blocks 700 , 702 , and 704 .
  • the artifact image may be reduced by a factor, or it may be offset by a factor, either of which may be controlled by a user. Note that when no modification of the artifact image is desired, the multiplicative factor may be set to 1.0 or the additive factor may be set to 0.0 resulting in no change to the artifact image.
  • a second red subtract block 700 the red artifact image 660 is subtracted from the flash red data 606 to produce the red final image 710 .
  • a second green subtract block 702 the green artifact image 662 is subtracted from the flash green data 607 to produce the green final image 712 .
  • a second blue subtract block 704 the blue artifact image 664 is subtracted from the flash blue data 608 to produce the blue final image 714 .
  • the red final image 710 , and the green final image 712 are combined with the blue final image 714 to produce the full-color final image 522 . It is also possible to create an embodiment of the present invention in a system that uses more than three primary colors.
  • the additional primary colors may act to expand the range of the color space beyond the area covered by, for example, a red/green/blue color space and may be treated within the present invention as another primary color. While the additional primary colors may not be necessary in many implementations, the present invention still allows their use in removal of flash artifacts. Thus, while three (red, green, and blue) color pipelines are shown in FIG. 6 and FIG. 7 there may be additional pipelines present as needed by some implementations of the present invention.
  • FIG. 8 is a flow chart of a method for removing flash artifacts from a pair of digital images on a pixel-by-pixel basis.
  • the example embodiment shown in FIG. 8 is similar to that shown in FIG. 1 , however the intensity calculations are done on a pixel-by-pixel basis.
  • FIG. 8 shows a method including capturing two digital images, an alternate embodiment of this method may capture images from slides, photos, or negatives as described in FIG. 3 .
  • a digital image is captured without the use of a flash.
  • a step 104 another digital image of the same subject is captured with the use of a flash. Note that it makes no difference in which order these two digital images are captured.
  • a threshold may be calculated and control passes to step 814 where the calculation is started. If an intensity threshold exists (possibly calculated earlier or set by a user) control passes to a step 802 where the pixel-by-pixel calculations begin. In a step 814 , an intensity difference histogram is created by subtracting the intensity data of the flash-less digital image from the intensity data of the flash digital image on a pixel-by-pixel basis.
  • This intensity difference histogram may have a shape similar to the example histogram of FIG. 2 .
  • an intensity threshold is calculated from the intensity difference histogram and control passes to a step 802 where the pixel-by-pixel calculations begin.
  • a pixel is selected.
  • the non-flash image intensity for the selected pixel is subtracted from the flash image intensity for the selected pixel resulting in a pixel difference.
  • a decision step 806 if the pixel difference is greater than the intensity threshold control is passed to a step 808 where the flash image data for the current pixel is overwritten by the non-flash image data for the current pixel, and control passes to another decision step 810 .
  • control may modify the flash image data for the current pixel by methods other than overwriting it by the non-flash image data. For example, an average of the flash and non-flash data may be used instead so that the area benefits from some of the brightening from the flash, but is less intense than the flash data. This is similar to the offset or factoring of the artifact data in other embodiments of the present invention described herein.
  • control is passed directly to the decision step 810 where the method determines if any pixels remain for calculation. If no pixels remain, control is passed to a finish step 812 where the process finishes and the final corrected flash image may then be stored. If pixels remain for calculation, control is passed back to a step 802 where another pixel is selected. Note that in some possible embodiments of the present invention it may not be necessary to run this process for each pixel in the image. In such cases the pixel selection step 802 may include sub-steps that insure that any pixels selected fall within the portion of the image selected for processing.
  • a pixel difference greater than the intensity threshold is required to correct the image data for the current pixel.
  • Other equivalent embodiments may check for a pixel difference greater than or equal to the intensity threshold. This is equivalent to decreasing the intensity threshold by one intensity unit.
  • the example embodiment of the present invention described in FIG. 8 will work equally well with color or gray scale data. Also, as shown in FIG. 6 and FIG. 7 , the method shown in FIG. 8 may be performed on individual axes of a color model either by running each axis separately, or by performing the operations in parallel. By performing the calculations on a pixel-by-pixel basis as described above, substantial savings in memory may be obtained.
  • this implementation does not require creation of a difference image, an artifact image, or of an additional final image.
  • the artifact image data is calculated pixel-by-pixel and only used to determine whether or not to overwrite the flash image data for the current pixel with the non-flash image data for the current pixel.
  • the flash image is selectively overwritten to become the final image, so that a final image in addition to the flash image is not needed.
  • a final image may be created by selectively writing either non-flash or flash image pixel data to a separate region of memory creating an additional final image.

Abstract

An image without use of a flash is taken, along with an image using a flash. A difference image is generated by subtracting the flash-less image from the flash image. A threshold is applied to the difference image such that only large differences in intensity remain in the difference image. This artifact image is then subtracted from the flash image, thereby removing flash artifacts such as specular reflections and red-eye. The threshold used may be automatically calculated or may be set by the user. For some applications it may be desirable to set separate thresholds for each dimension of the color space (such as red, green, and blue) used. Once again these separate thresholds may be automatically calculated or may be set by the user.

Description

    CROSS-REFERENCE TO RELATED APPLICATIONS
  • This application is a divisional of application Ser. No. 09/834,238 also entitled, “Method and Apparatus for the Removal of Flash Artifacts,” filed on Apr. 11, 2001 hereby incorporated herein by reference.
  • FIELD OF THE INVENTION
  • The present invention relates to the field of photography, and more particularly, to the field of flash photography.
  • BACKGROUND OF THE INVENTION
  • In the early days of photography, when people sat to have their portraits taken indoors, exposure times ranged up to ten minutes due to the slow speed of the photographic plates and the typically low light levels indoors. Often, the subjects of the photograph were subjected to being mechanically clamped in place to prevent movement during the exposure. With the advent of artificial lighting, initially the burning of magnesium powder, exposure times were shortened dramatically. Today, many cameras have a flash built-in, allowing users to take flash photographs at any opportunity.
  • However, the use of a flash does have its drawbacks. When the flash is positioned near the lens of the camera, the eyes of a human subject tend to glow red in the resulting photo. This problem is so common that many photography shops sell special pens used to remove the red-eye from the finished photo. Also, animals' eyes tend to glow similarly, but with different colors. Further, the flash photography of reflective objects may result in very bright specular reflections from the objects, thus detracting from the resulting photograph. There is a need in the art for a method and an apparatus to remove or modify these flash artifacts automatically.
  • SUMMARY OF THE INVENTION
  • An image without use of a flash is taken, along with an image using a flash. A difference image is generated by subtracting the flash-less image from the flashed image. A threshold is applied to the difference image such that only large differences in intensity remain in the difference image. This artifact image is then subtracted from the flash image, thereby removing flash artifacts such as specular reflections and red-eye. The threshold used may be automatically calculated or may be set by the user. For some applications it may be desirable to set separate thresholds for each dimension of the color space (such as red, green, and blue) used. Once again these separate thresholds may be automatically calculated or may be set by the user.
  • Other aspects and advantages of the present invention will become apparent from the following detailed description, taken in conjunction with the accompanying drawings, illustrating by way of example the principles of the invention.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • FIG. 1 is a flowchart of a process for removing flash artifacts from a pair of digital images.
  • FIG. 2 is an example intensity histogram of an example difference image.
  • FIG. 3 is a flowchart of a process for removing flash artifacts from a pair of photographic slides, prints, or negatives.
  • FIG. 4 is a flowchart of a process for removing flash artifacts from a pair of digital images.
  • FIG. 5 is a block diagram of a device for removing flash artifacts from a pair of digital images.
  • FIG. 6 is a block diagram of a device similar to that of FIG. 5 where each primary color is treated independently.
  • FIG. 7 is a block diagram of a device similar to that of FIG. 6 wherein the artifact image for each primary color is subtracted from the flash image for each primary color before the primary colors are combined to produce a full-color image.
  • FIG. 8 is a flow chart of a method for removing flash artifacts from a pair of digital images on a pixel-by-pixel basis.
  • DETAILED DESCRIPTION
  • FIG. 1 is a flowchart of a process for removing flash artifacts from a pair of digital images. In a step 102, a digital image is captured without the use of a flash. In a step 104, another digital image of the same subject is captured with the use of a flash. Note that it makes no difference in which order these two digital images are captured. However, it is desirable that the subject does not move between the capture of these two digital images. This may be accomplished by taking the two images in rapid succession. In a step 106, the flash-less digital image is subtracted from the flash digital image producing a difference image. In a black and white (gray scale) implementation of this invention, the digital images contain intensity data alone, and the difference image is a representation of the intensity changes in the image brought about by the use of the flash. In a step 108, an intensity histogram is generated from the difference image. This intensity histogram may have a shape similar to the example histogram of FIG. 2. From this intensity histogram, in a step 110, an intensity threshold is calculated. This intensity threshold is selected such that when applied to the difference image, in a step 112, only the overly bright reflections will be stored in the artifact image, and the desirable changes in intensity from the flash will no longer exist in the artifact image. If this threshold step (or a similar process) were not done, subtracting the artifact (equal to the difference image in this case) image from the flash image would simply generate the flash-less image. In some implementations of the present invention, it may be desirable to allow the user to set the intensity threshold, or it may be permanently configured within the digital image capture device, or it may be selected through the use of methods other than a histogram that produce similar intensity threshold results. In a step 112, the intensity threshold is applied to the difference image. All portions of the difference image with an intensity greater than the intensity threshold are stored in an artifact image. This artifact image now contains intensity differences for just those areas of the image where the intensity difference due to the use of the flash are greater than what is desired. In some implementations of the present invention, some processing of the artifact image may be done to modify the effect of the artifact removal. For example, it may not be desirable to completely remove the strobe artifact, but to remove a fraction of the brightness difference instead of the entire difference. The artifact image may be reduced by a factor, or it may be offset by a factor, either of which may be controlled by a user. Note that when no modification of the artifact image is desired, the multiplicative factor may be set to 1.0 or the additive factor may be set to 0.0 resulting in no change to the artifact image. This processing of the artifact image, as shown in step 114, may be included in any or all of the implementations of the present invention described herein. In a step 116, the modified artifact image is subtracted from the flash image resulting in a final image with the flash artifacts removed or modified. Also note that this algorithm helps correct for artifacts caused by portions of the image being too close to the flash. If there are parts of the subject much closer to the flash than other parts of the subject, the close parts may be unreasonably brightly lit by the flash and this method may replace those washed out parts of the image with the non-flashed image data.
  • FIG. 2 is an example intensity histogram of an example difference image. In this histogram the x-axis 200 represents intensity and the y-axis 202 represents frequency measured in number of pixels. In a digital camera the possible intensities are quantized, so it is relatively easy to calculate how many pixels are represented by each possible intensity and produce a histogram similar to that shown in FIG. 2. In an example embodiment of the present invention using gray scale data, the gray scale image data simply contains the intensity of each pixel and a histogram may be made directly from this data. Also, in some color space models, such as the YIQ model used in commercial color television broadcasts within the United States, one axis of the color model may contain the intensity data, and this component alone of the color image data may be used to create the histogram. In the YIQ color model example, the Y-axis is luminance and this component of the color image data is used to create the histogram. The YIQ color model is used in television images transmitted according to the National Television System Committee (NTSC) standard. RGB color data may be converted to YIQ data using a standard mapping algorithm. For example, the formula to extract the luminance (Y in YIQ) from RGB color data is Y=0.299R+0.587G+0.114B. This calculated luminance (or intensity) can then be used by the present invention to detect and correct intensity variations due to flashes in color images. There are many different models of color space. The present invention will work properly for all of them as long as some method exists to extract intensity, luminance, or brightness from the color data. The example histogram of FIG. 2 shows two distinct lobes of intensity. First, there is a lower lobe 204, that represents all the pixels where the intensity difference between the flash image and the flash-less image are present, but not extreme. However, there is also an upper lobe 206, where the intensity difference between the flash image and the flash-less image are extreme. This upper lobe will contain pixels with specular reflection of the flash and possibly other flash artifacts. From this example histogram, an intensity threshold 208 may easily be selected at the frequency minimum between the two lobes. Thus, when the artifact image is created from the difference image, all of the pixels with intensities less than the intensity threshold 208, have their intensity difference set to 0. Then, the pixels with intensities greater than the intensity threshold 208, have their intensity differences left as they were in the difference image. Note that some images may not have a nice easy bimodal distribution of intensity values as shown in this figure. In such a case, the intensity threshold will have to be chosen by a method other than looking for a local minimum of the data. For example, the method may default to a set threshold when no local minimum exists, or it may be set at a percentage of the maximum intensity difference, or it may be set such that a certain percentage of the pixels are above the threshold.
  • FIG. 3 is a flowchart of a process for removing flash artifacts from a pair of photographic slides, prints, or negatives. In this example embodiment of the present invention, the process of removing flash artifacts starts with two non-digital images; one taken with a flash, and the other taken without a flash. Then, digital images are created of both non-digital images, for example, through the use of a scanner. This process is very similar to that described in FIG. 1 with two exceptions. First, in a step 302, the flash-less digital image is captured from a flash-less slide, print, or negative, instead of directly from a subject. Second, in a step 304, the flash digital image is captured from a slide, print, or negative taken with the use of a flash, instead of directly from a subject. The rest of the process follows the steps of the process described in FIG. 1. In a step 106, the flash-less digital image is subtracted from the flash digital image producing a difference image. In a black and white (gray scale) implementation of this invention, the digital images contain intensity data alone, and the difference image is a representation of the intensity changes in the image brought about by the use of the flash. In a step 108, an intensity histogram is generated from the difference image. This intensity histogram may have a shape similar to the example histogram of FIG. 2. From this intensity histogram, in a step 110, an intensity threshold is calculated. This intensity threshold is selected such that when applied to the difference image, in a step 112, only the overly bright reflections will be stored in the artifact image, and the desirable changes in intensity from the flash will no longer exist in the artifact image. If this threshold step (or a similar process) were not done, subtracting the artifact image from the flash image would simply generate the flash-less image. In some implementations of the present invention, it may be desirable to allow the user to set the intensity threshold, or it may be permanently configured within the digital image capture device, or it may be selected through the use of methods other than a histogram that produce similar intensity threshold results. In a step 112, the intensity threshold is applied to the difference image. All portions of the difference image with an intensity greater than the intensity threshold are stored in an artifact image. This artifact image now contains intensity differences for just those areas of the image where the intensity difference due to the use of the flash are greater than what is desired. In some implementations of the present invention, some processing of the artifact image may be done to modify the effect of the artifact removal. For example, it may not be desirable to completely remove the strobe artifact, but to remove a fraction of the brightness difference instead of the entire difference. The artifact image may be reduced by a factor, or it may be offset by a factor, either of which may be controlled by a user. Note that when no modification of the artifact image is desired, the multiplicative factor may be set to 1.0 or the additive factor may be set to 0.0 resulting in no change to the artifact image. This processing of the artifact image, as shown in step 114, may be included in any or all of the implementations of the present invention described herein. In a step 116, the modified artifact image is subtracted from the flash image resulting in a final image with the flash artifacts removed or modified.
  • FIG. 4 is a flowchart of a process for removing flash artifacts from a pair of digital images. In this example embodiment of the present invention, color digital images are processed, and separate intensity thresholds 208 are calculated for each of the three primary colors within a color space. For example, many digital image capture devices use the red/green/blue (RGB) color space model comprising the three primary colors of red, green, and blue. In the RGB example, separate difference images would be created for each of the primary colors, and separate thresholds would be calculated for each of the primary colors. In this example, the method is applied to each of the primary colors and will remove artifacts such as red eye where the subject's eyes reflect back as a bright red color from the flash. In the case of red eye, the magnitude of the red component of the image will show a large increase in the area of the subject's eyes and this will be detected and removed or modified by the method. Many other color space models exist and are compatible with the present invention. In the RGB color space model, each of the three primary colors (red, green, and blue) is represented by one axis of the three-dimensional color space model. In contrast, in a hue/saturation/value (HSV) color space model the hue and saturation data define the shade of color and the value represents the intensity of the color. In a HSV color space model, the value data (the V in HSV) is used directly to determine the intensity threshold, whether by use of a histogram or another method. This process starts similar to the process described in FIG. 1. In a step 102, a digital image is captured without the use of a flash. In a step 104, another digital image of the same subject is captured with the use of a flash. Note that it makes no difference in which order these two digital images are captured. However, it is desirable that the subject does not move between the capture of these two digital images. This may be accomplished by taking the two images in rapid succession. In a step 400, separate difference images for each primary color (or axis of color space model) are calculated by subtracting the flash-less digital images from the flash digital image. For example, in an RGB color space, the two digital images are stored as separate red, green, and blue digital images. The red digital image contains data for the primary color red. Likewise, the green and blue digital images contain data for the primary colors green and blue. The red flash-less digital image is subtracted from the red flash image producing a red difference image. Likewise, the green and blue flash-less images are subtracted from the green and blue flash images producing green and blue difference images. In a step 402, histograms are created for each of these red, green and blue difference images. Note that in some color space models, such as the HSV model discussed above, there may only be one histogram generated. In these color space models where only one histogram is generated, the present method will serve to correct variations in intensity, but will not detect color shifts unless the image data is converted to a color space model, such as RGB, where data for each primary color is available. In a step 404, three thresholds 208 are calculated for each of the three difference images using a process similar to that from the detailed description of FIG. 2. In a step 406, each threshold 208 is applied to its corresponding difference image, producing three artifact images for the primary colors of the color space. Note that this step results in an additional function of the present invention, the correction of color shifts due to a flash. For example, if the red channel shows a large increase in magnitude, but the blue and green remain relatively unchanged, this represents a red color shift in that portion of the image. In embodiments of the present invention using a single intensity histogram, this color shift may remain uncorrected. By processing each channel separately, color shifts without an accompanying intensity shift may be detected and corrected. In a step 408, the three primary color artifact images are combined into a full color artifact image. In some implementations of the present invention, some processing of the artifact image may be done to modify the effect of the artifact removal. For example, it may not be desirable to completely remove the strobe artifact, but to remove a fraction of the brightness difference instead of the entire difference. The artifact image may be reduced by a factor, or it may be offset by a factor, either of which may be controlled by a user. Note that when no modification of the artifact image is desired, the multiplicative factor may be set to 1.0 or the additive factor may be set to 0.0 resulting in no change to the artifact image. This processing of the artifact image, as shown in step 410, may be included in any or all of the implementations of the present invention described herein. Finally, in a step 412, the full color artifact image is subtracted from the full color flash image producing a full color final image.
  • FIG. 5 is a block diagram of a device for removing flash artifacts from a pair of digital images. In this example embodiment of the present invention, it is assumed that both flash-less and flash digital image data is available at the inputs to the device. This image data may be captured by an image capture device as described in the detailed descriptions for FIG. 1 and FIG. 4, or alternatively may come from existing slides, prints, or negatives as described in the detailed description for FIG. 3. Flash-less digital image data 500 is subtracted from flash digital image data 502, in a first subtract block 504 resulting in difference digital image data 506. This difference digital image data 506 is then used to calculate an intensity threshold 510 in a threshold block 508. The intensity threshold 510 may be calculated through the use of a histogram as described in the detailed description for FIG. 2. Note that in a gray scale system, the gray scale data is intensity data and the difference image data is the used directly in calculating an intensity threshold 510. Alternative methods for determining the intensity threshold 510 such as a pre-set value, or user input, or other similar methods that are within the scope of the present invention. In a clipping block 512 all of the pixels in the difference digital image data 506 having an intensity less than the intensity threshold 510 are set to an intensity level of zero. This clipping block 512 produces the artifact image 514. In some implementations of the present invention, some processing of the artifact image may be done to modify the effect of the artifact removal. For example, it may not be desirable to completely remove the strobe artifact, but to remove a fraction of the brightness difference instead of the entire difference. The artifact image may be reduced by a factor, or it may be offset by a factor, either of which may be controlled by a user. Note that when no modification of the artifact image is desired, the multiplicative factor may be set to 1.0 or the additive factor may be set to 0.0 resulting in no change to the artifact image. This processing of the artifact image, as performed in the modification block 516, may be included in any or all of the implementations of the present invention described herein. The modified artifact image 518 is then subtracted from the flash digital image 502 in a second subtract block 520, producing the final flash-artifact-free image 522. In another example embodiment of the present invention, this process shown in FIG. 5 may be applied on a pixel-by-pixel basis. This type of embodiment is shown in FIG. 7 and described below.
  • FIG. 6 is a block diagram of a device similar to that of FIG. 5 where each primary color is treated independently. In this example embodiment of the present invention, it is assumed that both flash-less and flash digital image data are available at the inputs to the device. This image data may be captured by an image capture device as described in the detailed descriptions for FIG. 1 and FIG. 4, or alternatively may come from existing slides, prints, or negatives as described in the detailed description for FIG. 3. Flash-less digital image data 500 enters the device at a flash-less split block 600 where the image data is separated into color data for each primary color in the current color space. Flash digital image data 502 enters the device at a flash split block 602 where the image data is separated into color data for each primary color in the current color space. In this example embodiment of the present invention, an RGB color space is used and each primary color has a magnitude (instead of an intensity value). In other embodiments of the present invention, other color spaces such as Cyan/Magenta/Yellow (CMY) may be used instead of the RGB color space shown here. Note that some other color spaces may contain intensity data instead of saturation or magnitude data and the present invention is designed to work in either manner. Color spaces, such as HSV, where there are no separate primary colors, will allow use of the present method to detect and correct variations in intensity due to a flash. However, unless the data is converted to a color space model with primary colors (such as RGB or CMY), shifts in color due to the flash will not be detected. In this example embodiment, the flash-less split block 600 separates the flash-less digital image data into flash-less red data 603, flash-less green data 604, and flash-less blue data 605. Likewise, the flash split block 602 separates the flash digital image data into flash red data 606, flash green data 607, and flash blue data 608. In a first red subtraction block 610, the flash-less red data 603 is subtracted from the flash red data 606, producing red difference data 620. In a first green subtraction block 612, the flash- less green data 604 is subtracted from the flash green data 607, producing green difference data 622. In a first blue subtraction block 614, the flash-less blue data 605 is subtracted from the flash blue data 608, producing blue difference data 624. In a red threshold block 630 a red threshold 640 is determined using a histogram or other method. In a green threshold block 632 a green threshold 642 is determined using a histogram or other method. In a blue threshold block 634 a blue threshold 644 is determined using a histogram or other method. In a red clipping block 650 all of the pixels in the red difference data 620 having a value less than the red threshold 640 are set to a red saturation level of zero. This red clipping block 650 produces the red artifact image 660 that contains red difference data for all of the pixels within the image where the red difference due to the use of the flash is greater than desired. In a green clipping block 652 all of the pixels in the green difference data 622 having a value less than the green threshold 642 are set to a green level of zero. This green clipping block 652 produces the green artifact image 662 that contains green difference data for all of the pixels within the image where the green difference due to the use of the flash is greater than desired. In a blue clipping block 654 all of the pixels in the blue difference data 624 having a value less than the blue threshold 644 are set to a blue level of zero. This blue clipping block 654 produces the blue artifact image 664 that contains blue difference data for all of the pixels within the image where the blue difference due to the use of the flash is greater than desired. In an artifact combination block 666, the red artifact image 660, green artifact image 662, and blue artifact image 664 are combined to produce a full-color artifact image 514. In some implementations of the present invention, some processing of the artifact image may be included within the combination block 666 to modify the effect of the artifact removal. For example, it may not be desirable to completely remove the strobe artifact, but to remove a fraction of the brightness difference instead of the entire difference. The artifact image may be reduced by a factor, or it may be offset by a factor, either of which may be controlled by a user. Note that when no modification of the artifact image is desired, the multiplicative factor may be set to 1.0 or the additive factor may be set to 0.0 resulting in no change to the artifact image. This full-color artifact image 518 is then subtracted from the flash digital image 502 in a second subtract block 520, producing the final flash-artifact-free image 522. It is also possible to create an embodiment of the present invention in a system that uses more than three primary colors. The additional primary colors may act to expand the range of the color space beyond the area covered by, for example, a red/green/blue color space and may be treated within the present invention as another primary color. While the additional primary colors may not be necessary in many implementations, the present invention still allows their use in removal of flash artifacts. Thus, while three (red, green, and blue) color pipelines are shown in FIG. 6 and FIG. 7 there may be additional pipelines present as needed by some implementations of the present invention.
  • FIG. 7 is a block diagram of a device similar to that of FIG. 6 wherein the artifact image for each primary color is subtracted from the flash image for each primary color before the primary colors are combined to produce a full-color image. In this example embodiment of the present invention, it is assumed that both flash-less and flash digital image data are available at the inputs to the device. This image data may be captured by an image capture device as described in the detailed descriptions for FIG. 1 and FIG. 4, or alternatively may come from existing slides, prints, or negatives as described in the detailed description for FIG. 3. Flash-less digital image data 500 enters the device at a flash-less split block 600 where the image data is separated into color data for each primary color in the current color space. Flash digital image data 502 enters the device at a flash split block 602 where the image data is separated into color data for each primary color in the current color space. In this example embodiment of the present invention, an RGB color space is used and each primary color has a color value (instead of an intensity value). In other embodiments of the present invention, other color spaces such as Cyan/Magenta/Yellow (CMY) may be used instead of the RGB color space shown here. Note that some other color spaces may contain intensity data instead of saturation or magnitude data and the present invention is designed to work in either manner. Color spaces, such as HSV, where there are no separate primary colors, will allow use of the present method to detect and correct variations in intensity due to a flash. However, unless the data is converted to a color space model with primary colors (such as RGB or CMY), shifts in color due to the flash will not be detected. In this example embodiment the flash-less split block 600 separates the flash-less digital image data into flash-less red data 603, flash-less green data 604, and flash-less blue data 605. Likewise, the flash split block 602 separates the flash digital image data into flash red data 606, flash green data 607, and flash blue data 608. In a first red subtraction block 610, the flash-less red data 603 is subtracted from the flash red data 606, producing red difference data 620. In a first green subtraction block 612, the flash-less green data 604 is subtracted from the flash green data 607, producing green difference data 622. In a first blue subtraction block 614, the flash-less blue data 605 is subtracted from the flash blue data 608, producing blue difference data 624. In a red threshold block 630 a red threshold 640 is determined using a histogram or other method. In a green threshold block 632 a green threshold 642 is determined using a histogram or other method. In a blue threshold block 634 a blue threshold 644 is determined using a histogram or other method. In a red clipping block 650 all of the pixels in the red difference data 620 having a value less than the red threshold 640 are set to a red level of zero. This red clipping block 650 produces the red artifact image 660 that contains red difference data for all of the pixels within the image where the red difference due to the use of the flash is greater than desired. In a green clipping block 652 all of the pixels in the green difference data 622 having a value less than the green threshold 642 are set to a green level of zero. This green clipping block 652 produces the green artifact image 662 that contains green difference data for all of the pixels within the image where the green difference due to the use of the flash is greater than desired. In a blue clipping block 654 all of the pixels in the blue difference data 624 having a value less than the blue threshold 644 are set to a blue level of zero. This blue clipping block 654 produces the blue artifact image 664 that contains blue difference data for all of the pixels within the image where the blue difference due to the use of the flash is greater than desired. In some implementations of the present invention, some processing of the artifact images may be included within the clipping blocks 650, 652, and 654, or the second subtract blocks 700, 702, and 704, to modify the effect of the artifact removal. Alternatively, separate modification blocks may be included between the clipping blocks 650, 652, and 654, and the second subtract blocks 700, 702, and 704. For example, it may not be desirable to completely remove the strobe artifact, but to remove a fraction of the brightness difference instead of the entire difference. The artifact image may be reduced by a factor, or it may be offset by a factor, either of which may be controlled by a user. Note that when no modification of the artifact image is desired, the multiplicative factor may be set to 1.0 or the additive factor may be set to 0.0 resulting in no change to the artifact image. In a second red subtract block 700, the red artifact image 660 is subtracted from the flash red data 606 to produce the red final image 710. In a second green subtract block 702, the green artifact image 662 is subtracted from the flash green data 607 to produce the green final image 712. In a second blue subtract block 704, the blue artifact image 664 is subtracted from the flash blue data 608 to produce the blue final image 714. In a final combination block 720, the red final image 710, and the green final image 712, are combined with the blue final image 714 to produce the full-color final image 522. It is also possible to create an embodiment of the present invention in a system that uses more than three primary colors. The additional primary colors may act to expand the range of the color space beyond the area covered by, for example, a red/green/blue color space and may be treated within the present invention as another primary color. While the additional primary colors may not be necessary in many implementations, the present invention still allows their use in removal of flash artifacts. Thus, while three (red, green, and blue) color pipelines are shown in FIG. 6 and FIG. 7 there may be additional pipelines present as needed by some implementations of the present invention.
  • FIG. 8 is a flow chart of a method for removing flash artifacts from a pair of digital images on a pixel-by-pixel basis. The example embodiment shown in FIG. 8 is similar to that shown in FIG. 1, however the intensity calculations are done on a pixel-by-pixel basis. Note that while FIG. 8 shows a method including capturing two digital images, an alternate embodiment of this method may capture images from slides, photos, or negatives as described in FIG. 3. In a step 102, a digital image is captured without the use of a flash. In a step 104, another digital image of the same subject is captured with the use of a flash. Note that it makes no difference in which order these two digital images are captured. However, it is desirable that the subject does not move between the capture of these two digital images. This may be accomplished by taking the two images in rapid succession. In a decision step 800, if no intensity threshold currently exists, a threshold may be calculated and control passes to step 814 where the calculation is started. If an intensity threshold exists (possibly calculated earlier or set by a user) control passes to a step 802 where the pixel-by-pixel calculations begin. In a step 814, an intensity difference histogram is created by subtracting the intensity data of the flash-less digital image from the intensity data of the flash digital image on a pixel-by-pixel basis. Since the difference data is calculated on a pixel-by-pixel basis simply for the creation of a histogram, there is no need to retain the difference data for the entire image in a difference image. This intensity difference histogram may have a shape similar to the example histogram of FIG. 2. Once the histogram is completed, in a step 816, an intensity threshold is calculated from the intensity difference histogram and control passes to a step 802 where the pixel-by-pixel calculations begin. In a step 802, a pixel is selected. In a step 804, the non-flash image intensity for the selected pixel is subtracted from the flash image intensity for the selected pixel resulting in a pixel difference. In a decision step 806, if the pixel difference is greater than the intensity threshold control is passed to a step 808 where the flash image data for the current pixel is overwritten by the non-flash image data for the current pixel, and control passes to another decision step 810. Note that other embodiments of the present invention may modify the flash image data for the current pixel by methods other than overwriting it by the non-flash image data. For example, an average of the flash and non-flash data may be used instead so that the area benefits from some of the brightening from the flash, but is less intense than the flash data. This is similar to the offset or factoring of the artifact data in other embodiments of the present invention described herein. In the decision step 806, if the pixel difference is not greater than the intensity threshold, control is passed directly to the decision step 810 where the method determines if any pixels remain for calculation. If no pixels remain, control is passed to a finish step 812 where the process finishes and the final corrected flash image may then be stored. If pixels remain for calculation, control is passed back to a step 802 where another pixel is selected. Note that in some possible embodiments of the present invention it may not be necessary to run this process for each pixel in the image. In such cases the pixel selection step 802 may include sub-steps that insure that any pixels selected fall within the portion of the image selected for processing. Also note that in the example embodiment described above, a pixel difference greater than the intensity threshold is required to correct the image data for the current pixel. Other equivalent embodiments may check for a pixel difference greater than or equal to the intensity threshold. This is equivalent to decreasing the intensity threshold by one intensity unit. Further, note that the example embodiment of the present invention described in FIG. 8 will work equally well with color or gray scale data. Also, as shown in FIG. 6 and FIG. 7, the method shown in FIG. 8 may be performed on individual axes of a color model either by running each axis separately, or by performing the operations in parallel. By performing the calculations on a pixel-by-pixel basis as described above, substantial savings in memory may be obtained. Note that this implementation does not require creation of a difference image, an artifact image, or of an additional final image. The artifact image data is calculated pixel-by-pixel and only used to determine whether or not to overwrite the flash image data for the current pixel with the non-flash image data for the current pixel. Thus, the flash image is selectively overwritten to become the final image, so that a final image in addition to the flash image is not needed. Also, note that if it is desired to preserve the flash image (perhaps for later calculations) a final image may be created by selectively writing either non-flash or flash image pixel data to a separate region of memory creating an additional final image.
  • The foregoing description of the present invention has been presented for purposes of illustration and description. It is not intended to be exhaustive or to limit the invention to the precise form disclosed, and other modifications and variations may be possible in light of the above teachings. The embodiment was chosen and described in order to best explain the principles of the invention and its practical application to thereby enable others skilled in the art to best utilize the invention in various embodiments and various modifications as are suited to the particular use contemplated. It is intended that the appended claims be construed to include other alternative embodiments of the invention except insofar as limited by the prior art.

Claims (20)

1. A device for removing flash artifacts comprising:
a first subtract block with inputs comprising flash-less digital image data and flash digital image data, wherein said flash-less digital image data is subtracted from said flash digital image data producing difference data;
a threshold block connected to said first subtract block, that determines a threshold from said difference data;
a clipping block connected to said threshold block, that receives said difference data and said threshold and outputs artifact data, wherein said artifact data contains said difference data for said pixels with a value greater than said threshold; and
a second subtract block connected to said clipping block, wherein said second subtract block subtracts said artifact data from said flash digital image data, resulting in a final digital image.
2. A device for removing flash artifacts as recited in claim 1;
wherein said difference data comprises intensity data.
3. A device for removing flash artifacts as recited in claim 1;
wherein said threshold block calculates said threshold from a histogram.
4. A device for removing flash artifacts as recited in claim 1;
wherein said threshold block determines said threshold from a user input.
5. A device for removing flash artifacts comprising:
a first subtract block with inputs comprising flash-less digital image data and flash digital image data, wherein said flash-less digital image data is subtracted from said flash digital image data producing difference data;
a threshold block connected to said first subtract block, that determines a threshold from said difference data;
a clipping block connected to said threshold block, that receives said difference data and said threshold and outputs artifact data, wherein said artifact data contains said difference data for said pixels with a value greater than said threshold;
a factor block connected to said clipping block that modifies said artifact data by a factor and outputs factored artifact data; and
a second subtract block connected to said factor block, wherein said second subtract block subtracts said factored artifact data from said flash digital image data, resulting in a final digital image.
6. A device for removing flash artifacts as recited in claim 5;
wherein said factor block multiplies said artifact data by said factor.
7. A device for removing flash artifacts as recited in claim 5;
wherein said factor block adds said artifact data to said factor.
8. A device for removing flash artifacts as recited in claim 5;
wherein said difference data comprises intensity data.
9. A device for removing flash artifacts as recited in claim 5;
wherein said threshold block calculates said threshold from a histogram.
10. A device for removing flash artifacts as recited in claim 5;
wherein said threshold block determines said threshold from a user input.
11. A device for removing flash artifacts comprising:
a first subtract block with inputs comprising flash-less digital image data and flash digital image data, wherein said flash-less digital image data is subtracted from said flash digital image data producing difference data;
a threshold block connected to said first subtract block, that determines a threshold from said difference data;
a clipping block connected to said threshold block, that receives said difference data and said threshold and outputs artifact data, wherein said artifact data contains said difference data for said pixels with a value greater than or equal to said threshold; and
a second subtract block connected to said clipping block, wherein said second subtract block subtracts said artifact data from said flash digital image data, resulting in a final digital image.
12. A device for removing flash artifacts as recited in claim 11;
wherein said difference data comprises intensity data.
13. A device for removing flash artifacts as recited in claim 11;
wherein said threshold block calculates said threshold from a histogram.
14. A device for removing flash artifacts as recited in claim 11;
wherein said threshold block determines said threshold from a user input.
15. A device for removing flash artifacts comprising:
a first subtract block with inputs comprising flash-less digital image data and flash digital image data, wherein said flash-less digital image data is subtracted from said flash digital image data producing difference data;
a threshold block connected to said first subtract block, that determines a threshold from said difference data;
a clipping block connected to said threshold block, that receives said difference data and said threshold and outputs artifact data, wherein said artifact data contains said difference data for said pixels with a value greater than or equal to said threshold;
a factor block connected to said clipping block that modifies said artifact data by a factor and outputs factored artifact data; and
a second subtract block connected to said factor block, wherein said second subtract block subtracts said factored artifact data from said flash digital image data, resulting in a final digital image.
16. A device for removing flash artifacts as recited in claim 15;
wherein said factor block multiplies said artifact data by said factor.
17. A device for removing flash artifacts as recited in claim 15;
wherein said factor block adds said artifact data to said factor.
18. A device for removing flash artifacts as recited in claim 15;
wherein said difference data comprises intensity data.
19. A device for removing flash artifacts as recited in claim 15;
wherein said threshold block calculates said threshold from a histogram.
20. A device for removing flash artifacts as recited in claim 15;
wherein said threshold block determines said threshold from a user input.
US11/011,407 2001-04-11 2004-12-13 Method and apparatus for the removal of flash artifacts Abandoned US20050093994A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US11/011,407 US20050093994A1 (en) 2001-04-11 2004-12-13 Method and apparatus for the removal of flash artifacts

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US09/834,238 US6859565B2 (en) 2001-04-11 2001-04-11 Method and apparatus for the removal of flash artifacts
US11/011,407 US20050093994A1 (en) 2001-04-11 2004-12-13 Method and apparatus for the removal of flash artifacts

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US09/834,238 Division US6859565B2 (en) 2001-04-11 2001-04-11 Method and apparatus for the removal of flash artifacts

Publications (1)

Publication Number Publication Date
US20050093994A1 true US20050093994A1 (en) 2005-05-05

Family

ID=25266456

Family Applications (3)

Application Number Title Priority Date Filing Date
US09/834,238 Expired - Lifetime US6859565B2 (en) 2001-04-11 2001-04-11 Method and apparatus for the removal of flash artifacts
US11/011,407 Abandoned US20050093994A1 (en) 2001-04-11 2004-12-13 Method and apparatus for the removal of flash artifacts
US11/011,412 Expired - Lifetime US7027662B2 (en) 2001-04-11 2004-12-13 Method and apparatus for the removal of flash artifacts

Family Applications Before (1)

Application Number Title Priority Date Filing Date
US09/834,238 Expired - Lifetime US6859565B2 (en) 2001-04-11 2001-04-11 Method and apparatus for the removal of flash artifacts

Family Applications After (1)

Application Number Title Priority Date Filing Date
US11/011,412 Expired - Lifetime US7027662B2 (en) 2001-04-11 2004-12-13 Method and apparatus for the removal of flash artifacts

Country Status (2)

Country Link
US (3) US6859565B2 (en)
JP (1) JP4041687B2 (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20030174262A1 (en) * 2002-03-11 2003-09-18 Fujitsu Display Technologies Corporation Liquid crystal display and method of manufacturing the same
US20080149279A1 (en) * 2006-12-21 2008-06-26 Patricia Marlin Leighty Apparatus for displaying window covering samples
US20100080550A1 (en) * 2008-09-30 2010-04-01 Shinya Fujiwara Photography apparatus and photography method
US20130209057A1 (en) * 2012-02-09 2013-08-15 Motorola Solutions, Inc. Method and apparatus for capturing an image

Families Citing this family (135)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7630006B2 (en) 1997-10-09 2009-12-08 Fotonation Ireland Limited Detecting red eye filter and apparatus using meta-data
US7352394B1 (en) 1997-10-09 2008-04-01 Fotonation Vision Limited Image modification based on red-eye filter analysis
US7042505B1 (en) 1997-10-09 2006-05-09 Fotonation Ireland Ltd. Red-eye filter method and apparatus
US7738015B2 (en) 1997-10-09 2010-06-15 Fotonation Vision Limited Red-eye filter method and apparatus
JP3927353B2 (en) * 2000-06-15 2007-06-06 株式会社日立製作所 Image alignment method, comparison inspection method, and comparison inspection apparatus in comparison inspection
EP1288859A1 (en) * 2001-09-03 2003-03-05 Agfa-Gevaert AG Method for automatic detection of red-eye defecs in photographic images
JP4451583B2 (en) * 2001-12-27 2010-04-14 富士フイルム株式会社 Imaging apparatus, imaging method, and program
AU2003235586A1 (en) * 2002-01-16 2003-07-30 Andrei Feldman Oral implant template
US7239758B2 (en) * 2002-02-07 2007-07-03 Nikon Corporation Signal processing device for reducing noise of image signal, signal processing program, and signal processing method
DE10216806B4 (en) * 2002-04-16 2006-11-23 Institut Für Mikroelektronik Stuttgart Method and device for FPN correction of image signal values of an image sensor
US6944356B2 (en) * 2002-05-13 2005-09-13 Tektronix, Inc. Locating point of interest in an impaired image
JP5109221B2 (en) * 2002-06-27 2012-12-26 新世代株式会社 Information processing device equipped with an input system using a stroboscope
EP1688883B1 (en) * 2002-12-11 2009-04-01 FUJIFILM Corporation Image correction apparatus and image pickup apparatus
US7295720B2 (en) * 2003-03-19 2007-11-13 Mitsubishi Electric Research Laboratories Non-photorealistic camera
US7206449B2 (en) * 2003-03-19 2007-04-17 Mitsubishi Electric Research Laboratories, Inc. Detecting silhouette edges in images
US7738725B2 (en) * 2003-03-19 2010-06-15 Mitsubishi Electric Research Laboratories, Inc. Stylized rendering using a multi-flash camera
JP2004320128A (en) * 2003-04-11 2004-11-11 Mega Chips Corp Defective pixel correction device
JP4136793B2 (en) * 2003-05-29 2008-08-20 キヤノン株式会社 IMAGING DEVICE AND IMAGING DEVICE CONTROL METHOD
US7352909B2 (en) * 2003-06-02 2008-04-01 Seiko Epson Corporation Weighted overcomplete de-noising
US8170294B2 (en) * 2006-11-10 2012-05-01 DigitalOptics Corporation Europe Limited Method of detecting redeye in a digital image
US8896725B2 (en) 2007-06-21 2014-11-25 Fotonation Limited Image capture device with contemporaneous reference image capture mechanism
US7970182B2 (en) 2005-11-18 2011-06-28 Tessera Technologies Ireland Limited Two stage detection for photographic eye artifacts
US7587085B2 (en) * 2004-10-28 2009-09-08 Fotonation Vision Limited Method and apparatus for red-eye detection in an acquired digital image
US9129381B2 (en) * 2003-06-26 2015-09-08 Fotonation Limited Modification of post-viewing parameters for digital images using image region or feature information
US7920723B2 (en) * 2005-11-18 2011-04-05 Tessera Technologies Ireland Limited Two stage detection for photographic eye artifacts
US7689009B2 (en) * 2005-11-18 2010-03-30 Fotonation Vision Ltd. Two stage detection for photographic eye artifacts
US8036458B2 (en) * 2007-11-08 2011-10-11 DigitalOptics Corporation Europe Limited Detecting redeye defects in digital images
US8948468B2 (en) 2003-06-26 2015-02-03 Fotonation Limited Modification of viewing parameters for digital images using face detection information
US8330831B2 (en) * 2003-08-05 2012-12-11 DigitalOptics Corporation Europe Limited Method of gathering visual meta data using a reference image
US7574016B2 (en) 2003-06-26 2009-08-11 Fotonation Vision Limited Digital image processing using face detection information
US8254674B2 (en) * 2004-10-28 2012-08-28 DigitalOptics Corporation Europe Limited Analyzing partial face regions for red-eye detection in acquired digital images
US8682097B2 (en) 2006-02-14 2014-03-25 DigitalOptics Corporation Europe Limited Digital image enhancement with reference images
US8494286B2 (en) * 2008-02-05 2013-07-23 DigitalOptics Corporation Europe Limited Face detection in mid-shot digital images
US7792970B2 (en) 2005-06-17 2010-09-07 Fotonation Vision Limited Method for establishing a paired connection between media devices
US9692964B2 (en) 2003-06-26 2017-06-27 Fotonation Limited Modification of post-viewing parameters for digital images using image region or feature information
US7536036B2 (en) 2004-10-28 2009-05-19 Fotonation Vision Limited Method and apparatus for red-eye detection in an acquired digital image
US20050057663A1 (en) * 2003-07-16 2005-03-17 Thomas Graham Alexander Video processing
US20050031224A1 (en) * 2003-08-05 2005-02-10 Yury Prilutsky Detecting red eye filter and apparatus using meta-data
US20050140801A1 (en) * 2003-08-05 2005-06-30 Yury Prilutsky Optimized performance and performance for red-eye filter method and apparatus
US9412007B2 (en) * 2003-08-05 2016-08-09 Fotonation Limited Partial face detector red-eye filter method and apparatus
US8520093B2 (en) * 2003-08-05 2013-08-27 DigitalOptics Corporation Europe Limited Face tracker and partial face tracker for red-eye filter method and apparatus
WO2005024728A1 (en) * 2003-09-03 2005-03-17 Nec Corporation Form changing device, object action encoding device, and object action decoding device
US7835572B2 (en) * 2003-09-30 2010-11-16 Sharp Laboratories Of America, Inc. Red eye reduction technique
JP4610182B2 (en) * 2003-12-05 2011-01-12 株式会社日立ハイテクノロジーズ Scanning electron microscope
JP4464143B2 (en) * 2004-01-07 2010-05-19 キヤノン株式会社 Image processing method and apparatus
JP4379129B2 (en) * 2004-01-23 2009-12-09 ソニー株式会社 Image processing method, image processing apparatus, and computer program
US20110102643A1 (en) * 2004-02-04 2011-05-05 Tessera Technologies Ireland Limited Partial Face Detector Red-Eye Filter Method and Apparatus
DE102004016564A1 (en) * 2004-03-30 2006-02-16 Frie, Werner Rudolf, Dr. Flashlight reflection compensation system for camera with one flashlight involves taking two photographs with different intensities of flash and identifying and erasing bright reflections from flash
JP4610411B2 (en) * 2004-05-17 2011-01-12 ミツビシ・エレクトリック・リサーチ・ラボラトリーズ・インコーポレイテッド Method for generating a stylized image of a scene containing objects
JP4078334B2 (en) * 2004-06-14 2008-04-23 キヤノン株式会社 Image processing apparatus and image processing method
US20050276508A1 (en) * 2004-06-15 2005-12-15 Lockheed Martin Corporation Methods and systems for reducing optical noise
US7457477B2 (en) * 2004-07-06 2008-11-25 Microsoft Corporation Digital photography with flash/no flash extension
US7551797B2 (en) * 2004-08-05 2009-06-23 Canon Kabushiki Kaisha White balance adjustment
GB0419882D0 (en) * 2004-09-08 2004-10-13 Bamford Excavators Ltd Calculation module
US7403654B2 (en) * 2004-10-04 2008-07-22 Arcsoft, Inc. Enhanced automatic red eye removal
US8320641B2 (en) * 2004-10-28 2012-11-27 DigitalOptics Corporation Europe Limited Method and apparatus for red-eye detection using preview or other reference images
KR20060038681A (en) * 2004-11-01 2006-05-04 삼성테크윈 주식회사 Apparatus and method for removing hot pixel in digital camera
CN100517388C (en) * 2004-11-08 2009-07-22 松下电器产业株式会社 Normal information estimator, registration image group forming device, image collator and normal information estimating method
US20060104537A1 (en) * 2004-11-12 2006-05-18 Sozotek, Inc. System and method for image enhancement
DE102004062382A1 (en) * 2004-12-23 2006-07-13 Siemens Ag Image coding method and picture coding device
US7652717B2 (en) * 2005-01-11 2010-01-26 Eastman Kodak Company White balance correction in digital camera images
US7623732B1 (en) * 2005-04-26 2009-11-24 Mercury Computer Systems, Inc. Method and apparatus for digital image filtering with discrete filter kernels using graphics hardware
US7450756B2 (en) * 2005-04-28 2008-11-11 Hewlett-Packard Development Company, L.P. Method and apparatus for incorporating iris color in red-eye correction
US7443443B2 (en) * 2005-07-28 2008-10-28 Mitsubishi Electric Research Laboratories, Inc. Method and apparatus for enhancing flash and ambient images
JP4626436B2 (en) * 2005-07-29 2011-02-09 株式会社ニコン Electronic camera
JP4635779B2 (en) * 2005-08-18 2011-02-23 ソニー株式会社 Data processing apparatus, data processing method, and program
JP2007102189A (en) * 2005-09-07 2007-04-19 Ricoh Co Ltd Color shift correcting apparatus and method, image forming apparatus, color shift correcting program and recording medium
US8260008B2 (en) 2005-11-11 2012-09-04 Eyelock, Inc. Methods for performing biometric recognition of a human eye and corroboration of same
US7599577B2 (en) * 2005-11-18 2009-10-06 Fotonation Vision Limited Method and apparatus of correcting hybrid flash artifacts in digital images
WO2007095553A2 (en) 2006-02-14 2007-08-23 Fotonation Vision Limited Automatic detection and correction of non-red eye flash defects
US7796168B1 (en) * 2006-06-09 2010-09-14 Flir Systems, Inc. Methods and systems for detection and mitigation of image-flash in infrared cameras
DE602007012246D1 (en) 2006-06-12 2011-03-10 Tessera Tech Ireland Ltd PROGRESS IN EXTENDING THE AAM TECHNIQUES FROM GRAY CALENDAR TO COLOR PICTURES
WO2008102205A2 (en) * 2006-08-09 2008-08-28 Fotonation Vision Limited Detection of airborne flash artifacts using preflash image
JP5073996B2 (en) * 2006-09-20 2012-11-14 オリンパス株式会社 Image processing device
US8320702B2 (en) * 2006-09-28 2012-11-27 Jadak Technologies, Inc. System and method for reducing specular reflection
JP4747065B2 (en) * 2006-09-29 2011-08-10 富士通株式会社 Image generation apparatus, image generation method, and image generation program
US7908601B2 (en) * 2006-11-01 2011-03-15 International Business Machines Corporation Computer method and apparatus for merging software configurations when a change set spans multiple artifacts
JP4987442B2 (en) * 2006-11-21 2012-07-25 浜松ホトニクス株式会社 X-ray imaging method and X-ray imaging system
CN101563913B (en) * 2006-12-22 2011-10-19 诺基亚公司 Removal of artifacts in flash images
US8055067B2 (en) 2007-01-18 2011-11-08 DigitalOptics Corporation Europe Limited Color segmentation
JP4932504B2 (en) * 2007-01-18 2012-05-16 富士フイルム株式会社 Image processing method, apparatus, program, and imaging apparatus
US20080199073A1 (en) * 2007-02-20 2008-08-21 Microsoft Corporation Red eye detection in digital images
JP2010520567A (en) * 2007-03-05 2010-06-10 フォトネーション ビジョン リミテッド Red-eye false detection filtering using face position and orientation
JP4986747B2 (en) * 2007-07-09 2012-07-25 キヤノン株式会社 Imaging apparatus and imaging method
US7873236B2 (en) * 2007-08-28 2011-01-18 General Electric Company Systems, methods and apparatus for consistency-constrained filtered backprojection for out-of-focus artifacts in digital tomosythesis
JP2009077230A (en) * 2007-09-21 2009-04-09 Seiko Epson Corp Image processor, micro computer and electronic equipment
US8503818B2 (en) 2007-09-25 2013-08-06 DigitalOptics Corporation Europe Limited Eye defect detection in international standards organization images
US8018525B2 (en) 2007-12-21 2011-09-13 Nokia Corporation Camera flash module and method for controlling same
US8212864B2 (en) * 2008-01-30 2012-07-03 DigitalOptics Corporation Europe Limited Methods and apparatuses for using image acquisition data to detect and correct image defects
US8050493B2 (en) * 2008-03-31 2011-11-01 Konica Minolta Laboratory U.S.A., Inc. Method for generating a high quality scanned image of a document
US20090278921A1 (en) * 2008-05-12 2009-11-12 Capso Vision, Inc. Image Stabilization of Video Play Back
US8081254B2 (en) 2008-08-14 2011-12-20 DigitalOptics Corporation Europe Limited In-camera based method of detecting defect eye with high accuracy
TWI394443B (en) * 2008-09-19 2013-04-21 Altek Corp Image brightness correction method
TWI472222B (en) * 2008-09-19 2015-02-01 Altek Corp Image brightness compensation method and digital image pickup device with image brightness compensation
US8103120B2 (en) * 2008-09-22 2012-01-24 Solomon Systech Limited Method and apparatus of local contrast enhancement
JP2010074693A (en) * 2008-09-22 2010-04-02 Fujifilm Corp Digital camera, corrected image generating method and storage medium
US8295593B2 (en) * 2009-01-07 2012-10-23 Seiko Epson Corporation Method of detecting red-eye objects in digital images using color, structural, and geometric characteristics
JP5410510B2 (en) * 2009-04-16 2014-02-05 パナソニック株式会社 Imaging device, external flash detection method, program, and integrated circuit
FR2944884B1 (en) * 2009-04-27 2011-07-01 St Ericsson Sa St Ericsson Ltd METHOD OF CORRECTING A REPRESENTATION OF A CENTRAL REGION OF AN EYE ON A PHOTOGRAPHIC IMAGE OF SAID EYE WITH A FLASH.
JP5368960B2 (en) * 2009-12-09 2013-12-18 三星電子株式会社 Imaging apparatus and imaging method
US20110134283A1 (en) * 2009-12-09 2011-06-09 Samsung Electronics Co., Ltd. Photographing apparatus and photographing method
JP2011139115A (en) * 2009-12-25 2011-07-14 Saitama Univ High-speed camera equipment and image processing method for the same
US8339508B2 (en) * 2010-02-22 2012-12-25 Csr Technology Inc. Method and apparatus for low-light imaging enhancement
JP5141733B2 (en) * 2010-08-18 2013-02-13 カシオ計算機株式会社 Imaging apparatus, imaging method, and program
US8570433B1 (en) * 2010-08-25 2013-10-29 CSR Technology, Inc. Coloration artifact reduction
US8773577B2 (en) * 2010-10-27 2014-07-08 Qualcomm Incorporated Region of interest extraction
KR101317552B1 (en) * 2010-11-04 2013-10-16 가시오게산키 가부시키가이샤 Image capturing apparatus, method and storage medium capable of adjusting white balance
JP5056938B2 (en) * 2010-12-22 2012-10-24 パナソニック株式会社 Imaging apparatus and image data forming method
US8571271B2 (en) 2011-05-26 2013-10-29 Microsoft Corporation Dual-phase red eye correction
US8493459B2 (en) 2011-09-15 2013-07-23 DigitalOptics Corporation Europe Limited Registration of distorted images
US8493460B2 (en) 2011-09-15 2013-07-23 DigitalOptics Corporation Europe Limited Registration of differently scaled images
US8884553B2 (en) * 2011-10-19 2014-11-11 Justin Hai Current monitor for indicating condition of attached electrical apparatus
CN103150717B (en) * 2011-12-06 2016-02-10 株式会社理光 The detection method of image highlight area, content detection algorithm and content detection device
US9354748B2 (en) 2012-02-13 2016-05-31 Microsoft Technology Licensing, Llc Optical stylus interaction
US9075566B2 (en) 2012-03-02 2015-07-07 Microsoft Technoogy Licensing, LLC Flexible hinge spine
US9870066B2 (en) 2012-03-02 2018-01-16 Microsoft Technology Licensing, Llc Method of manufacturing an input device
US9158383B2 (en) 2012-03-02 2015-10-13 Microsoft Technology Licensing, Llc Force concentrator
TWI489861B (en) 2012-03-26 2015-06-21 Wistron Corp Method for checking camera and camera
US20130300590A1 (en) 2012-05-14 2013-11-14 Paul Henry Dietz Audio Feedback
US8964379B2 (en) 2012-08-20 2015-02-24 Microsoft Corporation Switchable magnetic lock
US8786767B2 (en) 2012-11-02 2014-07-22 Microsoft Corporation Rapid synchronized lighting and shuttering
US9536345B2 (en) * 2012-12-26 2017-01-03 Intel Corporation Apparatus for enhancement of 3-D images using depth mapping and light source synthesis
JP6351271B2 (en) * 2014-01-17 2018-07-04 オリンパス株式会社 Image composition apparatus, image composition method, and program
US10120420B2 (en) 2014-03-21 2018-11-06 Microsoft Technology Licensing, Llc Lockable display and techniques enabling use of lockable displays
US10324733B2 (en) 2014-07-30 2019-06-18 Microsoft Technology Licensing, Llc Shutdown notifications
FR3030974B1 (en) * 2014-12-19 2018-03-23 Stereolabs SYSTEM FOR SHOOTING THREE DIMENSIONS DURING DISPLACEMENT
US9747520B2 (en) * 2015-03-16 2017-08-29 Kla-Tencor Corporation Systems and methods for enhancing inspection sensitivity of an inspection tool
US9734433B1 (en) * 2015-03-25 2017-08-15 Amazon Technologies, Inc. Estimating photometric properties using a camera and a display screen
WO2016161481A1 (en) * 2015-04-08 2016-10-13 Wavefront Biometric Technologies Pty Limited Multi-biometric authentication
CN105959543B (en) * 2016-05-19 2019-04-30 努比亚技术有限公司 It is a kind of to remove reflective filming apparatus and method
US10981376B2 (en) * 2018-07-03 2021-04-20 Ricoh Company, Ltd. Inspection apparatus, image reading apparatus, image forming apparatus, inspection method, and recording medium
JP7247501B2 (en) * 2018-09-21 2023-03-29 富士フイルムビジネスイノベーション株式会社 Image processing device and program
CN112381775B (en) * 2020-11-06 2023-02-21 厦门市美亚柏科信息股份有限公司 Image tampering detection method, terminal device and storage medium
CN113362246A (en) * 2021-06-08 2021-09-07 百果园技术(新加坡)有限公司 Image banding artifact removing method, device, equipment and medium
CN115375689B (en) * 2022-10-25 2023-07-07 深圳华付技术股份有限公司 Tobacco shred barrel detection method, device, equipment and medium based on machine vision

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5748764A (en) * 1993-07-19 1998-05-05 Eastman Kodak Company Automated detection and correction of eye color defects due to flash illumination
US5782769A (en) * 1996-06-28 1998-07-21 Advanced Technology Laboratories, Inc. Ultrasonic diagnostic image flash suppression technique
US6285410B1 (en) * 1998-09-11 2001-09-04 Mgi Software Corporation Method and system for removal of flash artifacts from digital images
US6402694B1 (en) * 1999-05-06 2002-06-11 Medison Co., Ltd. Color imaging display method and apparatus for color doppler imaging system
US6728401B1 (en) * 2000-08-17 2004-04-27 Viewahead Technology Red-eye removal using color image processing

Family Cites Families (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4285588A (en) * 1980-07-24 1981-08-25 Eastman Kodak Company Apparatus and method for minimizing red-eye in flash photography
JPH07301842A (en) * 1994-05-02 1995-11-14 Asahi Optical Co Ltd Electronic still camera
US5568194A (en) * 1994-05-31 1996-10-22 Asahi Kogaku Kogyo Kabushiki Kaisha Adjusting a white balance in a camera using a flash
US6016354A (en) 1997-10-23 2000-01-18 Hewlett-Packard Company Apparatus and a method for reducing red-eye in a digital image
US6631208B1 (en) * 1998-05-29 2003-10-07 Fuji Photo Film Co., Ltd. Image processing method
JP3961734B2 (en) * 1999-12-24 2007-08-22 富士フイルム株式会社 Image processing method, apparatus, and recording medium
US6718051B1 (en) * 2000-10-16 2004-04-06 Xerox Corporation Red-eye detection method

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5748764A (en) * 1993-07-19 1998-05-05 Eastman Kodak Company Automated detection and correction of eye color defects due to flash illumination
US5782769A (en) * 1996-06-28 1998-07-21 Advanced Technology Laboratories, Inc. Ultrasonic diagnostic image flash suppression technique
US6285410B1 (en) * 1998-09-11 2001-09-04 Mgi Software Corporation Method and system for removal of flash artifacts from digital images
US6402694B1 (en) * 1999-05-06 2002-06-11 Medison Co., Ltd. Color imaging display method and apparatus for color doppler imaging system
US6728401B1 (en) * 2000-08-17 2004-04-27 Viewahead Technology Red-eye removal using color image processing

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20030174262A1 (en) * 2002-03-11 2003-09-18 Fujitsu Display Technologies Corporation Liquid crystal display and method of manufacturing the same
US20080068331A1 (en) * 2002-03-11 2008-03-20 Sharp Kabushiki Kaisha Liquid crystal display and method of manufacturing the same
US20080149279A1 (en) * 2006-12-21 2008-06-26 Patricia Marlin Leighty Apparatus for displaying window covering samples
US20100080550A1 (en) * 2008-09-30 2010-04-01 Shinya Fujiwara Photography apparatus and photography method
US7873271B2 (en) * 2008-09-30 2011-01-18 Fujifilm Corporation Photography apparatus and photography method
US20130209057A1 (en) * 2012-02-09 2013-08-15 Motorola Solutions, Inc. Method and apparatus for capturing an image

Also Published As

Publication number Publication date
US7027662B2 (en) 2006-04-11
US20050094895A1 (en) 2005-05-05
JP4041687B2 (en) 2008-01-30
US6859565B2 (en) 2005-02-22
JP2003008988A (en) 2003-01-10
US20020150306A1 (en) 2002-10-17

Similar Documents

Publication Publication Date Title
US7027662B2 (en) Method and apparatus for the removal of flash artifacts
CN107277356B (en) Method and device for processing human face area of backlight scene
US9852499B2 (en) Automatic selection of optimum algorithms for high dynamic range image processing based on scene classification
JP3668014B2 (en) Image processing method and apparatus
US9019402B2 (en) Dynamic range extension by combining differently exposed hand-held device-acquired images
US8488015B2 (en) Camera using preview image to select exposure
EP2426928B1 (en) Image processing apparatus, image processing method and program
JP5743696B2 (en) Image processing apparatus, image processing method, and program
JP6049343B2 (en) Image processing apparatus, image processing method, and program
JP5665436B2 (en) Image processing apparatus, image processing method, and program
JP5808142B2 (en) Image processing apparatus, image processing method, and program
JP5693271B2 (en) Image processing apparatus and method
JP6685188B2 (en) Imaging device, image processing device, control method thereof, and program
JP2015156615A (en) Image processing system, image processing method, control program, and recording medium
US20180025476A1 (en) Apparatus and method for processing image, and storage medium
CN110324529B (en) Image processing apparatus and control method thereof
JP2003234950A (en) Image processing device, image processing program, and image processing method
JP7114335B2 (en) IMAGE PROCESSING DEVICE, CONTROL METHOD FOR IMAGE PROCESSING DEVICE, AND PROGRAM
JP2003235050A (en) Image processing apparatus, image processing program, and image processing method
JP2001054014A (en) Electronic camera
JP2003199115A (en) Method for contrast enhancement in color digital image
JP2003209856A (en) White balance correction method
JP5854716B2 (en) Image processing apparatus, image processing method, and program
JP4857856B2 (en) Electronic camera having saturation adjustment function and image processing program
JPH11196324A (en) Method and device for outputting image

Legal Events

Date Code Title Description
STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO PAY ISSUE FEE