throbber
(19) )
`
`Europaisches Patentamt
`
`European Patent Office
`
`Office européen des brevets
`
`(11)
`
`EP 0 858 208 A1
`
`(12)
`
`EUROPEANPATENT APPLICATION
`
`(43) Date of publication:
`12.08.1998 Bulletin 1998/33
`
`(21) Application number: 98200205.7
`
`(22) Date offiling: 26.01.1998
`
`(84) Designated Contracting States:
`AT BE CH DE DK ES FI FR GB GRIEITLILU MC
`NL PT SE
`
`Designated Extension States:
`AL LT LV MK RO Sl
`
`(30) Priority: 07.02.1997 US 796350
`
`(71) Applicant: EASTMAN KODAK COMPANY
`Rochester, New York 14650 (US)
`
`(51) Int cl.6: HO4N 1/21, HO4N 1/409
`
`« Loveridge, Jennifer Clara,
`Eastman Kodak Company
`Rochester, New York 14650-2201 (US)
`¢ Mcintyre, Dale F., Eastman Kodak Company
`Rochester, New York 14650-2201 (US)
`
`(74) Representative: Buff, Michelet al
`KODAK INDUSTRIE
`
`Département Brevets - CRT
`Zone Industrielle - B.P. 21
`
`71102 Chalon sur Sa6ne Cédex (FR)
`
`
`
`(72) Inventors:
`¢ Weldy, John Allan, Eastman Kodak Company
`Rochester, New York 14650-2201 (US)
`
`(54)
`
`Methodof producing digital images with improved performance characteristic
`
`(57)©Amethodof producing a digital image with im-
`proved performance characteristics includes capturing
`at least two electronic images of a scene and digitizing
`the at least two electronic images of a scene. The meth-
`od further includes combining and processing the at
`least two digitized electronic images of the scene to pro-
`duce a combineddigital image of a scene with improved
`performance characteristics.
`
`ELECTRONIC IMAGE
`
`
`AT LEAST ONE
`MORE DIGITIZED
`ELECTRONIC IMAGE
`
`
`
`
`
`(estestocoumco.se|po; po[PeerFoamonaaneon: po[GaaracomeConsoneHIGHPASS FILTER a
`
`
`
`
`(covert70cominwanesoma
` 120 -
`
`
`
`
`
`
`EP0858208A1
`
`124 :
`
`
`
`:
`
`COMBINE IMAGES
`
`
`
`DIGITAL HARD
`COPY OUTPUT
`PRINTER
`
`82
`
`FIG. 3
`
`SAMS-1007
`SamsungElectronics Co. v. Yu etal.
`IPR2020-00492
`Page 00001
`
`Printed by Jouve, 75001 PARIS (FR)
`
`SAMS-1007
`Samsung Electronics Co. v. Yu et al.
`IPR2020-00492
`Page 00001
`
`

`

`Description
`
`EP 0 858 208 A1
`
`The present invention relates to producing improved digital images which can, for example, be produced from
`cameras that capture more than one record of a scene.
`Electronic image capture was spawnedbythe television industry over 50 years ago. Over the years, there have
`been advancesfrom black and white (B&W)only image captureto full color capture. Coupled with more recent advances
`in video recording technology, the camcorder (and more recently, the digital camcorder) has become commonplace.
`In the 1980s, SONY Corporation introduced the electronic still camera for capturing still images and storing them in
`an analog manner on a magnetic disk. More recently, the large reduction in digital memory costs has led to many
`manufacturers introducing digital still cameras. These cameras typically feature a full-frame or interline charge coupled
`device (CCD) which is used to capture full color scene information.
`Despite the rapid decrease in memory and CCD costs, image performance problems remain. One problem is the
`fact that when a single CCD is used, a captured color image has reduced spatial resolution compared to an otherwise
`equivalent B&W image. This is a consequence of the need for a color filter array (CFA) to enable the CCD to capture
`the color information from (slightly) different locations in the scene. Typically three color records will be captured si-
`multaneously, each sparsely sampled and pixel interleaved with respect to the pixel structure as defined by the CCD.
`Subsequent analogue or digital image processing, typically by spatial interpolation to upsample and rephase the color
`records,
`is required to produce a full color record corresponding to every location in the scene. This color sparse-
`sampling and subsequent image processing can lead to color artifacts. Alternatively, the scene can be low-passed
`filtered prior to being sensed by the CCD through the CFA: however, this low-passfiltering further lowers the spatial
`resolution of the scene.
`
`An alternative method for generating colorutilizes separate sensors for each color channel (typically three sensors,
`one eachfor capturing red, green, and blue information). While this solves the spatial resolution problem, these systems
`typically use a single lens followed by a beam splitter and aligning optics. There are light losses (and therefore logs in
`the light sensitivity of the system) with beam splitters and the aligning optics oftentimes requires adjustment owing to
`ternperature effects. In addition the beam splitter adds to the physical bulk of the camera.
`Another problem encountered with CCD sensorsis that the process used to manufacture these sensors and other
`silicon based devices such as digital memory yields, though small in number, defects. A defect that results in only a
`few single clement picture clements (pixels) being lost from an image is typically acceptable; however, those defects
`resulting in line (a complete row or column of pixels) defects are not acceptable. Therefore the yield of acceptable
`CCDs from these manufacturing processes is oftentimes low. As the probability of defects is typically proportional to
`the light sensitive area of the CCD, making smaller CCDs can increase the yield. The smaller light sensitive area of
`smaller CCDs also has an advantage in that shorter focal length lenses (and therefore potentially thinner cameras)
`are required to produce the same field of view as when captured on a device with a larger light-sensitive area. Unfor-
`tunately, in order to maintain spatial resolution of smaller CCDs, smaller pixel sizes are required. Smaller pixel sizes
`can suffer from increased noise owing to the smaller photon detection area, and the lower number of electrons (gen-
`erated from photons) that can be stored before reaching pixel saturation. To overcome this, larger area sensors can
`be made, but the above-mentioned problems with yield are exacerbated, resulting in high cost for larger area sensors.
`It is an object of the present invention to overcome the above mentioned problems and provide a method of pro-
`ducing a digital image with improved performance characteristics, comprising the stepsof:
`
`a) capturing at least two electronic images of a scene;
`b) digitizing the at least two electronic images of a scene; and
`¢) combining and processing the at least two digitized electronic images of the scene to produce a combined digital
`image of a scene with improved performance characteristics.
`
`The present invention overcomes the above-mentioned problems associated with electronic capture of scenes.
`By capturing at least two electronic images of a scene. it is possible to overcome the spatial resolution and noise
`problems associated with small area electronic sensors. Further improved characteristics can be obtained by capturing
`the at least two electronic images through a corresponding number of separate lenses (each pointing at the scene),
`thus overcoming the above-mentioned problems that can result when beam splitter optics are employed.
`
`FIG. 1a is a perspective view of a dual lens camera in accordance with the present invention;
`FIG. 1b and FIG. 1¢ are perspective views of multilens cameras with four and eight lenses respectively;
`FIG. 2 is a block diagram showing a central processing unit that can take the stored digital data frorn the dual lens
`camera and process the images in accordance with the present invention; and
`FIG. 3 is a flow chart showing in block diagram form the steps needed for the combining and image processing to
`produce the digital image with improved performancecharacteristics.
`
`10
`
`15
`
`20
`
`25
`
`30
`
`35
`
`40
`
`4&
`
`50
`
`55
`
`SAMS-1007 Page 00002
`
`SAMS-1007 Page 00002
`
`

`

`EP 0 858 208 Ai
`
`By the use of the term dual lens camera, is meantthat there are at least two separate images formed with their
`own unique, non-optically linked lens systems. For clarity, much of the following description details examples wherein
`only two images are captured and combined; however it will be evident to those skilled in the art that greater than two
`images could be captured and combined, as taught below,to provide digital images with improved performance char-
`acteristics.
`
`Turning to FIG. 1a, there is shown a dual lens electronic camera having a main camera body 1; imaging lenses
`3a and 3b; focusing mechanisms 6, a shutter switch including a self-return push-button switch 7; and a shutter speed
`dial 9. These components are the same as in a standard electronic camera.
`Turning now to FIG. 2, a block diagram is shown having a means to extract the digitized electronic images from
`the temporary digital storage within camera 76 and transfer these imagesto the central processing unit 78. Stored and
`transferred with the digitized scene information is identification information that allows the central processing unit 78
`to collate all of the images for the same scene when performing the subsequentdigital image processing. After the
`central processing unit has processed the images in accordance with the present invention, it produces output signals
`to a digital hard copy outputprinter 82, or alternatively displays the processed images on a cathode ray tube (CRT) or
`utilizes these processed imagesin other subsequent steps (not shown), such as image transmission, alternate digital
`storage, or the like. Digital hard copy output printers, for example ink jet printers and thermal printers, are well known
`in the art.
`
`Turning now to FIG. 3 which is a block diagram showing the algorithm for producing a combined digital image
`having improved characteristics; the two digitized electronic images are applied to the central processing unit 78. This
`block diagram and the corresponding description of the algorithm relate to two digitized electronic images, however it
`will be understood that the method can be extended to apply to more than two digitized electronic images. Although
`this algorithm is embodied in the central processing unit 78, it will be well understood that the algorithm can be stored
`on a computer program product such as, for example, magnetic storage media, such as magnetic disks (floppy disk)
`or magnetic tapes; optical storage media such as optical disks, optical tape, or machine readable barcode: solid state
`devices such as random access memory (RAM) or read only memory (ROM).
`The first step is to convert the at least two digitized electronic images to a commoncolor space. Although the
`images are digital representations of the same scene, the at least two electronic sensors need not have the same
`spectral sensitivities, and in this case, the color information will be apportioned differently among the color channels
`in the at least two electronic sensors. The images should be transformed into a commoncolor space with a common
`tone-scale, by means of color matrices and look-up tables (see, for example, W.K. Pratt, Digital Image Processing, pp
`50-93, Wiley Interscience 1978), or by means of 3-D LUTs, techniques which are well knownto thoseskilled in the art.
`In block 120 the number of pixels in each of color converted images must be substantially matched in number of
`pixels. In other words, there should be a common number ofpixels. It will be understood that the at least two electronic
`sensors need not have the same number of pixels, and in this case, the image with the lower number of pixels is
`upsampled by the method of, for example, bi-cubic interpolation to match the number of pixels of the other image.
`Other typesof interpolation techniques can also be used to upsample digitized images, such as spatial functionfitting,
`convolution, and Fourier domain filtering. These are well knownin the art, and described, for example, in W.K. Pratt,
`pp 113-116.
`In block 122, corrections are now made to the digitized electronic imagesto correctfor any difference in their global
`geometry, that is any geometrical transformation which, when applied to every pixel in one image, enables its geometry
`to be substantially mapped onto the geometry of the other. Examples of such transformations are translation, rotation,
`scaling, aspect ratio, and the geometrical differences between the lens systems that are used for the at least two
`electronic sensors. It will be understood that this correction need be applied to all but one of the digitized electronic
`images to enable them to be mapped onto the geomeiry of one of the digitized electronic images that is not to be
`corrected for geometry (referred to hereafter as the default image). Since there may be some slight loss in image
`quality, sharpness in particular, associated with the application of this correction, the correction would normally be
`applied to the digitized electronic imagesoriginally had fewer pixels at block 120.
`The correction will typically involve three steps, which are described in terms of correcting the geometry of one of
`the digitized electronic images to the geometry of the default digitized electronic image. First is the generation of a set
`of displacement vectors, typically with sub-pixel accuracy, which characterize a local x,y displacement between the
`two imagesat certain locations in the image-pair. A variety of techniques may be suitable, including block matching,
`and the methodofdifferentials, both well knownin the art (Image Processing, edited by D. Pearson, Chapter 3, "Motion
`and Motion Estimation," G. Thomas, pp 40-57, McGraw-Hill, 1991), but the preferred technique for this application is
`phasecorrelation. For amore complete disclosure of phase correlation techniques, see Pearsonibid. Phase correlation
`provides a method to generate displacement vectors which is robust in the presenceof noise and brightness changes
`in the record of the scene. The second step is the interpretation of that set of displacement vectors as a generalized
`geometrical transformation. Three commonly occurring transformations are described here, translation, magnification
`(Zoom) and rotation in the plane of the image, butit will be understood that a similar process can be used tointerpret
`
`10
`
`15
`
`20
`
`25
`
`30
`
`35
`
`40
`
`45
`
`50
`
`55
`
`SAMS-1007 Page 00003
`
`SAMS-1007 Page 00003
`
`

`

`EP 0 858 208 Ai
`
`all geometrical transformations of rigid bodies. In the case of translation, the transformation is defined, simply, as that
`x,y displacement which occurs mostfrequently in the set of displacement vectors. Otherwise, if two independenttrans-
`lation vectors are available from the vector set which map positions (x,,¥;) and (X2,yz2) in one image onto (x;,y;') and
`(x2'y2)) respectively in the second image, then the following transformation may be defined:
`Magnification, by a factor m (m#71), about an origin at (a, b), is defined as
`
`10
`
`15
`
`20
`
`25
`
`30
`
`35
`
`40
`
`45
`
`50
`
`55
`
`X;-a
`yb
`
`+
`
`a
`b
`
`=
`
`x';
`yi
`
`Hence, substituting the pair of translation vectors and rearranging yields
`
`
`Xp - X,'
`Xp'- MX,
`Vo'- MY.
`m=2— a= 2
`b-2
`2
`Xo > Xy
`(1-m)
`(1-m)
`
`The generalized equation for a rotation in the x,y plane about a center at (a, b) through an angle 6 is
`
`sme
`
`sma)
`
`be} +=
`
`which can be solvedin a similar fashion by substituting in the translation vectors to yield a, b, , and so on for the other
`transformations. The third step is the geometrical transformation of one of the pair of images according to the param-
`eters calculated and the transformation equations given in step 2. Typically, this is achieved using phase-shifting spatial
`interpolation, similar to the interpolation techniques referenced above.
`In the simplest implementation of this algorithm, the images are now combined by, for example, a numerical or
`geometric average on a pixel-wise basis, as shown in block 128.
`However, oftentimes local areas in the above mentioned simple combination suffer from poor quality resulting from,
`for example, differences in the local geometry among the images. A simple solution to this problem is to detect these
`local areas, and to change the way in which the images are combinedin these local areas. Specifically, where the
`difference between the default image and one of the digitized electronic images measured at each pixel location, is
`within a specified tolerance (depending on the inherent noise characteristics of the electronic recording medium), the
`pixel values of the two images are averaged. Where the difference between the same location pixel-values for the
`digitized electronic images are not within this tolerance, the pixel value from one of the images is chosento provide a
`value at that position for the resultant image. The choice of which image to use is made based on measuresoflocal
`noise or local sharpness, dependent onthe application, but one imageis used consistently either for any pair of images,
`or for each local area within the pair of images. This procedure is repeated for eachof the at least one digitized electronic
`imagesthat are not the default image, resulting in a varying number of images being averaged at any pixel location.
`An approach to provide improved image quality in local areas where the above mentionedtolerance is exceeded
`is to convert to a commonlocal geometry, as shownin block 124. Again, it will be understood that these local corrections
`need be applied to all but one of the digitized electronic images images to enable it to be mapped onto the geometry
`of the default digitized electronic image. The measurementof these differences in local geometry is achieved by tech-
`niques similar to those used for the measurementof global displacements, and involves the assignmentof a displace-
`mentvector, obtained, for example by the phase correlation technique, to each pixel in the image to be corrected. The
`resultant array of assigned vectors, which maps the values of each pixel in one image onto the corresponding pixel
`positions in the default image, is generally referred to as a vectorfield. The method is well known and described, for
`example, in Image Processing, edited by D. Pearson, Chapter 3, "Motion and Motion Estimation," G. Thomas, pp 53-54,
`McGraw-Hill, 1991. The pair of images may now be combined byfirst correcting one image by spatial interpolation
`according to the vector field, then combining the two images by averaging, pixel-by-pixel, the same location pixel-
`values. Alternatively, it may be moreefficient to perform these two steps simultaneously by performing an averageof
`the two images on a pixel-by-pixel basis, but where the corresponding pixel positions are defined according to the
`
`SAMS-1007 Page 00004
`
`SAMS-1007 Page 00004
`
`

`

`EP 0 858 208 Ai
`
`10
`
`15
`
`25
`
`30
`
`35
`
`20
`
`vector field. It will be understood to those skilled in the art that although the above the description relates to the use
`of the vector field to correct for differences in local geometry between two images which havealready been converted
`toacommonglobal geometry, a vector field can be generated which can be used to measure and correctfor differences
`in both the global and local geometries simultaneously.
`In the case wherethe twodigitized images are from electronic sensors of different spatial resolutions, that is are
`represented by a different number of pixels, improved sharpness (spatial detail) and reduced noise can be achieved
`by applying a high-passfilter at block 126, to thal image which contains the greatest amountof high spatial frequency
`information. The high-passfilter, which may be designed using well known techniques (as described, for example, in
`Chapter 3 of "Theory and Application of Digital Signal Processing", Rabiner & Gold, Prentice-Hall 1975), should have
`a spatial-frequency response which correspondsto the differencein the effective spatial frequency responsesof the
`two sensors of different spatial resolutions. The digital images are combined in block 128 by adding the high-pass
`filtered digital image to the digital image which contains less information at high spatial frequencies, according to the
`considerations described in the previous paragraph.
`It will be further appreciated that the particular two lens/two sensor embodiment described above can be varied
`and modified to greater than two lens/two sensor systems wherein all the greater than two lens/sensor systems are
`used to simultaneously capture light information from the same scene. FIG. 1b shows a multi-lens camera with main
`body 50, four imaging lenses 51a, 51b, 51¢ and 51d, shutter switch 52 and focusing mechanism 16a. Similarly, FIG.
`1c shows a multi-lens camera with main body 102, eight imaging lenses 103a-h, and shutter switch 105. As more than
`one representation of the scene is captured, it is possible to select scene exposure conditions in a way that provides
`for improving various characteristics.
`One example arisesif the at least two electronic images are not at the same focus position. Most scenestypically
`have scene objects at different distances from the camera, and thereforeit is often the case, particularly with large
`apertures (small lens f-numbers), that only part of the scene is in focus. By utilizing image combination as described
`above, where the image with the best signal to noise is selected on a local basis, a combined image is produced in
`which more scene objects are sharper than in any of the input images. In this case, the "sharpest" of the imagesis
`defined as whichever image hasthe highestlocal signal to noise ratio, as determined, for example, by comparing the
`magnitude of the high-passfiltered image to the a priori noise level for the corresponding image capture means as
`measured over the same band of spatial frequencies, for each of the images. Alternatively, it is possible to divide each
`of the digitized images from electronic sensors, after conversion to a commonglobal and local geometry,into high and
`low frequency components by meansof high-pass and low-passfilters. The low frequency component componentsof
`the images are averaged and the high frequency component of the image with the best signal to noise (as defined
`above).
`A second example occurs whenthe electronic images do not have the same depth of field. The depth of field of
`an image capture system is a function of both lens focal length and aperture (f-number). It may be thatthe light sensitive
`areas of the two or more image sensors are different. In this case the smallest image sensor utilizes a shorter focal
`length lens (for the same angle of view of the scene) and, therefore, has the greatest the depth offield for a given f-
`number. An improved performance can be achieved which is similar to that obtained when the imagesare captured at
`different focus positions; this time the difference in local sharpness (improved signal-to-noise) having resulted from
`capture with different depth of field (or combination of depth of field and focus position). This variation can be particularly
`useful in situations where large size lens apertures are used, for example, in cases of low scene illumination or when
`very short exposure times are required, e.g., for capturing fast moving action without image blur.
`Further improvement in image performance can be achieved byutilizing alternate types of image combinations.
`For example, as described earlier, image sensors (CCD) and other electronic capture devices cannot sense and fully
`separate and form three independent color channels at a single capture site (CCD pixel). Given this limitation, CCDs
`and other electronic capture media typically have to apportionthe total light sensing elementsinto at leastthree different
`color sensitive types (e.g. red, green and blue; or cyan, yellow, green and white) in order to capture a color image.
`Image processing algorithms are utilized to reconstruct a full color image but, fundamentally there is a loss in spatial
`resolution in order to obtain a color image. Furthermore, when colorfilters are attached to the surface of the image
`sensor, they, by definition, limit the number of photons captured by the sensor to those photons of wavelength that are
`transmitted by the colorfilter. In consequence, it is possible to utilize single capture sites (CCD pixels) that are smaller
`in physical dimension (and therefore generate a higher resolution digitized electronic image from the same area of
`silicon) if no colorfilters are attached to the surface of the capture sites, since more photons per unit area will be
`captured than whenacolor filter is present. Hence, in situations where the camera is operating at the limits of light
`sensitivity, there is an advantagein utilizing a high-resolution (small capture site) monochrome (no color filter) image
`sensor, in conjunction with either a single lower-resolution sensor with a color filter array attached and larger pixels,
`or a set of lower resolution (larger pixels) color arrays, each sensitive to bands of wavelengths corresponding to a
`"single" color. Techniques of this type have been employedin the pastin relation to high-resolution CCD scanners of
`motion-picture film for the television industry (see, for example, the U.S. Patent assigned to R.A. Sharman and R.T.
`
`40
`
`45
`
`50
`
`55
`
`SAMS-1007 Page 00005
`
`SAMS-1007 Page 00005
`
`

`

`EP 0 858 208 Ai
`
`Lees related to IR 246253X, A quad-linear sensor for high definition film scanning). However, because these scanners
`capture images from film, they are not subject to factors such as depth of field or differences in local geometry. This
`invention enables the advantages of a combination of high-resolution monochrome sensor and lower-resolution color
`sensors to be usedin the context of a digital camera by employing the techniques described earlier to convert the
`images from both sensor-types to a common global and local geometry. More specifically, a typical implementation of
`the method comprises: first the generation of acommon number of pixels by up-interpolating the color image to generate
`the same number of pixels as the monochrome image; secondly the conversion of both images to a commonglobal
`and local geometry by the methods described earlier; third, the application of a high-passfilter to the monochrome
`image, wherethe high-passfilter is complementary to the differencein the effectivefiltration between the monochrome
`and color channels (sensor and subsequentsignal processing for interpolation); finally the additionif the high spatial-
`frequency monochrome signal to the up-converted color channels with common global and local geometry to produce
`a full color image of improved characteristic.
`A further alternate type of image combination may be achieved, for example, with the use of three or more non-
`optically linked lens systems and three or more electronic sensors, wherein each of the multiple lenses is spectrally
`filtered to one of at least three color types (e.g. red, green and blue; white, yellow, and green; or cyan, yellow, green
`and white). By utilizing the above described global and local combination techniques to combine the at least three
`images, a full color image can be reproduced. This provides a further method to overcome the problem associated
`with the use of a single CCD to capture a full color image, described in the paragraph above, while also overcoming
`the problemsof alignment of the optics as described in the background. In addition, by utilizing greater than three
`spectrally filtered and captured images, it is possible to provide further improved color reproduction by being able to
`better estimate the original spectral power distribution of scene objects.
`
`Parts List
`
`1
`3a
`3b
`6
`7
`
`9
`16a
`50
`51a
`51b
`51c
`51d
`52
`76
`
`main camera body
`imaging lens
`imaging lens
`focusing mechanisms
`shutter switch
`
`shutter speed dial
`focusing mechanism
`main body
`imaging lens
`imaging lens
`imaging lens
`imaging lens
`shutter switch
`camera
`
`10
`
`15
`
`20
`
`25
`
`30
`
`35
`
`40
`
`45
`
`central processing unit
`78
`digital hard copy output printer
`82
`102=main body
`103a
`imaging lens
`1036
`imaging lens
`103c
`imaging lens
`103d
`imaging lens
`103e
`imaging lens
`103f
`imaging lens
`103g
`imaging lens
`103h
`imaging lens
`50
`105—shutter switch
`120
`block
`122
`block
`124
`block
`126
`block
`128
`block
`
`55
`
`SAMS-1007 Page 00006
`
`SAMS-1007 Page 00006
`
`

`

`Claims
`
`EP 0 858 208 Ai
`
`1. A method of producing a digital image with improved performance characteristics, comprising the stepsof:
`
`a) capturing at least two electronic images of a scene;
`b) digitizing the at least two electronic images of a scene; and
`c) combining and processing the at least two digitized electronic images of the scene to produce a combined
`digital image of a scene with improved performance characteristics.
`
`2. The method of claim 5 wherein the combining and processing step includes high-passfiltering of the digitized first
`and the at least one more electronic images containing more information at high spatial frequencies, and adding
`these high-passfiltered images to the at least one image containing less information at high spatial frequencies.
`
`15
`
`The method of claim 5 wherein the combining and processing step includes low-pass and high-passfiltering the
`digitized first and at least one more electronic images, and combining the imagesbyfirst averaging the low-pass
`filtered images and then adding the high-passfiltered images containing the most information.
`
`4. A method of producing a digital image with improved performance characteristics, comprising the stepsof:
`
`20
`a) producingafirst electronic image of a scene by a camera;
`b) producing at least one more electronic image of the same scene by the same camera but fromadifferent
`position;
`Cc) digitizing the first and the at least one more electronic images of a scene; and
`d)} combining and processing the digitized first electronic image and the at least one moredigitized electronic
`image to produce another digital image with improved performance characieristics.
`
`25
`
`5. Amethod of producing a digital image with improved performance characteristics, comprising the steps of:
`
`a) producingafirst electronic image of a scene by a camera;
`30
`b) producing at least one moreelectronic image of the same scene by the same camera but fromadifferent
`position;
`Cc) digitizing the first and the at least one more electronic images of a scene;
`d) storing the digitized electronic images of the scene on digital storage medial and
`€) combining and processing the digitized first electronic image and the al least one moredigitized electronic
`image stored on digital storage media to produce another digital image with improved performance charac-
`teristics.
`
`35
`
`40
`
`45
`
`50
`
`55
`
`6. Acomputer program product for producing a digital image with improved performance characteristics from the at
`least two digitized electronic images of the same scene and having a computer program stored thereon for per-
`forming the steps of:
`
`(i) converting to a common global geometry; and
`(ii) combining and processing the converted digitized electronic images to produce another digital image with
`improved performance characteristics.
`
`A computer program productfor producing a digital image with improved performance characteristics from the at
`least two digitized electronic images of the same scene and having a computer program stored thereon for per-
`forming the steps of:
`
`(i) converting the images to a commoncolor space;
`(ii) converting the images to a common number of pixels;
`(ili) converting to a commonglobal geometry; and
`(iv) combining and processing the converted digitized electronic images to produce another digital image with
`improved performance characteristics.
`
`The computer program product of claim 19 wherein the conversion step (iii) includes globally aligning the digitized
`first and at least one more electronic images, then, when thereis any misalignmentin any local area, measuring
`and correcting as appropriate for each of these locally misaligned areas of the digitized first and at least one more
`
`SAMS-1007 Page 00007
`
`SAMS-1007 Page 00007
`
`

`

`EP 0 858 208 Ai
`
`electronic images and combining the appropriately corrected images to produce a combined digital image with
`improved performance characteristics.
`
`10
`
`15
`
`20
`
`25
`
`30
`
`35
`
`40
`
`45
`
`50
`
`55
`
`SAMS-1007 Page 00008
`
`SAMS-1007 Page 00008
`
`

`

`EP 0 858 208 Ai
`
`{x
`
`SAMS-1007 Page 00009
`
`SAMS-1007 Page 00009
`
`

`

`EP 0 858 208 Ai
`
`76
`
`78
`
`82
`
`CENTRAL
`ara
`wiTH TEMPORARY L—»?ROCESSING
`
`DIGITAL STORAGE
`
`DIGITAL HARD
`(One
`
`FIG. 2
`
`SAMS-1007 Page 00010
`
`SAMS-1007 Page 00010
`
`

`

`EP 0 858 208 Ai
`
`FIRST DIGITIZED
`ELECTRONIC IMAGE
`
`
`MORE DIGITIZED
`ELECTRONIC IMAGE
`
`
`mew eee wm He ee ee te eh hee
`
` AT LEAST ONE
`
`118
`
`120
`
`122
`
`124
`
`SAMS-1007 Page 00011
`
`HIGH PASS FILTER
`
`COMBINE IMAGES
`
`a D
`
`IGITAL HARD
`COPY OUTPUT
`_ PRINTER
`
`FIG. 3
`
`SAMS-1007 Page 00011
`
`

`

`EP 0 858 208 Ai
`
`0
`
`European Patent
`
`Office
`
`EUROPEAN SEARCH REPORT
`
`Application Number
`
`EP 98 20 0205
`
`
`
`DOCUMENTS CONSIDERED TO BE RELEVANT
`
`
`
`HO4N1/21
`HO4N1/409
`
`
`
`
`
`
`
`
`
`
`PATENT ABSTRACTS OF JAPAN
`
`
`vol. 013, no. 459 (E-832), 17 October 1989
`
`
`& JP @1 17728

This document is available on Docket Alarm but you must sign up to view it.


Or .

Accessing this document will incur an additional charge of $.

After purchase, you can access this document again without charge.

Accept $ Charge
throbber

Still Working On It

This document is taking longer than usual to download. This can happen if we need to contact the court directly to obtain the document and their servers are running slowly.

Give it another minute or two to complete, and then try the refresh button.

throbber

A few More Minutes ... Still Working

It can take up to 5 minutes for us to download a document if the court servers are running slowly.

Thank you for your continued patience.

This document could not be displayed.

We could not find this document within its docket. Please go back to the docket page and check the link. If that does not work, go back to the docket and refresh it to pull the newest information.

Your account does not support viewing this document.

You need a Paid Account to view this document. Click here to change your account type.

Your account does not support viewing this document.

Set your membership status to view this document.

With a Docket Alarm membership, you'll get a whole lot more, including:

  • Up-to-date information for this case.
  • Email alerts whenever there is an update.
  • Full text search for other cases.
  • Get email alerts whenever a new case matches your search.

Become a Member

One Moment Please

The filing “” is large (MB) and is being downloaded.

Please refresh this page in a few minutes to see if the filing has been downloaded. The filing will also be emailed to you when the download completes.

Your document is on its way!

If you do not receive the document in five minutes, contact support at support@docketalarm.com.

Sealed Document

We are unable to display this document, it may be under a court ordered seal.

If you have proper credentials to access the file, you may proceed directly to the court's system using your government issued username and password.


Access Government Site

We are redirecting you
to a mobile optimized page.





Document Unreadable or Corrupt

Refresh this Document
Go to the Docket

We are unable to display this document.

Refresh this Document
Go to the Docket