throbber
IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, VOL. 7, NO. 4, AUGUST 1997
`
`575
`
`Multiresolution Image Sensor
`
`Sabrina E. Kemeny, Member, IEEE, Roger Panicacci, Bedabrata Pain,
`Larry Matthies, and Eric R. Fossum, Senior Member, IEEE
`
`Abstract— The recent development of the CMOS active pixel
`sensor (APS) has, for the first time, permitted large scale in-
`tegration of supporting circuitry and smart camera functions
`on the same chip as a high-performance image sensor. This
`paper reports on the demonstration of a new 128  128 CMOS
`APS with programmable multiresolution readout capability. By
`placing signal processing circuitry on the imaging focal plane, the
`image sensor can output data at varying resolutions which can
`decrease the computational load of downstream image processing.
`For instance, software intensive image pyramid reconstruction
`can be eliminated. The circuit uses a passive switched capacitor
`network to average arbitrarily large neighborhoods of pixels
`which can then be read out at any user-defined resolution by
`configuring a set of digital shift registers. The full resolution
`frame rate is 30 Hz with higher rates for all other image
`resolutions. The sensor achieved 80 dB of dynamic range while
`dissipating only 5 mW of power. Circuit error was less than
`34 dB and introduced no objectionable fixed pattern noise or
`other artifacts into the image.
`
`Index Terms— Focal plane array, image processing, imager,
`multimedia, sensor.
`
`I. INTRODUCTION
`
`problems become especially severe for image processing tasks
`1024) that
`performed on large format imagers (e.g., 1024
`are read out at video rates (30 frames/s).
`CMOS active pixel sensor (APS) technology allows the
`integration of support electronics and smart camera functions
`onto the same chip as a high-performance image sensor
`[6]. The integration of support electronics such as timing
`and control, correlated double sampling, and analog to dig-
`ital conversion leads to fewer components, thus increasing
`system robustness while reducing system mass and cost.
`The implementation of programmable multiresolution readout
`allows unprecedented camera functionality which eases the
`performance requirements of downstream image processing.
`The CMOS APS technology also enjoys other advantages over
`its charge-coupled device (CCD) counterpart such as ultra low
`lower than comparable CCD
`power performance (50–100
`systems) and increased radiation hardness [7]. CMOS APS
`architectures [Fig. 1(a)] allow – addressability of the array
`for windows of interest and sparse sampling readout of the
`array. Unfortunately, sparse sampling the array, for example,
`by reading out every fourth pixel of every fourth row, reduces
`the amount of image data by a factor of 1/16 but introduces
`aliasing into the image. In the multiresolution sensor, regions
`of the array are averaged together (block or kernel averaging)
`and read out [Fig. 1(b)], leading to data reduction without
`aliasing effects.
`128 photogate
`The multiresolution CMOS APS is a 128
`block
`array that is programmable to read out any size
`of pixels (kernel). Each kernel value represents the average
`of all the pixel values in its region. Averaging is done in the
`column readout circuitry so that the average value is based
`–
`on a full resolution image. Combining the sensor’s
`addressability with programmable resolution, the device can
`achieve true electronic zoom capability. In a standard digital
`camera, electronic zoom is achieved by mapping each pixel in
`a small area of interest to several display pixels. In contrast,
`the multiresolution sensor allows one to read out a small area
`of interest at a higher resolution than the full frame such
`that each pixel may be mapped to an individual display pixel
`much like an optical zoom lens allows one to capture more
`detail in a small area. This capability can also be used to
`increase processing speed of tracking algorithms where course
`resolution image data can be quickly read out and processed
`to determine an area of interest followed by read out of the
`area of interest at a higher resolution in the subsequent frame
`as illustrated in Fig. 2.
`Details of the multiresolution sensor operation are discussed
`in Section II. Section III presents the test results from the fab-
`1051–8215/97$10.00 ª
`
`FOR a variety of image processing tasks, such as biological
`
`vision modeling, stereo range finding, pattern recognition,
`target tracking, and transmission of compressed images, it is
`desirable to have image data available at varying resolutions
`to increase processing speed and efficiency. The user can then
`obtain a frame of data at the lowest resolution necessary for
`the task at hand and eliminate unnecessary processing steps.
`The multiresolution image data is usually generated through
`an image pyramid approach (implemented in software), and
`has been used extensively in recent years [1]–[4]. Typically,
`each image level is a low-pass filtered and down-sampled
`version of the prior level, although block averaging and
`down sampling can also be used to generate the pyramid
`[5]. In software, construction of the multiresolution pyramid is
`often the most computationally intensive and time consuming
`portion of the image processing task. For applications where
`power consumption is of concern, the power consumed by
`the processor while performing this task can be critical. These
`
`Manuscript received September 30, 1996; revised January 31, 1997. This
`paper was recommended by Guest Editors B. Sheu, C.-Y. Wu, H.-D. Lin,
`and M. Ghanbari. This work was performed by the Center for Space
`Microelectronics Technology, Jet Propulsion Laboratory, California Institute
`of Technology and was sponsored in part by the Jet Propulsion Laboratory
`Director’s Discretionary Fund (DDF) and the National Aeronautics and Space
`Administration, Office of Space Access and Technology.
`S. E. Kemeny, R. Panicacci, and E. R. Fossum are with Photobit, La
`Crescenta, CA 91214 USA.
`B. Pain and L. Matthies are with the Jet Propulsion Laboratory–California
`Institute of Technology, Pasadena, CA 91109 USA.
`Publisher Item Identifier S 1051-8215(97)05896-5.
`
`1997 IEEE
`
`Magna 2023
`TRW v. Magna
`IPR2015-00436
`
`

`
`576
`
`IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, VOL. 7, NO. 4, AUGUST 1997
`
`(a)
`
`(b)
`
`Fig. 3. Multiresolution image sensor block diagram.
`
`(a) Programmable multiresolution CMOS active pixel sensor archi-
`Fig. 1.
`tecture and (b) example of column’s functional configuration for 3  3 block
`averaging. (Actual neighborhood size is programmable.)
`
`Ideal switch and capacitor model for six columns configured for 3
`Fig. 4.
` 3 block averaging.
`
`the array is a network of capacitors to store pixel reset and
`signal levels. The column circuitry also contains an additional
`capacitor and a set of switches to the adjacent column to
`perform averaging on any size square array of pixels called
`a kernel (rectangular kernels are also possible). Resolution of
`the sensor is set by the size of the kernel. Large kernels sizes
`–
`are set for low resolution imaging requirements. The
`addressability of the sensor allows the user to zoom into areas
`of interest.
`Fig. 3 shows a block diagram of the sensor. A decoder at
`the side of the array selects a row of pixels for readout. Each
`pixel is controlled by a photogate signal enabling readout
`of integrated charge, a reset signal, and select signal
`to
`enable the buffered pixel data to drive the column output
`line. Column parallel circuitry at the bottom of the array
`samples the addressed row of pixel data simultaneously. The
`kernel size determines how a set of shift registers in the
`column circuits are configured. These shift registers control
`how the columns containing stored readout data are averaged
`and where the averaged row data is stored for subsequent
`
`Fig. 2. Sensor X –Y addressability and multiresolution readout allows the
`user to zoom into an area of interest with increased resolution.
`
`ricated chip. Finally, applications of the sensor are discussed
`in Section IV.
`
`II. DESIGN AND OPERATION
`
`A. Design Overview
`128 photogate pixel array sim-
`The sensor contains a 128
`ilar to previous APS arrays demonstrated at the Jet Propulsion
`Laboratory (JPL) [6]–[10]. At the bottom of each column in
`
`

`
`KEMENY et al.: MULTIRESOLUTION IMAGE SENSOR
`
`577
`
`Fig. 5. Multiresolution column processing circuitry for three columns.
`
`the bottom of the array
`processing. A second decoder at
`controls which columns containing the processed data are
`read out. The sensor’s differential output circuitry performs
`correlated double sampling (CDS) to suppress pixel kTC noise,
`noise, and fixed pattern noise.
`Row pixel data is read onto a column averaging capacitor
`with switches to its neighboring columns that are subse-
`quently enabled resulting in averaged column data for that
`row (Fig. 4). Averaged column data for that row is stored on a
`second bank of capacitors in one of the columns of the kernel.
`Data from successively read out rows is stored in each of the
`remaining columns in the kernel. Shift registers in the readout
`circuitry are configured according to kernel size to determine
`which switches are enabled to perform averaging and where
`the averaged column data is stored.
`rows of the kernel are read, they are averaged
`Once all
`by connecting the second bank of capacitors together and
`mixing the charge. A shift register configured to enable dummy
`switches to correct for switch feedthrough effects is also
`included. Both reset and signal levels are processed for a kernel
`so that correlated double sampling and double-delta sampling
`operations can be performed.
`Operation will be illustrated by stepping through the se-
`3 block (kernel) averaging (Fig. 4). In this
`quence for 3
`case, every third column average (CA) switch is open (i.e.,
`
`deselected, denoted by bit 0 over the switch), while the
`other switches are closed (i.e., selected, denoted by bit 1
`over the switch). Pixel signals are sampled onto the column
`averaging capacitors by globally pulsing (closing) the signal
`select switches (S). Charge redistributes such that the voltage
`on each capacitor in each block of three capacitors
`is
`the same such that
`
`is the horizontal size (number of columns) of the
`where
`block average (kernel),
`the pixel voltage value of the
`( – )th column, and
`is the average value for the th
`row in the kernel. These kernel row averages are then sampled
`-capacitor block of the row
`onto the first capacitor in the
`averaging bank of capacitors. Column averaging is repeated
`with the next pixel row (
`) and these new
`kernel averages are sampled onto the second capacitor in the -
`capacitor blocks of the row averaging bank of capacitors. The
`rows have been processed and
`process is repeated until all
`samples have been collected in
`adjacent capacitors in the
`row averaging bank. The temporal switching sequence (digital
`3 kernel case (Fig. 5). After
`pattern) is shown for the 3
`the
`-samples
`have been collected,
`charge is redistributed by pulsing the row averaging (RA)
`
`

`
`578
`
`IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, VOL. 7, NO. 4, AUGUST 1997
`
`Fig. 6. Transistor-level schematic of column circuit. Capacitors are poly-diffusion linear capacitors.
`
`(a)
`
`(b)
`
`(a) 4  4 block averaged time and (b) 1/4 subsampled image (no
`Fig. 8.
`averaging).
`
`for the first 3
`
`3 kernel
`
`Fig. 7. Sensor’s full resolution image (128  128).
`
`switches with the same pattern used for the column averaging
`switches, resulting in the final block average
`
`These kernel values are then scanned out of the array by
`reading out every th capacitor in the row average bank. The
`row averaging capacitors are then reset (circuitry not shown)
`and the process is repeated to generate the next row of kernels.
`Note that in the configuration described above, kernels must
`be either square or rectangular, where the number of rows must
`be less than or equal to the number of columns.
`
`is the vertical size (number of rows) in the kernel.
`where
`The constant factor of 1/2 arises from charge sharing between
`the column and row averaging capacitors when the column
`average is sampled onto the row averaging capacitor. Thus,
`
`B. Column Processing Circuitry
`Shown in Fig. 5 is the actual column parallel circuitry
`for three columns. There is one bank of column averaging
`capacitors and two banks of row averaging capacitors (rather
`
`

`
`KEMENY et al.: MULTIRESOLUTION IMAGE SENSOR
`
`579
`
`(a)
`
`(b)
`
`Fig. 9.
`
`(a) Full resolution sensor output for one row of pixels and (b) 4  4 kernel output from sample image.
`
`than the single bank shown in Fig. 4). One bank stores
`the row average pixel reset levels, and the other stores the
`row average pixel signal levels in order to perform on-chip
`double correlated sampling. The column averaging bank is
`used sequentially to horizontally average together the kernel
`row reset levels followed by the signal levels. The kernel
`reset switch to ground is shown as well as the column buffer
`amplifier for generating V R and V S. The buffer amplifier
`is only enabled when the column is selected for readout.
`The digital patterns shown are an example of the timing for
`a 3
`3 kernel. They are generated by gating the output of
`the configuration shift registers and the timing signals shown
`in Fig. 5. The global timing signals are CA (enable column
`averaging), RA (enable row averaging), VS (sample signal
`
`onto row averaging capacitor), and VR (sample reset onto row
`averaging capacitor). Each of these global signals is gated with
`the output of one of the two configuration shift registers. The
`CA and RA signal are gated with the output of the same shift
`register (CARA shift register). The VS and VR signals are
`gated with the output of the second shift register (VSVR shift
`register).
`is
`The transistor-level schematic of the column circuit
`shown in Fig. 6. The signals CA , RA , VS , and VR are the
`outputs from the corresponding global signals gated with the
`shift register output bit for that column. The buffer amplifier
`is a p-channel source follower. The CB signal is part of the
`double delta sampling readout scheme as reported in [9] used
`to reduce column fixed-pattern noise.
`
`

`
`580
`
`IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, VOL. 7, NO. 4, AUGUST 1997
`
`TABLE I
`SENSOR CHARACTERISTICS
`
`III. EXPERIMENTAL RESULTS
`The sensor was fabricated through MOSIS in the HP 1.2- m
`128 photogate sensor
`n-well CMOS process. The 128
`has a pixel pitch of 24 m. The total chip size is 4.8
`6.6 mm. Table I lists some of the sensor characteristics for
`full resolution operation.
`Fig. 7 shows a full resolution image of a test pattern used to
`demonstrate the sensor’s block averaging. Fig. 8(b) shows the
`same image for the sensor operating in a subsampling mode
`where every fourth column of every fourth row is read with
`no averaging. Because the test target contains relatively high
`spatial frequency patterns, the subsampled image produces
`32 image with the full
`dramatic aliasing. Comparing this 32
`128 image shows the appearance of both
`resolution 128
`fewer stripes and diagonal stripes rather than parallel stripes
`relative to the edge of the square. The 32
`32 image with 4
`4 kernel averaging [Fig. 8(a)] reduces this effect because the
`pixel array is read at full resolution and subsequently averaged.
`To measure how well the multiresolution sensor performs
`averaging, a test pattern containing a black and white stripe
`was imaged. The black–white edge (defocused) was positioned
`so that half the pixels in the kernels on the edge are black.
`Thus, the sensor output of the kernels aligned on top of the
`edge ideally should equal one-half of the difference between
`the totally white and black pixels. To measure the individual
`pixels in the kernel, subsampled data was first measured.
`Based on this subsampled raw image data, block averages
`were calculated for the pattern. This data was compared to
`the multiresolution sensor’s output at different kernel sizes.
`An example of the sensor’s output for one of the rows is
`shown in Fig. 9 where the sensor’s full resolution row data and
`4 kernel output data are shown. The row shown [Fig. 9(a)]
`4
`is one of four rows used to calculate the average from the full
`4 kernel
`resolution image for comparison with the on-chip 4
`4 kernel producing an
`average. Fig. 9(b) illustrates the 4
`output voltage at the average value of the four pixels at the
`black–white stripe edge (pixels in columns 65–68). Image data
`2, 4
`4, and 8
`8 kernel sizes were acquired for
`for 2
`this test pattern. Analysis of kernel data for the entire frame
`versus the off-chip block average data based on full resolution
`data shows that the sensor is accurate to within 2% ( 35 dB)
`of the ideal average value. The use of dummy switches for
`
`Fig. 10. Photograph of completed programmable multiresolution APS (128
` 128 array) IC.
`
`switch feedthrough compensation did not have a significant
`effect of the averaging accuracy.
`Table I lists the results from sensor characterization tests
`similar to those described in [9]. The sensor exhibited very low
`fixed pattern noise and dissipated very little power. Overhead
`for performing on-chip block averaging is a small percentage
`of the sensor readout time and total power consumption. For
`lower resolutions, the frame rate increase above 30 Hz is
`approximately proportional to the number of pixels,
`,
`in the kernel.
`
`IV. APPLICATIONS
`Multiresolution readout capability is useful in a wide variety
`of imaging applications where real world systems impose con-
`straints on format choices, processing speed, and bandwidth.
`A few applications that greatly benefit from such capability
`are described below.
`today,
`Data Reduction: In many imaging applications
`bandwidth limitations
`impose severe constraints on the
`manipulation and transmission of image data. From computer
`telephony to internet Vmail, transmission of image data is
`becoming increasingly common. Tremendous amounts of
`compression are needed to realize these functions (e.g., a com-
`pression ratio of 320 : 1 is required to transmit a video graphics
`adaptor (VGA) 640
`480-resolution image across a 28.8-kb/s
`phone line at video rates). Even without transmission, interfac-
`
`

`
`KEMENY et al.: MULTIRESOLUTION IMAGE SENSOR
`
`581
`
`(a)
`
`(b)
`
`(c)
`
`(d)
`
`Fig. 11. Programmable multiresolution sensor output for (a) full resolution, (b) 2  2 kernel, (c) 4  4 kernel, and (d) 8  8 kernel configurations.
`
`ing standard electronic cameras to a PC poses a compression
`challenge (e.g., 500 kbytes to 2 Mbytes per second enhanced
`parallel port (EPP), 12 MB for universal serial bus (USB)
`interface). One approach to achieving such compression is to
`reduce the amount of data being transmitted or compressed.
`In current CCD-based systems, such image editing may be
`realized by either throwing away image data (e.g., transmit the
`center of the image) or through image pyramid reconstruction.
`Unfortunately in the latter case, the required decimation and
`low-pass filtering is usually implemented in software, located
`beyond the bandwidth limited transmission link. Alternatively,
`memory and specialized hardware (digital signal processing
`(DSP) or custom application-specific integrated circuit (ASIC)
`[11]) may be integrated in the camera head to implement the
`image reconstruction.
`In contrast to these approaches, the multiresolution sensor
`may be programmed to readout a lower resolution image at
`any user-defined frame rate (including video or faster) with
`no additional hardware (decimation and averaging occur on-
`chip) or software overhead. The same optics setting can be
`used for the varying resolutions maintaining a constant field
`of view such that no part of the scene needs to be discarded.
`Robotics: The primary motivation for the work described
`in this paper was to reduce the computational complexity of
`algorithms devoted to autonomous navigation for vehicles in
`space. For power- and size-constrained missions, the multires-
`olution imager serves a dual purpose: navigational sensor and
`science sensor. In the former, for instance, low resolution data
`may be used for stereo vision-based autonomous navigation,
`while high-resolution visual data can be obtained for public
`relations and science.
`Target Tracking: In both commercial and military applica-
`tions, real-time tracking poses a difficult challenge. In a variety
`of defense imaging systems, for instance, high-speed target
`acquisition, tracking, and homing are essential operations. The
`multiresolution sensor can play a pivotal role in reducing the
`amount of image data that must be processed. For example,
`depending on the distance to the expected target, the sensor
`can be read out at the lowest resolution necessary such that
`the target covers a small number of pixels. Once potential
`targets are identified, small windows around each possible
`target can be read out at high resolution leading to better clutter
`rejection and faster processing speed. Tracking and homing
`can also benefit from similar optimal adjustment of resolution
`and windowing.
`
`tracking can be used
`In the commercial arena, subject
`in applications such as PC videoconferencing and perime-
`ter surveillance to reduce the required bandwidth of video
`transmissions. In the former, a low resolution “coarse” image
`is quickly read out and the subject of interest is identified.
`In subsequent frames, a high resolution window around the
`subject is read out and transmitted. Frequent repetition of this
`process is used to update the desired readout window. In the
`case of perimeter surveillance, the multiresolution capability
`can be exploited in a variety of ways. For example, low
`resolution imagery can be continually transmitted to a central
`processing workstation until movement or another triggering
`mechanism is alerted which would then signal the sensor or
`sensor bank to switch into high-resolution mode and start
`recording image data.
`Biological Vision: There is a trend among some researchers
`today [12], [13] to mimic simple biological vision systems in
`silicon. The multiresolution architecture can be extended to
`help realize these goals. Specifically, in the case of a retina,
`a foveated architecture in which the center pixels are read out
`at high resolution while the outer pixels are readout at lower
`resolutions is required. In the current sensor, kernel size is
`limited to either square or rectangular pixels and kernels must
`be of uniform size in the vertical direction. It is possible to
`vary the kernel size in the horizontal direction. In order to
`mimic biological systems, programmability of kernel size in
`both directions in a single frame would be required but could
`be realized with an extension of the approach described here.
`
`V. SUMMARY
`The multiresolution sensor, shown in Fig. 10, demonstrates
`the versatility of CMOS active pixel
`image sensors. On-
`chip column circuitry performs block averaging using pro-
`grammable kernel sizes. The images of George Washington in
`Fig. 11 from a dollar bill illustrate the sensor’s multiresolution
`ability. Shown are images at full resolution, 2
`2, 4
`4,
`and 8
`8 kernel configurations. The accuracy of averaging
`is within 2% of the average calculated from full resolution
`image data. With power consumption as low as 5 mW and
`30-Hz minimum frame rate operation for any resolution, this
`programmable multiresolution sensor can significantly reduce
`camera system complexity and power where multiresolution
`image processing is required, yet retain very high imaging
`performance of 80-dB dynamic range comparable to the very
`best CCD’s.
`
`

`
`582
`
`IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, VOL. 7, NO. 4, AUGUST 1997
`
`REFERENCES
`
`[1] P. J. Burt, “Fast filter transforms for image processing,” Comput.
`Graphics, Image Processing, vol. 16, pp. 20–51, 1981.
`[2] C. H. Anderson, “Filter-subtract-decimate hierarchical pyramid signal
`analyzing and synthesizing technique,” U.S. patent 4 718 104, Jan. 5,
`1988.
`[3] L. H. Matthies, “Stereo vision for planetary rovers: Stochastic modeling
`to near real-time implementation,” Int. J. Comput. Vision, vol. 8, no. 1,
`pp. 71–91, July 1992.
`[4] D. H. Ballard and C. M Brown, Computer Vision. Englewood Cliffs,
`NJ: Prentice-Hall, 1982.
`[5] H. P. Moravec, “Obstacle avoidance and navigation in the real world
`by a seeing robot rover,” Ph.D dissertation, Stanford Univ., Sept. 1980.
`[6] S. Mendis, S. E. Kemeny, and E. R. Fossum, “CMOS active pixel image
`sensor,” IEEE Trans. Electron Devices, vol. 41, no. 3, pp. 452–453,
`1994.
`[7] E. R. Fossum, “Active pixel sensors—Are CCD’s dinosaurs?,” in CCD’s
`and Optical Sensors III, Proc. SPIE, 1993, vol. 1900, pp. 2–14.
`[8] S. Mendis, S. E. Kemeny, R. Gee, B. Pain, and E. R. Fossum, “Progress
`in CMOS active pixel image sensors,” in Charge-Coupled Devices and
`Solid State Optical Sensors IV, Proc. SPIE, 1994, vol. 2172, pp. 19–29.
`[9] R. H. Nixon, S. E. Kemeny, C. O. Staller, and E. R. Fossum, “128 
`128 CMOS photodiode-type active pixel sensor with on-chip timing,
`control and signal chain electronics,” in Proc. SPIE, 1995, vol. 2415,
`pp. 117–123.
`[10] E. R. Fossum, “Low power camera-on-a-chip using CMOS APS tech-
`nology,” in 1995 Symp. Low Power Electronics, San Jose, CA, Oct.
`9–10, 1995, pp. 74–77.
`[11] G. S. van der Wal, “The Sarnoff pyramid chip,” in Proc. Computer
`Architecture for Machine Perception, CAMP-91, Paris, Dec. 16, 1991.
`[12] C. A. Mead, “Adaptive retina,” in Analog VLSI Implementation of Neural
`Networks. Boston, MA: Kluwer, 1989.
`[13] J. Van der Spiegel, G. Kreider, C. Claeys, I. Debusschere, G. Sandini,
`P. Dario, F. Fantini, P. Bellutti, and G. Soncini, “A foveated retina-
`like sensor using CCD technology,” in Analog VLSI Implementation of
`Neural Networks. Boston, MA: Kluwer, 1989.
`
`Sabrina E. Kemeny (S’84–M’91) received the A.S.
`degree from the University of Vermont, Burlington,
`in 1977. In 1981, she returned to academia, enrolling
`in Columbia University, New York, NY, where she
`received the B.S.E.E. degree in 1986,
`the M.S.
`degree in 1987, and the Ph.D. degree in 1991, all in
`electrical engineering.
`She began her professional career as a registered
`nurse after completing the A.S. degree. As a Ph.D.
`student, her work centered on the design and imple-
`mentation of charge-coupled device (CCD) image
`sensors and on-chip image processing using charge domain circuits. In the
`summer of 1988, she worked for the Ford Aerospace and Communications
`Corporation and designed the linear CCD image sensor flown on the Mars
`Observer Mission. Following completion of the Ph.D. degree, she joined the
`NASA Jet Propulsion Laboratory (JPL) in 1991, as a Member of the Technical
`Staff in the Microdevices Technology Section. At JPL she designed concurrent
`processing ASIC’s including neural networks and a high-speed SIMD path
`planner integrated circuit. She was a key member of the APS R&D Team and
`was one of the three original inventors of the APS technology. She managed
`several tasks at JPL and served as Study Team Leader on two studies related
`to imaging systems and on-board data processing. In 1995, she left JPL to
`form Photobit, La Crescenta, CA, and serves as CEO. She has published more
`than 25 technical papers and holds two patents with several patents pending.
`While an undergraduate at Columbia, Dr. Kemeny received the Helen
`Rubenstein Outstanding Women of Science Scholarship Award. Her high-
`speed SIMD path planner integrated circuit was nominated by JPL for an
`R&D 100 Award. In 1994, she received the JPL Exceptional Service Award
`for Concurrent Processor Development, a Group Achievement Award for
`Concurrent Processor Development, and a Group Achievement Award for the
`STRVIB Payload Development Team. In 1996, she received the NASA Group
`Achievement Award for her contributions to the JPL CMOS APS research
`effort.
`
`Roger Panicacci earned the B.S. degree in elec-
`trical engineering from the University of California,
`Berkeley, in 1985 and the M.S.E.E. degree from the
`University of California, Santa Barbara, in 1989.
`He is both a Senior Engineer and an original
`founder of Photobit, La Crescenta, CA. He is re-
`sponsible for VLSI circuit design and layout, test,
`and integration of Photobit APS products. At Photo-
`bit, he completed the design of 576  432 element
`sensor with on-chip ADC that was 100% successful
`on first silicon. Prior to Photobit, he was with the
`Jet Propulsion Laboratory’s Advanced Imager Technology Group where his
`responsibilities included imager chip architecture, digital control logic, analog
`circuitry, and board level test system. He has been the primary designer on
`several APS chips including a multiresolution robotics imaging APS, a very
`high frame rate APS for NRL, and JPL’s 1024  1024 element sensor with on-
`chip analog to digital conversion. Prior to JPL, he was with Kodak Berkeley
`Research for four years in their VLSI design group where he was responsible
`for the design of a number of macrocells (FIFO’s, ROM’s, PLA’s I/O cells), a
`mixed analog-digital 1-GHz optical sensor chip for a high-speed optical tape
`drive, a number of board-level designs for communication applications, and
`the design of a data block decoder/error corrector for the Kodak PhotoCD
`product. Prior to Kodak, he was with Delco System Operations for four years
`as a VLSI Design Engineer.
`In 1996, Mr. Panicacci received the NASA Group Achievement Award for
`his contributions to the JPL CMOS APS research effort.
`
`Bedabrata Pain received the Bachelor of Technol-
`ogy degree in 1986 from the Indian Institute of
`Technology, Kharagpur, India, and the Masters and
`Ph.D. degrees in electrical engineering at Columbia
`University, NY, in 1989 and 1993, respectively.
`In 1993, he joined the Jet Propulsion Laboratory
`(JPL) as a Post-Doctoral Research Associate. Since
`1994, he has been a Member of the Technical Staff
`at JPL. Currently, he heads the advanced imager and
`focal-plane technology group at JPL and is involved
`in research and development of CMOS active pixel
`sensors, infrared sensors, and integrated charged particle detectors. His current
`research interests include low-noise, low-power mixed analog/digital VLSI,
`and integrated sensor technology.
`Dr. Pain is the recipient of two NASA achievement awards for his contri-
`bution to the active pixel sensor technology and the airborne visible/infrared
`imaging spectrometer.
`
`Larry Matthies received the Ph.D. degree in computer science from Carnegie
`Mellon University, Pittsburgh, PA, in 1989.
`He is now a Member of the Technical Staff at the Jet Propulsion Laboratory
`(JPL) in Pasadena, CA. His research to date has focused on 3-D sensing
`and representation methods for autonomous navigation. In particular, he has
`applied stochastic modeling and recursive estimation methods to 3-D shape
`and motion estimation from stereo vision, from image sequences, and from
`the combination of stereo vision and sonar. At JPL, he has developed a real-
`time stereo vision system for range estimation in cross-country navigation.
`His research interests include sensing, world modeling, and planning for
`robotics and related applications. He has also worked and taught in the field of
`computer graphics at the National Research Council of Canada, the Tektronix
`Corporation, and the University of Waterloo.
`
`

`
`KEMENY et al.: MULTIRESOLUTION IMAGE SENSOR
`
`583
`
`Eric R. Fossum (S’80–M’84–SM’91) received the
`B.S. degree in physics (with Honors) and engineer-
`ing from Trinity College, Hartford, CT, in 1979, the
`M.S. degree in applied physics in 1980 from Yale
`University, New Haven, CT, and the Ph.D. degree
`from Yale in electrical engineering in 1984.
`During the summers of 1981–1983, he was with
`the Hughes Aircraft Company, Canoga Park, CA,
`working on various problems related to infrared
`focal-plane array detector and readout structures. He
`joined Columbia University in 1984 as an Assistant
`Professor and was engaged primarily in research on silicon charged-coupled
`devices (CCD’s) for on-focal-plane image processing and III–V CCD’s for
`very high speed signal processing. Other activities included novel devices
`and structures for fiber-based optical interconnects and the low energy ion
`beam modification of semiconductor surfaces. He was promoted to Associate
`Professor in 1989. He joined the Jet Propulsion Laboratory (JPL) in 1990
`as a Technical Assistant Section Manager. At JPL he was responsible for
`developing visible and infrared imaging technology and programs in a section
`of approximately 110 researchers and engineers. He initiated and led JPL’s
`pioneering research in CMOS active pixel
`image sensors. In 1995, he
`helped to found Photobit, La Crescenta, CA, and in 1996, joined Photobit
`as Chief Scientist. At Photobit, he is responsible for advanced research
`and development of CMOS active pixel sensors for commercial, industrial,
`government, and consumer applications. He also serves as Chairman of t

This document is available on Docket Alarm but you must sign up to view it.


Or .

Accessing this document will incur an additional charge of $.

After purchase, you can access this document again without charge.

Accept $ Charge
throbber

Still Working On It

This document is taking longer than usual to download. This can happen if we need to contact the court directly to obtain the document and their servers are running slowly.

Give it another minute or two to complete, and then try the refresh button.

throbber

A few More Minutes ... Still Working

It can take up to 5 minutes for us to download a document if the court servers are running slowly.

Thank you for your continued patience.

This document could not be displayed.

We could not find this document within its docket. Please go back to the docket page and check the link. If that does not work, go back to the docket and refresh it to pull the newest information.

Your account does not support viewing this document.

You need a Paid Account to view this document. Click here to change your account type.

Your account does not support viewing this document.

Set your membership status to view this document.

With a Docket Alarm membership, you'll get a whole lot more, including:

  • Up-to-date information for this case.
  • Email alerts whenever there is an update.
  • Full text search for other cases.
  • Get email alerts whenever a new case matches your search.

Become a Member

One Moment Please

The filing “” is large (MB) and is being downloaded.

Please refresh this page in a few minutes to see if the filing has been downloaded. The filing will also be emailed to you when the download completes.

Your document is on its way!

If you do not receive the document in five minutes, contact support at support@docketalarm.com.

Sealed Document

We are unable to display this document, it may be under a court ordered seal.

If you have proper credentials to access the file, you may proceed directly to the court's system using your government issued username and password.


Access Government Site

We are redirecting you
to a mobile optimized page.





Document Unreadable or Corrupt

Refresh this Document
Go to the Docket

We are unable to display this document.

Refresh this Document
Go to the Docket