throbber
as) United States
`a2) Patent Application Publication 10) Pub. No.: US 2012/0140243 Al
`
` Colonna de Lega (43) Pub. Date: Jun. 7, 2012
`
`
`US 20120140243A1
`
`Publication Classification
`
`(51)
`
`Int. Cl
`(2006.01)
`GO1B 11/24
`(52) U.S. C1. cc ceeeseseeesersereeceeceeceeceeceeeneenes 356/609
`(57)
`ABSTRACT
`Methods for forming a three-dimensional image ofa test
`object include directing light to a surface of best-focus of an
`imaging optic, where thelight has an intensity modulation in
`at least one direction in the surface of best-focus, scanning a
`test object relative to the imaging optic so that a surface ofthe
`measurement object passes through the surface ofbest-focus
`of the imaging optic as the test object is scanned, acquiring,
`for each of a series of positions of the test object during the
`scan, a single image of the measurement object using the
`imaging optic, in which the intensity modulation ofthe light
`in the surface ofbest-focusis different for successive images,
`and forming a three-dimensional image of the test object
`
`(54) NON-CONTACT SURFACE
`CHARACTERIZATION USING MODULATED
`ILLUMINATION
`
`(75)
`
`Inventor:
`
`Xavier M. Colonna de Lega,
`Middlefield, CT (US)
`
`(73) Assignee:
`
`Zygo Corporation, Middfield, CT
`(US)
`
`(21) Appl. No.:
`
`13/309,244
`
`(22)
`
`Filed:
`
`Dec. 1, 2011
`
`Related U.S. Application Data
`
`(60) Provisional application No. 61/419,386, filed on Dec.
`3, 2010.
`
`based on the acquired images.
`
`Align Ex. 1017
`U.S. Patent No. 9,962,244
`
`0001
`
`Align Ex. 1017
`U.S. Patent No. 9,962,244
`
`0001
`
`

`

`Patent Application Publication
`
`Jun. 7,2012 Sheet 1 of 18
`
`US 2012/0140243 Al
`
`
`
`0002
`
`0002
`
`

`

`Patent Application Publication
`
`Jun. 7,2012 Sheet 2 of 18
`
`US 2012/0140243 Al
`
`
`
`0003
`
`0003
`
`

`

`Patent Application Publication
`
`Jun. 7,2012 Sheet 3 of 18
`
`US 2012/0140243 Al
`
`318
`
`302
`
`
`
`
`
`Scan
`
`
`
`(e) }
`
`:
`
`FIG. 3
`
`0004
`
`0004
`
`

`

`Patent Application Publication
`
`Jun. 7,2012 Sheet 4 of 18
`
`US 2012/0140243 Al
`
`
`
`0005
`
`0005
`
`

`

`Patent Application Publication
`
`Jun. 7,2012 Sheet 5 of 18
`
`US 2012/0140243 Al
`
`
`
`FIG. 6
`
`0006
`
`0006
`
`

`

`Patent Application Publication
`
`US 2012/0140243 Al
`
`Jun. 7,2012 Sheet 6 of 18
`
`FIG.7
`
`0007
`
`0007
`
`

`

`Patent Application Publication
`
`Jun. 7,2012 Sheet 7 of 18
`
`US 2012/0140243 Al
`
`08‘Sls a8‘Sis V8‘Old
`
`0008
`
`0008
`
`

`

`Patent Application Publication
`
`Jun. 7,2012 Sheet 8 of 18
`
`US 2012/0140243 Al
`
`Spectral magnitude
`
`
`
`Frequency
`FIG. 9A
`
`Spectral phase
`
`Frequency
`
`FIG. 9B
`
`0009
`
`0009
`
`

`

`Patent Application Publication
`
`US 2012/0140243 Al
`
`Jun. 7,2012 Sheet 9 of 18
`
`d0b‘Sls
`
`0010
`
`0010
`
`

`

`US 2012/0140243 Al
`
`\
`
`\SS
`
`yy
`
`ie
`
`esGYyy
`esiyLie
`
`Jun. 7,2012 Sheet 10 of 18
`
`Yue
`
`Patent Application Publication
`
`S
`
`SSS
`SS
`\
`AX
`
`0011
`
`0011
`
`

`

`Patent Application Publication
`
`Jun. 7,2012 Sheet 11 of 18
`
`US 2012/0140243 Al
`
`
`
`0012
`
`0012
`
`

`

`Patent Application Publication
`
`Jun. 7,2012 Sheet 12 of 18
`
`US 2012/0140243 Al
`
`1402
`
`LO
`com)
`wzel
`
`o/©=

`
`FIG.14
`
`t
`I
`t
`I
`
`t
`!
`|
`1
`
`1403
`
`1404
`
`Fg
`
`JANE
`YY
`
`1
`
`Ww
`
`{O
`
`g |—_re
`
`—tee
`w_oo
`
`1416
`
`0013
`
`0013
`
`

`

`Patent Application Publication
`
`Jun. 7,2012 Sheet 13 of 18
`
`US 2012/0140243 Al
`
` cOSL
`
`OOSL
`
`0014
`
`0014
`
`

`

`Patent Application Publication
`
`Jun. 7,2012 Sheet 14 of 18
`
`US 2012/0140243 Al
`
`
`
`Vv
`a
`
`Oo
`ra)
`Sre
`
`3
`So
`—
`
`CO
`ere
`
`ae)
`3

`ve
`
`N O
`
`o©e
`
`r
`
`
`en1 tga i
`S
`rir dia’
`o
`
`
`
`
`3 fFoekyr tt tog wo
`
`O
`=_—
`
`
`
`~ eya
`
`1606
`
`1614
`
`0015
`
`0015
`
`

`

`Patent Application Publication
`
`Jun. 7,2012 Sheet 15 of 18
`
`US 2012/0140243 Al
`
`dZbSls
`
`0016
`
`0016
`
`

`

`Patent Application Publication
`
`Jun. 7,2012 Sheet 16 of 18
`
`US 2012/0140243 Al
`
`cO8l
`
`aslOld
`
`cOst
`
`V8-Sls
`
`cO8L
`
`
`
`0017
`
`0017
`
`
`

`

`Patent Application Publication
`
`Jun. 7,2012 Sheet 17 of 18
`
`US 2012/0140243 Al
`
`wi
`
`61‘Old
`
`O61
`Vlas
`
`0018
`
`0018
`
`
`
`

`

`Patent Application Publication
`
`Jun. 7,2012 Sheet 18 of 18
`
`US 2012/0140243 Al
`
`2000
`
`Scan
`
` KWfNEY
`
`foNIY
`
`FIG. 20
`
`0019
`
`0019
`
`

`

`US 2012/0140243 Al
`
`Jun. 7, 2012
`
`NON-CONTACT SURFACE
`CHARACTERIZATION USING MODULATED
`ILLUMINATION
`
`CROSS REFERENCE TO RELATED
`APPLICATIONS
`
`[0001] This application claims benefit ofprovisional patent
`application No. 61/419,386, entitled “Non-Contact Surface
`Characterization Using Modulated Hlumination,” filed on
`Dec. 3, 2010, the entire contents of which are hereby incor-
`porated by reference.
`
`TECHNICAL FIELD
`
`[0002] This disclosure relates to methods of non-contact
`surface characterization using modulated illumination and
`devices configured to perform the same.
`
`BACKGROUND
`
`[0003] The field of non-contact surface characterization
`includes characterization and/or measurementof information
`
`about an object’s surface including, for example, shape and
`surface finish measurements. Tools capable of non-contact
`surface characterization can be useful in manufacturing dur-
`ing stages such as process development, quality control and
`process control on the production floor. Non-contact surface
`characterization methods can include obtaining information
`about the three-dimensional (3D) shape and location of an
`object by analyzing the frequency content of an imageofthe
`object in an optical system.
`
`SUMMARY
`
`Ingeneral, the subject matter described in this speci-
`[0004]
`fication relates to methods and systems for obtaining infor-
`mation about a surface of an object. The methods disclosed
`can include, for example, directing a structured illumination
`pattern to a surface of best-focus of an imaging optic while
`imaging a field in the surface of best-focus onto a multi-
`element detector. A structured illumination pattern is a non-
`uniform illumination pattern that contains some form of
`encoding based on light intensity. Examples of structured
`illumination patterns includeperiodicintensity patterns (e.g.,
`modulated in one or two dimensions). An object is scanned
`through the surface of best-focus while a series of images of
`the illumination patterns on the object surface are sequen-
`tially acquired by the multi-element detector. For each suc-
`cessive image obtained, the illumination pattern is modified
`so that its intensity at each point in the field corresponding to
`a detector element is modulated from one frameto the next.
`
`The resulting signal acquired at each detector element has a
`modulation, with the maximum amplitude occurring at or
`near the position of the object surface as it intersects the
`surface ofbest-focusofthe objective. Based on an analysis of
`the signals provided by the sequence of images, measure-
`ments of topography and surface texture of the object can be
`obtained and, in some implementations, a three-dimensional
`imageof the object’s surface can be produced.
`[0005]
`In some embodiments, a detector plane and pattern-
`generating plane are mapped onto a surface in object space
`that is nominally conformal to the shape of the object to be
`characterized including, for example, a plane, a sphere, a
`parabola, a cylinder, a cone, or an aspheric optic.
`[0006]
`In certain embodiments, the signal recorded at a
`pixel(1.e., element) on the detector has a modulation envelope
`
`that emulates some of the characteristics of a scanning low-
`coherence interferometer (SWLI) signal. Conversion of scan
`data to topography or reflectivity data may therefore be
`accomplished by application of envelope-detection algo-
`rithms developed for SWLI such as a frequency domain
`analysis (“FDA”) or least squares analysis (“LSQ”).
`[0007]
`In some embodiments, the device used for creating
`the projected illumination pattern is programmable, which
`allows adapting the frequency content, orientation andspatial
`intensity distribution to optimize the measurement capability
`for a given object.
`[0008] The structured illumination modulation scheme can
`be optimized to enable a fast autofocus scan of an optical
`system. An application in the context of laser eye surgery is
`the localization of the position of laser optics with respect to
`a critical component that makes contact with the cornea.
`Another application is rapid focusing of a low-coherence
`interferometer,
`in which case the autofocus scan rate is
`selected to average out the interference signal from the
`detected signal.
`[0009]
`In certain embodiments, the apparatus presents to
`the user an enhanced imageofthe object that combines height
`information with additional surface information, such as
`color, absorption, texture,etc.
`[0010] Various aspects of the subject matter described in
`this specification are summarizedas follows.
`[0011]
`In general, one aspect of the subject matter
`described in this specification can be embodied in methods
`for forming a three-dimensional image of a test object, in
`which the methods include directing light to a surface of
`best-focus of an imaging optic, where the light has an inten-
`sity modulation in at least one direction in the surface of
`best-focus, scanning a test object relative to the imaging optic
`so that a surface ofthe measurementobjectpasses through the
`surface of best-focus of the imaging optic as the test objectis
`scanned, acquiring, for each of a series ofpositionsofthe test
`object during the scan, a single image of the measurement
`object using the imaging optic, in which the intensity modu-
`lation of the light in the surface of best-focus is different for
`successive images, and forming a three-dimensional image of
`the test object based on the acquired images.
`[0012] These and other embodiments can each optionally
`include one or more ofthe following features. For example,in
`some implementations, directing the light to the surface of
`best-focus includes imaging a spatial light modulator (SLM)
`to the surface of best-focus. The intensity modulation of the
`light in the surface of best-focus can be varied using the
`spatial light modulator.
`[0013]
`In some implementations, directing the light to the
`surface of best-focus includes imaging a pattern-generating
`plane onto a surface in object space. The surface can be
`conformal to a shape of the test object. The shapeof the test
`object can be planar, spherical, parabolic, cylindrical, coni-
`cal, or aspheric.
`[0014]
`In some implementations, the intensity modulation
`is a periodic modulation. The periodic modulation can be a
`sinusoidal modulation. The phaseofthe periodic modulation
`can be varied by less than 27 between each successive image.
`The phase of the periodic modulation can be varied by x or
`less between each successive image. The phase ofthe peri-
`odic modulation can be varied by 1/2 between each succes-
`sive image.
`[0015]
`In some implementations, the intensity modulation
`is a two-dimensional intensity modulation. The scan posi-
`
`0020
`
`0020
`
`

`

`US 2012/0140243 Al
`
`Jun. 7, 2012
`
`In some implementations, the SLM can be a liquid
`[0023]
`crystal panel, or include a micro-mirrorarray.
`[0024]
`In some implementations, the imaging optic has a
`numerical aperture greater than 0.6, a numerical aperture
`greater than 0.8, a numerical aperture greater than 0.9, or a
`numerical aperture of 0.95.
`the system further
`[0025]
`In some implementations,
`includescolorfilters arranged to filter the wavelength of light
`forming the images at the multi-element detector.
`[0026] The one or more optical elements can include, for
`example, a fisheye lens, an endoscope, and/or a zoom lens.
`[0027]
`Implementations disclosed herein can offer several
`advantages. For example,
`in some implementations,
`the
`methods and apparatus can be used to provide non-contact
`three-dimensional imaging of an object. The imaging can be
`performed using short measurement times and/or to achieve
`high resolution images of objects’ surfaces. In some imple-
`mentations, the imaging is performed with reduced sensitiv-
`ity to environmental perturbations such as, for example,
`vibration or acoustic noise.
`
`arranged to direct light from the spatial light modulator to
`tions can be evenly spaced. The intensity modulation can be
`form an image of the SLM atthe surface of best-focus during
`selected based on a slope ofthe test object surface.
`operation ofthe system, a scanning stage arrangedto scan the
`[0016]
`Insome implementations, forming the three-dimen-
`test object relative to the microscope object during operation
`sional image includes identifying, for multiple different loca-
`of the system so that a surface ofthe test object intersects the
`tionsofthe test object surface, the scan position correspond-
`surface of best-focus, a multi-element detector positioned
`ing to where eachlocationintersects the surface ofbest-focus.
`relative to the microscope such that the microscope forms an
`An intensity of the acquired images as a function of scan
`image ofa field at the surface of best-focus on the multi-
`position at each of the different locations can include an
`element detector during operation of the system, and an elec-
`amplitude modulation. Identifying the scan position corre-
`tronic control module in communication with the scanning
`sponding to where each location intersects the surface of
`stage,
`the spatial
`light modulator, and the multi-element
`best-focus can include identifying the scan position where a
`detector, in which during operation, the system causes the
`modulation amplitudeis largest.
`multi-element detector to acquire a single image ofthe test
`[0017]
`Insome implementations, forming the three-dimen-
`object for each of multiple scan positions of the test object
`sional image includesderiving an intensity signal for each of
`relative to the imaging optic, causes the SLM to variably
`multiple different locations of the test object surface, each
`modulate the intensity of light at the surface of best-focus in
`intensity signal corresponding to the intensity of the acquired
`at least one direction so that the intensity modulation of the
`images at the corresponding location as a function of scan
`light is different for successive images, and formsa three-
`position. Forming the three-dimensional image can include
`dimensional image of the test object based on the acquired
`identifying a scan position corresponding to a maximum
`images.
`amplitude of a modulation of the intensity signal for each
`[0022] This and other embodiments can each optionally
`location. Identifying the scan position can include transform-
`include one or moreofthe following features. In some imple-
`ing each intensity signal into a frequency domain.
`mentations, the system includes, for example, a light source
`[0018]
`In some implementations, the test object is a lens
`arranged to direct light to the SLM during operation of the
`element. The three-dimensional image can be, for example, a
`system. The SLM can beareflective SLM ora transmissive
`monochrome image. Alternatively,
`the three-dimensional
`SLM.
`image can be a color image.
`[0019]
`In general, another aspect of the subject matter
`described in this specification can be embodied in methods
`for forming a three-dimensional image of a test object, in
`which the methodseach include forming an imageofa spatial
`light modulatorat a surface ofbest-focus of an imagingoptic,
`scanning a test object relative to the imaging optic so that a
`surface ofthe measurementobject passes through a surface of
`best-focus of the imaging optic as the test object is scanned,
`acquiring, for each of a series of positions of the test object
`during the scan, a single imageof the test object using the
`imaging optic, in which the spatial light modulator varies an
`intensity modulationin the light forming the image sothat the
`modulation ofthe light at the surface ofbest-focusis different
`for successive images, and forming a three-dimensional
`imageofthe test object based on the acquired images.
`[0020]
`In general, another aspect of the subject matter
`described in this specification can be embodied in methods
`for forming a three-dimensional image of a test object, in
`which the methodseach includedirecting light to a surface of
`best-focus of an imaging optic, where the light has an inten-
`sity modulation in at least one direction in the surface of
`best-focus, scanning a test object relative to the imaging optic
`so that a surface of the measurementobject passes through a
`surface of best-focus of the imaging optic as the test objectis
`scanned, imaging the surface ofbest-focus to a multi-element
`detector, acquiring, for each ofa series ofpositions ofthe test
`object during the scan, a single intensity measurementat one
`or more elements of the multi-element detector, in which the
`intensity modulationofthe light in the surface ofbest-focusis
`different for successive positions of the test object during the
`scan, and forming a three-dimensional image of the test
`object based on the acquired intensity measurements.
`[0021]
`In general, another aspect of the subject matter
`describedin this specification can be embodied in systems for
`forming a three-dimensional imageofa test object, in which
`the systems each include a microscope including an imaging
`optic, the imaging optic having a surface of best-focus, a
`spatial
`light modulator, one or more optical elements
`
`Implementationsoffer other advantagesas well. For
`[0028]
`example, the methods andstructures disclosed herein can, in
`some implementations, provide similar depth sectioning
`capabilities as a conventional confocal microscope. The use
`of structured illumination profiling can offer certain benefits
`compared to confocal microscopy (e.g., reduced source
`brightness requirements) and interference microscopy. For
`example, in some implementations, structured illumination
`has reduced source brightness requirements relative to con-
`ventional interference microscopes. Alternatively, or in addi-
`tion, restrictions on which non-interferometric microscope
`objectives can be used can be reduced. For example, in some
`implementations, higher numerical aperture objectives are
`available relative to the objectives available in Michelson or
`Mirau interferometers.
`
`[0029] The details of one or more embodiments are set
`forth in the accompanying drawings and the description
`
`0021
`
`0021
`
`

`

`US 2012/0140243 Al
`
`Jun. 7, 2012
`
`below. Other features and advantages will be apparent from
`the description, drawings, and from the claims.
`
`BRIEF DESCRIPTION OF DRAWINGS
`
`FIG. 11s a schematic diagram of an exemplary sys-
`[0030]
`tem for providing a 3D image of a sample object.
`[0031]
`FIG. 21s a schematic diagram of an exemplary sys-
`tem for providing a 3D image of a sample object.
`[0032]
`FIG. 3 is a schematic illustrating various scan posi-
`tions of a sample object.
`[0033]
`FIG. 4isa graph ofa simulatedlightintensity signal.
`[0034] FIG.5is a graph of an experimentalintensity signal
`recorded by a pixel of an image-recording device.
`[0035]
`FIG. 6 is an exemplary topography map of a 13-um
`step height step generated by FDA processingoflight inten-
`sity signals.
`[0036]
`FIG. 7 is a graph of a simulated intensity signal.
`[0037]
`FIGS. 8A-8Bare plots of a simulated analog inten-
`sity signal overlaid with a corresponding digitized sampling
`of the intensity signal.
`[0038]
`FIG. 8C is a plot of the simulated analog intensity
`signal of FIG. 8B overlaid with the corresponding digitized
`intensity signal of FIG. 8B, where zeros have been inserted
`into the digitized intensity signal.
`[0039]
`FIG. 9A is a plot of the magnitude ofthe digitized
`sample signal from FIG. 8A and the phase of the digitized
`intensity signal from FIG. 8C.
`[0040] FIG.9B isa plotofthe phaseofthe digitized sample
`signal from FIG. 8A and the magnitudeofthe digitized inten-
`sity signal from FIG. 8C.
`[0041]
`FIG. 10A is a plot of a simulated analog intensity
`signal overlaid with a correspondingdigitized sampling ofthe
`simulated intensity signal.
`[0042]
`FIG. 10B is a plot of the simulated analog intensity
`signal of FIG. 10A overlaid with the digitized intensity signal
`of FIG. 10A, where zero values have been inserted into the
`digitized intensity signal.
`[0043]
`FIGS. 11-13 are exemplary images of a microlens
`onto which an illumination pattern is projected.
`[0044]
`FIG. 14 is a schematic diagram of an exemplary
`structured illumination imaging system.
`[0045]
`FIG. 15 is a schematic diagram of an exemplary
`imaging system.
`[0046]
`FIG. 16 is a schematic diagram of an exemplary
`imaging system.
`[0047]
`FIGS. 17A-17B are schematic diagrams of exem-
`plary imaging systems.
`[0048]
`FIGS. 18A-18B are schematic diagrams of exem-
`plary imaging systems.
`[0049] FIG.19A is a3D graphof experimental topography
`data collected on a sample object.
`[0050]
`FIG. 19B is a color image of the surface of the
`sample object of FIG. 19A.
`[0051]
`FIG. 19C is an exemplary color image produced by
`combining the topographical data of FIG. 19A with the color
`image of FIG. 19B.
`[0052]
`FIG. 20 is a schematic of an exemplary system for
`performing laser eye surgery.
`
`DETAILED DESCRIPTION
`
`Acquiring the Light Intensity Signal
`[0053]
`FIG. 1 is a schematic diagram of an exemplary
`structured illumination system 100 for providing a 3D image
`
`of a sample object 102. An illumination portion 103 of the
`system 100 includes an illumination source 104, a spatial
`light modulator (SLM) 106, one or more lenses (108a-108c),
`an aperture 110 and a beam combiner 112. An imagingpor-
`tion 105 of the system 100 includes an image-recording
`device 114, a beam splitter 116, an objective lens 118 and a
`tube lens 120. For the purposes of this disclosure, the coor-
`dinate system is defined such that the z-axis is parallel to the
`optical axis of the objective lens 118 and the x and y-axes are
`parallel to the lens’ plane ofbest-focus such that xyz forms an
`orthogonal coordinate system.
`[0054] During operation of the system 100, light 101 gen-
`erated from the illumination source 104 passes through the
`lenses 108a, 1084 and the aperture 110, wherethe light 101
`then is incident on the beam splitter 112. Beam splitter 112
`reflects a portion of the incidentlight 101 onto the SLM 106.
`In the present example, the SLM 106 is configured to modify
`and reflect the light incidenton its surface so as to produce an
`illumination pattern characterized as either binary (i.e., the
`imaging pattern has regions wherelight is present and regions
`wherelight is absent) or quasi-continuous(i.e., the imaging
`pattern can be approximated by continuous varying levels of
`light intensity).
`[0055] The illumination pattern reflected from the SLM
`106 then passes through the beam splitter 112 and lens 108c
`to a second beam splitter 116 where the illumination pattern
`thenis directed through the objective lens 118 and preferen-
`tially fills the back pupil of the objective 118. The illumina-
`tion pattern is re-imaged in object space to the plane ofbest-
`focus (focal plane) of the objective lens 118. Lightreflecting
`and/or scattering off a surface of the sample object 102 then
`proceeds through the objective lens 118, the beam splitter
`116, and the tube lens 120 onto a light-detecting surface/plane
`of the image-recording device 114 whereit is recorded. The
`recordedlight thus acquired can be storedin digital format as
`an array of light intensity signals, with each light intensity
`signal being acquired from a corresponding pixel of the
`image-recording device 114.
`[0056] During imaging, the object 102 is translated verti-
`cally with respect to the objective lens 118 (i.e., toward or
`away from the objective lens 118). The sample object 102 can
`be displaced or actuated by an electromechanical transducer
`(not shown) andassociated drive electronics controlled by a
`computer 122 so as to enable precise scans along a direction
`of translation of the object 102. Examples of transducers
`include, but are not
`limited to, piezoelectric transducers
`(PZT), stepper motors and voice coils. Alternatively, or in
`addition, the objective lens 118 maybetranslated vertically
`with respectto a position of the sample object 102. Again, an
`electromechanical transducer and associated drive electron-
`ics may be usedto controlthe translation. The image-record-
`ing device 114 simultaneously records imaging data as the
`object 102 is scanned through the plane of best-focus of the
`objective lens 118 such that multiple light intensity signals
`will be recorded overtime. That is, an imageofthe object and
`illumination pattern is captured by the image-recording
`device in the form ofthe light intensity signal at correspond-
`ing scan positions. For example, if the image-recording
`device 114 includes a 128x128 array of pixels and if 64
`images are stored during a scan, then there will be approxi-
`mately 16,000 light intensity signals each 64 data points in
`length. In some implementations,
`the scan positions are
`
`0022
`
`0022
`
`

`

`US 2012/0140243 Al
`
`Jun. 7, 2012
`
`evenly spaced,i.e., the translation distance between succes-
`sive images captured by the image-recording device is the
`same.
`
`as a halogen bulb or metal halide lamp, with or without
`spectral bandpass filters; a broadband laser diode; a light-
`emitting diode; a combination of several light sources of the
`sameor different types; an arc lamp; any source in the visible
`Furthermore, the SLM 106 spatially modulates the
`[0057]
`spectral region (about 400 to 700 nm), any source in the IR
`illumination pattern as the sample object 102 is translated,
`spectral region (about 0.7 to 300 um), any source in the UV
`such that the object 102 is illuminated with a different pattern
`spectral region (about 10 to 400 nm). For broadbandapplica-
`at different corresponding vertical positions of the scan. The
`tions, the source preferably has a net spectral bandwidth
`continuous modification of the illumination pattern during
`broader than 5% of the mean wavelength, or more preferably
`the scan results in a modulated light intensity signal at each
`greater than 10%, 20%, 30%, or even 50% ofthe mean wave-
`pixel of the image-recording device 114. After the data has
`length. The source may also include one or more diffuser
`been acquired, the computer 122 can processthe light inten-
`elements to increasethe spatial extent of the input light being
`sity signals in accordance with, for example, pattern match-
`emitted from the source. Examples of narrow band sources
`ing techniques, and output data indicative of a surface topog-
`includealaser or a broadband source in combination with a
`raphy of the sample object 102.
`narrowbandfilter.
`[0058] When a surface of the sample object 102 is located
`[0061] The source can be a spatially-extended source or a
`at the plane of best-focus of the objective 118, an in-focus
`point source. In some implementations,it is preferable to use
`imageofthe illumination pattern is formed on a surface ofthe
`a spatially-extended source (e.g., when the surface of an
`object 102. The in-focus image exhibits the highest contrast
`object being imaged is smooth) to avoid observing a high-
`achievable for a given illumination numerical aperture (NA)
`contrast pattern regardless of the position of the object with
`of the system 100. The intensity of the image of the object
`respect to the plane of best focus. The image-recording
`surface that is reflected back to the image-recording device
`devices 114, 214 can includeaplurality of detector elements,
`114 at each pixel is proportional to the product of the local
`e.g., pixels, arranged in at least one and more generally two
`object surface reflectivity and the local intensity of the pro-
`dimensions. Examples of image-recording devices include
`jected illumination pattern. In contrast, when the surface of
`digital cameras, multi-element charge coupled devices
`the object 102 is located away from the plane ofbest-focus of
`(CCDs) and complementary metal oxide semiconductor
`the objective lens 118, the image ofthe illumination pattern
`(CMOS)detectors. Other image-recording devices may be
`on the surface of the object 102 is blurred and thus exhibits
`used as well. In some implementations, one or more color
`reduced contrast. The image of the surface of the object 102
`filters can be includedin the system tofilter the wavelength of
`passing back through the objective lens 118 towards the
`light forming the imagesat the image-recording device. The
`image-recording device then is blurred again. The result is
`one or more colorfilters can be arranged to be an integral
`that the contrast of the projected illumination pattern, as seen
`componentofthe image-recording device or as separate from
`on the image-recording device, is a function of the vertical
`the image-recording device.
`displacementofthe object surface with respectto the plane of
`best-focus of the objective. Accordingly, the depth profiling
`[0062]
`In some implementations, the objective lenses 118,
`capability of the system 100 arises from the detection of the
`218 can be incorporated as components of any standard
`position of the object 102 for which the pattern contrast is a
`microscope. For example, the systems can include a micro-
`maximumat each pixel ofthe image-recording device 114, as
`scope configured for use with one or more different objective
`the object 102 is scanned through the best-focus plane.
`lenses, each providing a different magnification. The NA of
`the objective lenses 118, 218 can be about0.1 orgreater(e.g.,
`[0059]
`In some implementations, the SLM 106 can be used
`about0.2 or greater, about 0.3 or greater, about 0.4 or greater,
`in a transmission arrangement as opposedto reflection. For
`about0.5 or greater, about 0.6 or greater, about 0.7 or greater,
`example, FIG. 2 shows an exemplary system 200 that
`about 0.8 or greater, about 0.9 or greater, or about 0.95 or
`includes the same components as system 100, except that a
`greater). Other NA valuesare possible as well.
`single beam splitter 216 is used in the arrangement and aSLM
`is configured to transmit incident light rather than reflect the
`[0063] Examples of SLMsthat can be usedtoreflect light,
`incident light (computer 122 also is omitted for ease of view-
`similar to the arrangement shown in FIG.1, include liquid
`ing). In the example shownin FIG.2, the transmission SLM
`crystal on silicon (LCOS) devices or micromirror array
`206 modifies incident light 201 generated by an illumination
`devices, e.g., a digital micromirror device (DMD). To spa-
`source 204 to produce a binary or quasi-continuousillumina-
`tially modulate illumination patterns produced by a LCOS
`tion pattern. The illumination pattern produced by the SLM
`device, auser can electronically control, using the appropriate
`206 passes through a lens 208c andis incident on the beam
`hardware and software, the direction and amountof light
`splitter 216, where the pattern then is directed through an
`reflected by each pixel of the LCOS device. Likewise, to
`objective lens 218 and imagedontoa plane of best-focus for
`spatially modulate the illumination pattern produced by a
`the objective. The illumination pattern reflects off a sample
`DMD,a user can electronically control the direction and/or
`object 202 and passes back throughthe objective lens 218, the
`orientation of each individual mirror to vary the amount of
`beam splitter 216, a tube lens 220 andis finally detected by a
`light reflected at each mirror. Examples of SLMsthat can be
`image-recording device 214. Similar to the system 100, the
`used to transmit light, similar to the arrangement shown in
`illumination pattern produced by the SLM 206 can be spa-
`FIG.2 include liquid crystal device (LCD) modulators that
`tially modulated as the image data is acquired, such that
`can be electronically controlled to vary the amountoflight
`different light intensity patterns are produced at correspond-
`transmitted. Alternatively, in some implementations, trans-
`ing different scan positions of the object.
`mission SLMscaninclude intensity masks, such as gratings,
`[0060]
`In each of the foregoing examples, the illumination
`Ronchirulings, or other patterned surfaces that have regions
`sources 104, 204 can include, but are not limited to, spec-
`of varying degrees of absorption for the wavelength of inci-
`trally-broadband sources or narrow band sources. Examples
`dent light. To spatially modulate the illumination patterns
`of broadband sources include: an incandescent source, such
`produced by the transmission SLMs formed from intensity
`
`0023
`
`0023
`
`

`

`US 2012/0140243 Al
`
`Jun. 7, 2012
`
`masks, the SLMs can be mounted on a mechanical frame that
`provides controlled in-plane motion of the SLM. The motion
`ofthe frame can be provided using one or more mechanical or
`electronic actuators, including, for example, piezoelectric
`actuators, stepper-motors, or voice coils. Other SLMs can be
`used in transmission or reflection arrangements as well. In
`some implementations, an electronic controller may store in
`memory instructions for how to generate one or moreillumi-
`nation patterns. In some implementations, the SLM usedfor
`creating the projected pattern is programmable, such that the
`frequency content, orientation and spatial intensity distribu-
`tion ofa particular pattern can be optimized for measurement
`ofa given object. It should be notedthatthe variousstructured
`illumination systems described in this disclosure are not
`restricted to

This document is available on Docket Alarm but you must sign up to view it.


Or .

Accessing this document will incur an additional charge of $.

After purchase, you can access this document again without charge.

Accept $ Charge
throbber

Still Working On It

This document is taking longer than usual to download. This can happen if we need to contact the court directly to obtain the document and their servers are running slowly.

Give it another minute or two to complete, and then try the refresh button.

throbber

A few More Minutes ... Still Working

It can take up to 5 minutes for us to download a document if the court servers are running slowly.

Thank you for your continued patience.

This document could not be displayed.

We could not find this document within its docket. Please go back to the docket page and check the link. If that does not work, go back to the docket and refresh it to pull the newest information.

Your account does not support viewing this document.

You need a Paid Account to view this document. Click here to change your account type.

Your account does not support viewing this document.

Set your membership status to view this document.

With a Docket Alarm membership, you'll get a whole lot more, including:

  • Up-to-date information for this case.
  • Email alerts whenever there is an update.
  • Full text search for other cases.
  • Get email alerts whenever a new case matches your search.

Become a Member

One Moment Please

The filing “” is large (MB) and is being downloaded.

Please refresh this page in a few minutes to see if the filing has been downloaded. The filing will also be emailed to you when the download completes.

Your document is on its way!

If you do not receive the document in five minutes, contact support at support@docketalarm.com.

Sealed Document

We are unable to display this document, it may be under a court ordered seal.

If you have proper credentials to access the file, you may proceed directly to the court's system using your government issued username and password.


Access Government Site

We are redirecting you
to a mobile optimized page.





Document Unreadable or Corrupt

Refresh this Document
Go to the Docket

We are unable to display this document.

Refresh this Document
Go to the Docket