`a2) Patent Application Publication 10) Pub. No.: US 2012/0140243 Al
`
` Colonna de Lega (43) Pub. Date: Jun. 7, 2012
`
`
`US 20120140243A1
`
`Publication Classification
`
`(51)
`
`Int. Cl
`(2006.01)
`GO1B 11/24
`(52) U.S. C1. cc ceeeseseeesersereeceeceeceeceeceeeneenes 356/609
`(57)
`ABSTRACT
`Methods for forming a three-dimensional image ofa test
`object include directing light to a surface of best-focus of an
`imaging optic, where thelight has an intensity modulation in
`at least one direction in the surface of best-focus, scanning a
`test object relative to the imaging optic so that a surface ofthe
`measurement object passes through the surface ofbest-focus
`of the imaging optic as the test object is scanned, acquiring,
`for each of a series of positions of the test object during the
`scan, a single image of the measurement object using the
`imaging optic, in which the intensity modulation ofthe light
`in the surface ofbest-focusis different for successive images,
`and forming a three-dimensional image of the test object
`
`(54) NON-CONTACT SURFACE
`CHARACTERIZATION USING MODULATED
`ILLUMINATION
`
`(75)
`
`Inventor:
`
`Xavier M. Colonna de Lega,
`Middlefield, CT (US)
`
`(73) Assignee:
`
`Zygo Corporation, Middfield, CT
`(US)
`
`(21) Appl. No.:
`
`13/309,244
`
`(22)
`
`Filed:
`
`Dec. 1, 2011
`
`Related U.S. Application Data
`
`(60) Provisional application No. 61/419,386, filed on Dec.
`3, 2010.
`
`based on the acquired images.
`
`Align Ex. 1017
`U.S. Patent No. 9,962,244
`
`0001
`
`Align Ex. 1017
`U.S. Patent No. 9,962,244
`
`0001
`
`
`
`Patent Application Publication
`
`Jun. 7,2012 Sheet 1 of 18
`
`US 2012/0140243 Al
`
`
`
`0002
`
`0002
`
`
`
`Patent Application Publication
`
`Jun. 7,2012 Sheet 2 of 18
`
`US 2012/0140243 Al
`
`
`
`0003
`
`0003
`
`
`
`Patent Application Publication
`
`Jun. 7,2012 Sheet 3 of 18
`
`US 2012/0140243 Al
`
`318
`
`302
`
`
`
`
`
`Scan
`
`
`
`(e) }
`
`:
`
`FIG. 3
`
`0004
`
`0004
`
`
`
`Patent Application Publication
`
`Jun. 7,2012 Sheet 4 of 18
`
`US 2012/0140243 Al
`
`
`
`0005
`
`0005
`
`
`
`Patent Application Publication
`
`Jun. 7,2012 Sheet 5 of 18
`
`US 2012/0140243 Al
`
`
`
`FIG. 6
`
`0006
`
`0006
`
`
`
`Patent Application Publication
`
`US 2012/0140243 Al
`
`Jun. 7,2012 Sheet 6 of 18
`
`FIG.7
`
`0007
`
`0007
`
`
`
`Patent Application Publication
`
`Jun. 7,2012 Sheet 7 of 18
`
`US 2012/0140243 Al
`
`08‘Sls a8‘Sis V8‘Old
`
`0008
`
`0008
`
`
`
`Patent Application Publication
`
`Jun. 7,2012 Sheet 8 of 18
`
`US 2012/0140243 Al
`
`Spectral magnitude
`
`
`
`Frequency
`FIG. 9A
`
`Spectral phase
`
`Frequency
`
`FIG. 9B
`
`0009
`
`0009
`
`
`
`Patent Application Publication
`
`US 2012/0140243 Al
`
`Jun. 7,2012 Sheet 9 of 18
`
`d0b‘Sls
`
`0010
`
`0010
`
`
`
`US 2012/0140243 Al
`
`\
`
`\SS
`
`yy
`
`ie
`
`esGYyy
`esiyLie
`
`Jun. 7,2012 Sheet 10 of 18
`
`Yue
`
`Patent Application Publication
`
`S
`
`SSS
`SS
`\
`AX
`
`0011
`
`0011
`
`
`
`Patent Application Publication
`
`Jun. 7,2012 Sheet 11 of 18
`
`US 2012/0140243 Al
`
`
`
`0012
`
`0012
`
`
`
`Patent Application Publication
`
`Jun. 7,2012 Sheet 12 of 18
`
`US 2012/0140243 Al
`
`1402
`
`LO
`com)
`wzel
`
`o/©=
`¢
`
`FIG.14
`
`t
`I
`t
`I
`
`t
`!
`|
`1
`
`1403
`
`1404
`
`Fg
`
`JANE
`YY
`
`1
`
`Ww
`
`{O
`
`g |—_re
`
`—tee
`w_oo
`
`1416
`
`0013
`
`0013
`
`
`
`Patent Application Publication
`
`Jun. 7,2012 Sheet 13 of 18
`
`US 2012/0140243 Al
`
` cOSL
`
`OOSL
`
`0014
`
`0014
`
`
`
`Patent Application Publication
`
`Jun. 7,2012 Sheet 14 of 18
`
`US 2012/0140243 Al
`
`
`
`Vv
`a
`
`Oo
`ra)
`Sre
`
`3
`So
`—
`
`CO
`ere
`
`ae)
`3
`©
`ve
`
`N O
`
`o©e
`
`r
`
`
`en1 tga i
`S
`rir dia’
`o
`
`
`
`
`3 fFoekyr tt tog wo
`
`O
`=_—
`
`
`
`~ eya
`
`1606
`
`1614
`
`0015
`
`0015
`
`
`
`Patent Application Publication
`
`Jun. 7,2012 Sheet 15 of 18
`
`US 2012/0140243 Al
`
`dZbSls
`
`0016
`
`0016
`
`
`
`Patent Application Publication
`
`Jun. 7,2012 Sheet 16 of 18
`
`US 2012/0140243 Al
`
`cO8l
`
`aslOld
`
`cOst
`
`V8-Sls
`
`cO8L
`
`
`
`0017
`
`0017
`
`
`
`
`Patent Application Publication
`
`Jun. 7,2012 Sheet 17 of 18
`
`US 2012/0140243 Al
`
`wi
`
`61‘Old
`
`O61
`Vlas
`
`0018
`
`0018
`
`
`
`
`
`Patent Application Publication
`
`Jun. 7,2012 Sheet 18 of 18
`
`US 2012/0140243 Al
`
`2000
`
`Scan
`
` KWfNEY
`
`foNIY
`
`FIG. 20
`
`0019
`
`0019
`
`
`
`US 2012/0140243 Al
`
`Jun. 7, 2012
`
`NON-CONTACT SURFACE
`CHARACTERIZATION USING MODULATED
`ILLUMINATION
`
`CROSS REFERENCE TO RELATED
`APPLICATIONS
`
`[0001] This application claims benefit ofprovisional patent
`application No. 61/419,386, entitled “Non-Contact Surface
`Characterization Using Modulated Hlumination,” filed on
`Dec. 3, 2010, the entire contents of which are hereby incor-
`porated by reference.
`
`TECHNICAL FIELD
`
`[0002] This disclosure relates to methods of non-contact
`surface characterization using modulated illumination and
`devices configured to perform the same.
`
`BACKGROUND
`
`[0003] The field of non-contact surface characterization
`includes characterization and/or measurementof information
`
`about an object’s surface including, for example, shape and
`surface finish measurements. Tools capable of non-contact
`surface characterization can be useful in manufacturing dur-
`ing stages such as process development, quality control and
`process control on the production floor. Non-contact surface
`characterization methods can include obtaining information
`about the three-dimensional (3D) shape and location of an
`object by analyzing the frequency content of an imageofthe
`object in an optical system.
`
`SUMMARY
`
`Ingeneral, the subject matter described in this speci-
`[0004]
`fication relates to methods and systems for obtaining infor-
`mation about a surface of an object. The methods disclosed
`can include, for example, directing a structured illumination
`pattern to a surface of best-focus of an imaging optic while
`imaging a field in the surface of best-focus onto a multi-
`element detector. A structured illumination pattern is a non-
`uniform illumination pattern that contains some form of
`encoding based on light intensity. Examples of structured
`illumination patterns includeperiodicintensity patterns (e.g.,
`modulated in one or two dimensions). An object is scanned
`through the surface of best-focus while a series of images of
`the illumination patterns on the object surface are sequen-
`tially acquired by the multi-element detector. For each suc-
`cessive image obtained, the illumination pattern is modified
`so that its intensity at each point in the field corresponding to
`a detector element is modulated from one frameto the next.
`
`The resulting signal acquired at each detector element has a
`modulation, with the maximum amplitude occurring at or
`near the position of the object surface as it intersects the
`surface ofbest-focusofthe objective. Based on an analysis of
`the signals provided by the sequence of images, measure-
`ments of topography and surface texture of the object can be
`obtained and, in some implementations, a three-dimensional
`imageof the object’s surface can be produced.
`[0005]
`In some embodiments, a detector plane and pattern-
`generating plane are mapped onto a surface in object space
`that is nominally conformal to the shape of the object to be
`characterized including, for example, a plane, a sphere, a
`parabola, a cylinder, a cone, or an aspheric optic.
`[0006]
`In certain embodiments, the signal recorded at a
`pixel(1.e., element) on the detector has a modulation envelope
`
`that emulates some of the characteristics of a scanning low-
`coherence interferometer (SWLI) signal. Conversion of scan
`data to topography or reflectivity data may therefore be
`accomplished by application of envelope-detection algo-
`rithms developed for SWLI such as a frequency domain
`analysis (“FDA”) or least squares analysis (“LSQ”).
`[0007]
`In some embodiments, the device used for creating
`the projected illumination pattern is programmable, which
`allows adapting the frequency content, orientation andspatial
`intensity distribution to optimize the measurement capability
`for a given object.
`[0008] The structured illumination modulation scheme can
`be optimized to enable a fast autofocus scan of an optical
`system. An application in the context of laser eye surgery is
`the localization of the position of laser optics with respect to
`a critical component that makes contact with the cornea.
`Another application is rapid focusing of a low-coherence
`interferometer,
`in which case the autofocus scan rate is
`selected to average out the interference signal from the
`detected signal.
`[0009]
`In certain embodiments, the apparatus presents to
`the user an enhanced imageofthe object that combines height
`information with additional surface information, such as
`color, absorption, texture,etc.
`[0010] Various aspects of the subject matter described in
`this specification are summarizedas follows.
`[0011]
`In general, one aspect of the subject matter
`described in this specification can be embodied in methods
`for forming a three-dimensional image of a test object, in
`which the methods include directing light to a surface of
`best-focus of an imaging optic, where the light has an inten-
`sity modulation in at least one direction in the surface of
`best-focus, scanning a test object relative to the imaging optic
`so that a surface ofthe measurementobjectpasses through the
`surface of best-focus of the imaging optic as the test objectis
`scanned, acquiring, for each of a series ofpositionsofthe test
`object during the scan, a single image of the measurement
`object using the imaging optic, in which the intensity modu-
`lation of the light in the surface of best-focus is different for
`successive images, and forming a three-dimensional image of
`the test object based on the acquired images.
`[0012] These and other embodiments can each optionally
`include one or more ofthe following features. For example,in
`some implementations, directing the light to the surface of
`best-focus includes imaging a spatial light modulator (SLM)
`to the surface of best-focus. The intensity modulation of the
`light in the surface of best-focus can be varied using the
`spatial light modulator.
`[0013]
`In some implementations, directing the light to the
`surface of best-focus includes imaging a pattern-generating
`plane onto a surface in object space. The surface can be
`conformal to a shape of the test object. The shapeof the test
`object can be planar, spherical, parabolic, cylindrical, coni-
`cal, or aspheric.
`[0014]
`In some implementations, the intensity modulation
`is a periodic modulation. The periodic modulation can be a
`sinusoidal modulation. The phaseofthe periodic modulation
`can be varied by less than 27 between each successive image.
`The phase of the periodic modulation can be varied by x or
`less between each successive image. The phase ofthe peri-
`odic modulation can be varied by 1/2 between each succes-
`sive image.
`[0015]
`In some implementations, the intensity modulation
`is a two-dimensional intensity modulation. The scan posi-
`
`0020
`
`0020
`
`
`
`US 2012/0140243 Al
`
`Jun. 7, 2012
`
`In some implementations, the SLM can be a liquid
`[0023]
`crystal panel, or include a micro-mirrorarray.
`[0024]
`In some implementations, the imaging optic has a
`numerical aperture greater than 0.6, a numerical aperture
`greater than 0.8, a numerical aperture greater than 0.9, or a
`numerical aperture of 0.95.
`the system further
`[0025]
`In some implementations,
`includescolorfilters arranged to filter the wavelength of light
`forming the images at the multi-element detector.
`[0026] The one or more optical elements can include, for
`example, a fisheye lens, an endoscope, and/or a zoom lens.
`[0027]
`Implementations disclosed herein can offer several
`advantages. For example,
`in some implementations,
`the
`methods and apparatus can be used to provide non-contact
`three-dimensional imaging of an object. The imaging can be
`performed using short measurement times and/or to achieve
`high resolution images of objects’ surfaces. In some imple-
`mentations, the imaging is performed with reduced sensitiv-
`ity to environmental perturbations such as, for example,
`vibration or acoustic noise.
`
`arranged to direct light from the spatial light modulator to
`tions can be evenly spaced. The intensity modulation can be
`form an image of the SLM atthe surface of best-focus during
`selected based on a slope ofthe test object surface.
`operation ofthe system, a scanning stage arrangedto scan the
`[0016]
`Insome implementations, forming the three-dimen-
`test object relative to the microscope object during operation
`sional image includes identifying, for multiple different loca-
`of the system so that a surface ofthe test object intersects the
`tionsofthe test object surface, the scan position correspond-
`surface of best-focus, a multi-element detector positioned
`ing to where eachlocationintersects the surface ofbest-focus.
`relative to the microscope such that the microscope forms an
`An intensity of the acquired images as a function of scan
`image ofa field at the surface of best-focus on the multi-
`position at each of the different locations can include an
`element detector during operation of the system, and an elec-
`amplitude modulation. Identifying the scan position corre-
`tronic control module in communication with the scanning
`sponding to where each location intersects the surface of
`stage,
`the spatial
`light modulator, and the multi-element
`best-focus can include identifying the scan position where a
`detector, in which during operation, the system causes the
`modulation amplitudeis largest.
`multi-element detector to acquire a single image ofthe test
`[0017]
`Insome implementations, forming the three-dimen-
`object for each of multiple scan positions of the test object
`sional image includesderiving an intensity signal for each of
`relative to the imaging optic, causes the SLM to variably
`multiple different locations of the test object surface, each
`modulate the intensity of light at the surface of best-focus in
`intensity signal corresponding to the intensity of the acquired
`at least one direction so that the intensity modulation of the
`images at the corresponding location as a function of scan
`light is different for successive images, and formsa three-
`position. Forming the three-dimensional image can include
`dimensional image of the test object based on the acquired
`identifying a scan position corresponding to a maximum
`images.
`amplitude of a modulation of the intensity signal for each
`[0022] This and other embodiments can each optionally
`location. Identifying the scan position can include transform-
`include one or moreofthe following features. In some imple-
`ing each intensity signal into a frequency domain.
`mentations, the system includes, for example, a light source
`[0018]
`In some implementations, the test object is a lens
`arranged to direct light to the SLM during operation of the
`element. The three-dimensional image can be, for example, a
`system. The SLM can beareflective SLM ora transmissive
`monochrome image. Alternatively,
`the three-dimensional
`SLM.
`image can be a color image.
`[0019]
`In general, another aspect of the subject matter
`described in this specification can be embodied in methods
`for forming a three-dimensional image of a test object, in
`which the methodseach include forming an imageofa spatial
`light modulatorat a surface ofbest-focus of an imagingoptic,
`scanning a test object relative to the imaging optic so that a
`surface ofthe measurementobject passes through a surface of
`best-focus of the imaging optic as the test object is scanned,
`acquiring, for each of a series of positions of the test object
`during the scan, a single imageof the test object using the
`imaging optic, in which the spatial light modulator varies an
`intensity modulationin the light forming the image sothat the
`modulation ofthe light at the surface ofbest-focusis different
`for successive images, and forming a three-dimensional
`imageofthe test object based on the acquired images.
`[0020]
`In general, another aspect of the subject matter
`described in this specification can be embodied in methods
`for forming a three-dimensional image of a test object, in
`which the methodseach includedirecting light to a surface of
`best-focus of an imaging optic, where the light has an inten-
`sity modulation in at least one direction in the surface of
`best-focus, scanning a test object relative to the imaging optic
`so that a surface of the measurementobject passes through a
`surface of best-focus of the imaging optic as the test objectis
`scanned, imaging the surface ofbest-focus to a multi-element
`detector, acquiring, for each ofa series ofpositions ofthe test
`object during the scan, a single intensity measurementat one
`or more elements of the multi-element detector, in which the
`intensity modulationofthe light in the surface ofbest-focusis
`different for successive positions of the test object during the
`scan, and forming a three-dimensional image of the test
`object based on the acquired intensity measurements.
`[0021]
`In general, another aspect of the subject matter
`describedin this specification can be embodied in systems for
`forming a three-dimensional imageofa test object, in which
`the systems each include a microscope including an imaging
`optic, the imaging optic having a surface of best-focus, a
`spatial
`light modulator, one or more optical elements
`
`Implementationsoffer other advantagesas well. For
`[0028]
`example, the methods andstructures disclosed herein can, in
`some implementations, provide similar depth sectioning
`capabilities as a conventional confocal microscope. The use
`of structured illumination profiling can offer certain benefits
`compared to confocal microscopy (e.g., reduced source
`brightness requirements) and interference microscopy. For
`example, in some implementations, structured illumination
`has reduced source brightness requirements relative to con-
`ventional interference microscopes. Alternatively, or in addi-
`tion, restrictions on which non-interferometric microscope
`objectives can be used can be reduced. For example, in some
`implementations, higher numerical aperture objectives are
`available relative to the objectives available in Michelson or
`Mirau interferometers.
`
`[0029] The details of one or more embodiments are set
`forth in the accompanying drawings and the description
`
`0021
`
`0021
`
`
`
`US 2012/0140243 Al
`
`Jun. 7, 2012
`
`below. Other features and advantages will be apparent from
`the description, drawings, and from the claims.
`
`BRIEF DESCRIPTION OF DRAWINGS
`
`FIG. 11s a schematic diagram of an exemplary sys-
`[0030]
`tem for providing a 3D image of a sample object.
`[0031]
`FIG. 21s a schematic diagram of an exemplary sys-
`tem for providing a 3D image of a sample object.
`[0032]
`FIG. 3 is a schematic illustrating various scan posi-
`tions of a sample object.
`[0033]
`FIG. 4isa graph ofa simulatedlightintensity signal.
`[0034] FIG.5is a graph of an experimentalintensity signal
`recorded by a pixel of an image-recording device.
`[0035]
`FIG. 6 is an exemplary topography map of a 13-um
`step height step generated by FDA processingoflight inten-
`sity signals.
`[0036]
`FIG. 7 is a graph of a simulated intensity signal.
`[0037]
`FIGS. 8A-8Bare plots of a simulated analog inten-
`sity signal overlaid with a corresponding digitized sampling
`of the intensity signal.
`[0038]
`FIG. 8C is a plot of the simulated analog intensity
`signal of FIG. 8B overlaid with the corresponding digitized
`intensity signal of FIG. 8B, where zeros have been inserted
`into the digitized intensity signal.
`[0039]
`FIG. 9A is a plot of the magnitude ofthe digitized
`sample signal from FIG. 8A and the phase of the digitized
`intensity signal from FIG. 8C.
`[0040] FIG.9B isa plotofthe phaseofthe digitized sample
`signal from FIG. 8A and the magnitudeofthe digitized inten-
`sity signal from FIG. 8C.
`[0041]
`FIG. 10A is a plot of a simulated analog intensity
`signal overlaid with a correspondingdigitized sampling ofthe
`simulated intensity signal.
`[0042]
`FIG. 10B is a plot of the simulated analog intensity
`signal of FIG. 10A overlaid with the digitized intensity signal
`of FIG. 10A, where zero values have been inserted into the
`digitized intensity signal.
`[0043]
`FIGS. 11-13 are exemplary images of a microlens
`onto which an illumination pattern is projected.
`[0044]
`FIG. 14 is a schematic diagram of an exemplary
`structured illumination imaging system.
`[0045]
`FIG. 15 is a schematic diagram of an exemplary
`imaging system.
`[0046]
`FIG. 16 is a schematic diagram of an exemplary
`imaging system.
`[0047]
`FIGS. 17A-17B are schematic diagrams of exem-
`plary imaging systems.
`[0048]
`FIGS. 18A-18B are schematic diagrams of exem-
`plary imaging systems.
`[0049] FIG.19A is a3D graphof experimental topography
`data collected on a sample object.
`[0050]
`FIG. 19B is a color image of the surface of the
`sample object of FIG. 19A.
`[0051]
`FIG. 19C is an exemplary color image produced by
`combining the topographical data of FIG. 19A with the color
`image of FIG. 19B.
`[0052]
`FIG. 20 is a schematic of an exemplary system for
`performing laser eye surgery.
`
`DETAILED DESCRIPTION
`
`Acquiring the Light Intensity Signal
`[0053]
`FIG. 1 is a schematic diagram of an exemplary
`structured illumination system 100 for providing a 3D image
`
`of a sample object 102. An illumination portion 103 of the
`system 100 includes an illumination source 104, a spatial
`light modulator (SLM) 106, one or more lenses (108a-108c),
`an aperture 110 and a beam combiner 112. An imagingpor-
`tion 105 of the system 100 includes an image-recording
`device 114, a beam splitter 116, an objective lens 118 and a
`tube lens 120. For the purposes of this disclosure, the coor-
`dinate system is defined such that the z-axis is parallel to the
`optical axis of the objective lens 118 and the x and y-axes are
`parallel to the lens’ plane ofbest-focus such that xyz forms an
`orthogonal coordinate system.
`[0054] During operation of the system 100, light 101 gen-
`erated from the illumination source 104 passes through the
`lenses 108a, 1084 and the aperture 110, wherethe light 101
`then is incident on the beam splitter 112. Beam splitter 112
`reflects a portion of the incidentlight 101 onto the SLM 106.
`In the present example, the SLM 106 is configured to modify
`and reflect the light incidenton its surface so as to produce an
`illumination pattern characterized as either binary (i.e., the
`imaging pattern has regions wherelight is present and regions
`wherelight is absent) or quasi-continuous(i.e., the imaging
`pattern can be approximated by continuous varying levels of
`light intensity).
`[0055] The illumination pattern reflected from the SLM
`106 then passes through the beam splitter 112 and lens 108c
`to a second beam splitter 116 where the illumination pattern
`thenis directed through the objective lens 118 and preferen-
`tially fills the back pupil of the objective 118. The illumina-
`tion pattern is re-imaged in object space to the plane ofbest-
`focus (focal plane) of the objective lens 118. Lightreflecting
`and/or scattering off a surface of the sample object 102 then
`proceeds through the objective lens 118, the beam splitter
`116, and the tube lens 120 onto a light-detecting surface/plane
`of the image-recording device 114 whereit is recorded. The
`recordedlight thus acquired can be storedin digital format as
`an array of light intensity signals, with each light intensity
`signal being acquired from a corresponding pixel of the
`image-recording device 114.
`[0056] During imaging, the object 102 is translated verti-
`cally with respect to the objective lens 118 (i.e., toward or
`away from the objective lens 118). The sample object 102 can
`be displaced or actuated by an electromechanical transducer
`(not shown) andassociated drive electronics controlled by a
`computer 122 so as to enable precise scans along a direction
`of translation of the object 102. Examples of transducers
`include, but are not
`limited to, piezoelectric transducers
`(PZT), stepper motors and voice coils. Alternatively, or in
`addition, the objective lens 118 maybetranslated vertically
`with respectto a position of the sample object 102. Again, an
`electromechanical transducer and associated drive electron-
`ics may be usedto controlthe translation. The image-record-
`ing device 114 simultaneously records imaging data as the
`object 102 is scanned through the plane of best-focus of the
`objective lens 118 such that multiple light intensity signals
`will be recorded overtime. That is, an imageofthe object and
`illumination pattern is captured by the image-recording
`device in the form ofthe light intensity signal at correspond-
`ing scan positions. For example, if the image-recording
`device 114 includes a 128x128 array of pixels and if 64
`images are stored during a scan, then there will be approxi-
`mately 16,000 light intensity signals each 64 data points in
`length. In some implementations,
`the scan positions are
`
`0022
`
`0022
`
`
`
`US 2012/0140243 Al
`
`Jun. 7, 2012
`
`evenly spaced,i.e., the translation distance between succes-
`sive images captured by the image-recording device is the
`same.
`
`as a halogen bulb or metal halide lamp, with or without
`spectral bandpass filters; a broadband laser diode; a light-
`emitting diode; a combination of several light sources of the
`sameor different types; an arc lamp; any source in the visible
`Furthermore, the SLM 106 spatially modulates the
`[0057]
`spectral region (about 400 to 700 nm), any source in the IR
`illumination pattern as the sample object 102 is translated,
`spectral region (about 0.7 to 300 um), any source in the UV
`such that the object 102 is illuminated with a different pattern
`spectral region (about 10 to 400 nm). For broadbandapplica-
`at different corresponding vertical positions of the scan. The
`tions, the source preferably has a net spectral bandwidth
`continuous modification of the illumination pattern during
`broader than 5% of the mean wavelength, or more preferably
`the scan results in a modulated light intensity signal at each
`greater than 10%, 20%, 30%, or even 50% ofthe mean wave-
`pixel of the image-recording device 114. After the data has
`length. The source may also include one or more diffuser
`been acquired, the computer 122 can processthe light inten-
`elements to increasethe spatial extent of the input light being
`sity signals in accordance with, for example, pattern match-
`emitted from the source. Examples of narrow band sources
`ing techniques, and output data indicative of a surface topog-
`includealaser or a broadband source in combination with a
`raphy of the sample object 102.
`narrowbandfilter.
`[0058] When a surface of the sample object 102 is located
`[0061] The source can be a spatially-extended source or a
`at the plane of best-focus of the objective 118, an in-focus
`point source. In some implementations,it is preferable to use
`imageofthe illumination pattern is formed on a surface ofthe
`a spatially-extended source (e.g., when the surface of an
`object 102. The in-focus image exhibits the highest contrast
`object being imaged is smooth) to avoid observing a high-
`achievable for a given illumination numerical aperture (NA)
`contrast pattern regardless of the position of the object with
`of the system 100. The intensity of the image of the object
`respect to the plane of best focus. The image-recording
`surface that is reflected back to the image-recording device
`devices 114, 214 can includeaplurality of detector elements,
`114 at each pixel is proportional to the product of the local
`e.g., pixels, arranged in at least one and more generally two
`object surface reflectivity and the local intensity of the pro-
`dimensions. Examples of image-recording devices include
`jected illumination pattern. In contrast, when the surface of
`digital cameras, multi-element charge coupled devices
`the object 102 is located away from the plane ofbest-focus of
`(CCDs) and complementary metal oxide semiconductor
`the objective lens 118, the image ofthe illumination pattern
`(CMOS)detectors. Other image-recording devices may be
`on the surface of the object 102 is blurred and thus exhibits
`used as well. In some implementations, one or more color
`reduced contrast. The image of the surface of the object 102
`filters can be includedin the system tofilter the wavelength of
`passing back through the objective lens 118 towards the
`light forming the imagesat the image-recording device. The
`image-recording device then is blurred again. The result is
`one or more colorfilters can be arranged to be an integral
`that the contrast of the projected illumination pattern, as seen
`componentofthe image-recording device or as separate from
`on the image-recording device, is a function of the vertical
`the image-recording device.
`displacementofthe object surface with respectto the plane of
`best-focus of the objective. Accordingly, the depth profiling
`[0062]
`In some implementations, the objective lenses 118,
`capability of the system 100 arises from the detection of the
`218 can be incorporated as components of any standard
`position of the object 102 for which the pattern contrast is a
`microscope. For example, the systems can include a micro-
`maximumat each pixel ofthe image-recording device 114, as
`scope configured for use with one or more different objective
`the object 102 is scanned through the best-focus plane.
`lenses, each providing a different magnification. The NA of
`the objective lenses 118, 218 can be about0.1 orgreater(e.g.,
`[0059]
`In some implementations, the SLM 106 can be used
`about0.2 or greater, about 0.3 or greater, about 0.4 or greater,
`in a transmission arrangement as opposedto reflection. For
`about0.5 or greater, about 0.6 or greater, about 0.7 or greater,
`example, FIG. 2 shows an exemplary system 200 that
`about 0.8 or greater, about 0.9 or greater, or about 0.95 or
`includes the same components as system 100, except that a
`greater). Other NA valuesare possible as well.
`single beam splitter 216 is used in the arrangement and aSLM
`is configured to transmit incident light rather than reflect the
`[0063] Examples of SLMsthat can be usedtoreflect light,
`incident light (computer 122 also is omitted for ease of view-
`similar to the arrangement shown in FIG.1, include liquid
`ing). In the example shownin FIG.2, the transmission SLM
`crystal on silicon (LCOS) devices or micromirror array
`206 modifies incident light 201 generated by an illumination
`devices, e.g., a digital micromirror device (DMD). To spa-
`source 204 to produce a binary or quasi-continuousillumina-
`tially modulate illumination patterns produced by a LCOS
`tion pattern. The illumination pattern produced by the SLM
`device, auser can electronically control, using the appropriate
`206 passes through a lens 208c andis incident on the beam
`hardware and software, the direction and amountof light
`splitter 216, where the pattern then is directed through an
`reflected by each pixel of the LCOS device. Likewise, to
`objective lens 218 and imagedontoa plane of best-focus for
`spatially modulate the illumination pattern produced by a
`the objective. The illumination pattern reflects off a sample
`DMD,a user can electronically control the direction and/or
`object 202 and passes back throughthe objective lens 218, the
`orientation of each individual mirror to vary the amount of
`beam splitter 216, a tube lens 220 andis finally detected by a
`light reflected at each mirror. Examples of SLMsthat can be
`image-recording device 214. Similar to the system 100, the
`used to transmit light, similar to the arrangement shown in
`illumination pattern produced by the SLM 206 can be spa-
`FIG.2 include liquid crystal device (LCD) modulators that
`tially modulated as the image data is acquired, such that
`can be electronically controlled to vary the amountoflight
`different light intensity patterns are produced at correspond-
`transmitted. Alternatively, in some implementations, trans-
`ing different scan positions of the object.
`mission SLMscaninclude intensity masks, such as gratings,
`[0060]
`In each of the foregoing examples, the illumination
`Ronchirulings, or other patterned surfaces that have regions
`sources 104, 204 can include, but are not limited to, spec-
`of varying degrees of absorption for the wavelength of inci-
`trally-broadband sources or narrow band sources. Examples
`dent light. To spatially modulate the illumination patterns
`of broadband sources include: an incandescent source, such
`produced by the transmission SLMs formed from intensity
`
`0023
`
`0023
`
`
`
`US 2012/0140243 Al
`
`Jun. 7, 2012
`
`masks, the SLMs can be mounted on a mechanical frame that
`provides controlled in-plane motion of the SLM. The motion
`ofthe frame can be provided using one or more mechanical or
`electronic actuators, including, for example, piezoelectric
`actuators, stepper-motors, or voice coils. Other SLMs can be
`used in transmission or reflection arrangements as well. In
`some implementations, an electronic controller may store in
`memory instructions for how to generate one or moreillumi-
`nation patterns. In some implementations, the SLM usedfor
`creating the projected pattern is programmable, such that the
`frequency content, orientation and spatial intensity distribu-
`tion ofa particular pattern can be optimized for measurement
`ofa given object. It should be notedthatthe variousstructured
`illumination systems described in this disclosure are not
`restricted to