throbber
(19) United States
`(12) Patent Application Publication (10) Pub. No.: US 2012/0140243 A1
`
` Colonna de Lega (43) Pub. Date: Jun. 7, 2012
`
`
`US 20120140243A1
`
`(54) NON-CONTACT SURFACE
`CHARACTERIZATION USING MODULATED
`ILLUMINATION
`
`(75)
`
`Inventor:
`
`Xavier M. Colonna de Lega,
`Middlefield, CT (US)
`
`(73) Assignee:
`
`Zygo Corporation, Middfield, CT
`(US)
`
`(21) Appl. NO’:
`
`13/309’244
`
`(22)
`
`Filed:
`
`Dec. 1, 2011
`
`Related US. Application Data
`
`(60) Provisional application No. 61/419,386, filed on Dec.
`3, 2010.
`
`Publication Classification
`
`(51)
`
`Int Cl
`(2006.01)
`G013 11/24
`(52) US. Cl. ........................................................ 356/609
`
`ABSTRACT
`(57)
`Methods for forming a three-dimensional image of a test
`object include directing light to a surface of best-focus of an
`imaging optic, where the light has an intensity modulation in
`at least one direction in the surface of best-focus, seaming a
`test object relative to the imaging optic so that a surface ofthe
`measurement object passes through the surface ofbest-focus
`of the imaging optic as the test object is scanned, acquiring,
`for each of a series of positions of the test object during the
`scan, a single image of the measurement object using the
`imaging optic, in which the intensity modulation of the light
`in the surface ofbest-focus is different for successive images,
`and forming a three-dimensional image of the test object
`based on the acquired images.
`
`7—‘7
`
`[A
`‘II'iEEI
`‘gllgifl
`
`Align EX. 1017
`US. Patent No. 9,962,244
`
`0001
`
`Align Ex. 1017
`U.S. Patent No. 9,962,244
`
`0001
`
`

`

`Patent Application Publication
`
`Jun. 7, 2012 Sheet 1 of 18
`
`US 2012/0140243 A1
`
`‘ggllglfl
`
`A‘:II§IIII'
`‘Iln'k
`‘ll'nlmgll
`
`0002
`
`0002
`
`

`

`Patent Application Publication
`
`Jun. 7, 2012 Sheet 2 of 18
`
`US 2012/0140243 A1
`
`
`
` IE\
`
`0003
`
`0003
`
`

`

`Patent Application Publication
`
`Jun. 7, 2012 Sheet 3 of 18
`
`US 2012/0140243 A1
`
`318
`
`
`
`
`
`
`
`
`
`
`
`/-\ (D \_/
`
`FIG. 3
`
`0004
`
`0004
`
`

`

`Patent Application Publication
`
`Jun. 7, 2012 Sheet 4 of 18
`
`US 2012/0140243 A1
`
`
`
`
`
`.,A.I.a=====§i._EEEE
`
`Miii—.01
`iE—«l lug—ail
`Iii—ml
`
`»..,Egg—=5
`
`0005
`
`0005
`
`

`

`Patent Application Publication
`
`Jun. 7, 2012 Sheet 5 of 18
`
`US 2012/0140243 A1
`
`
`
`FIG. 6
`
`0006
`
`0006
`
`

`

`Patent Application Publication
`
`Jun. 7, 2012 Sheet 6 of 18
`
`US 2012/0140243 A1
`
`i,
`
`,_
`
`i.A,
`
`i.
`
`.i
`i,
`
`
`
`<
`
`——-
`
`)
`
`c _
`
`-
`————
`—--
`
`FIG.7
`
`,lllllllllJi_. ill-n-
`
`.___ Vs
`
`——-——-
`_—————
`-——-—— 3
`
`
`‘
`1
`
`__
`,—
`fill-l-
`_-
`’»
`
`0007
`
`0007
`
`

`

`Patent Application Publication
`
`Jun. 7, 2012 Sheet 7 of 18
`
`US 2012/0140243 A1
`
`
`
`mm.67..
`
`Ow.OE
`
`
`
`0008
`
`0008
`
`

`

`Patent Application Publication
`
`Jun. 7, 2012 Sheet 8 of 18
`
`US 2012/0140243 A1
`
`Spectral magnitude
`
`
`
`Frequency
`
`FIG. 9A
`
`Spectral phase
`
`180___-
`90l
`
`
`
`IVAYA‘_——
`
`Frequency
`
`FIG. QB
`
`0009
`
`0009
`
`

`

`Patent Application Publication
`
`Jun. 7, 2012 Sheet 9 of 18
`
`US 2012/0140243 A1
`
`mow.®_n_
`
` <0?.07..
`
`0010
`
`0010
`
`

`

`Patent Application Publication
`
`Jun. 7, 2012 Sheet 10 of 18
`
`US 2012/0140243 A1
`
`//k 134%4%//A//4
`
`
`
`\\
`\\ x
`
`/%/3.?.//
`
`0011
`
`0011
`
`

`

`Patent Application Publication
`
`Jun. 7, 2012 Sheet 11 of 18
`
`US 2012/0140243 A1
`
`
`
`0012
`
`0012
`
`

`

`Patent Application Publication
`
`Jun. 7, 2012 Sheet 12 of 18
`
`US 2012/0140243 A1
`
`1402
`
`I
`
`M’’4’
`
`”rl’’//z”’’//,x/r’II//’zz,r/’’If,’l,z’2’z’I’l’l’l”//Ir,”
`
`1405
`
`1406
`
`1414
`
`1400
`
`1403
`
`1404
`
`1416
`
`0013
`
`0013
`
`

`

`Patent Application Publication
`
`Jun. 7, 2012 Sheet 13 0f 18
`
`US 2012/0140243 A1
`
`
`
`0014
`
`0014
`
`

`

`Patent Application Publication
`
`Jun. 7, 2012 Sheet 14 of 18
`
`US 2012/0140243 A1
`
`1602
`
`(:3
`03
`(0V"
`
`1605
`
`E‘EE
`
`1606
`
`1614
`
`1616
`
`0015
`
`FIG.16
`
`1603
`
`1604
`
`0015
`
`

`

`Patent Application Publication
`
`Jun. 7, 2012 Sheet 15 of 18
`
`US 2012/0140243 A1
`
`mt
`
`.mVE
`
`0016
`
`0016
`
`

`

`Patent Application Publication
`
`Jun. 7, 2012 Sheet 16 0f 18
`
`US 2012/0140243 A1
`
`
`
`mwr.07.
`
`mowr
`
`
`
`<wr.muE
`
`NO 2
`
`2
`
`mom?W.
`
`meme
`
`
`
`0017
`
`0017
`
`
`

`

`Patent Application Publication
`
`Jun. 7, 2012 Sheet 17 of 18
`
`US 2012/0140243 A1
`
`2.05
`
`02.07.
`
`2mm
`
`com
`
`owm02
`
`ow3
`
`omcm
`
`cm:
`
`03
`
`03
`
`02
`
`0018
`
`0018
`
`
`
`

`

`Patent Application Publication
`
`Jun. 7, 2012 Sheet 18 of 18
`
`US 2012/0140243 A1
`
`
`
`
`2020
`
`Scan
`
`2003
`
`/
`
`I|
`2006
`
`" AEE"
`
`Il'mlmlll
`[an]
`AAIIIIiIIIIV
`
`
`
`
`FIG. 20
`
`0019
`
`0019
`
`

`

`US 2012/0140243 A1
`
`Jun. 7, 2012
`
`NON-CONTAC T SURFACE
`CHARACTERIZATION USING MODULATED
`ILLUMINATION
`
`CROSS REFERENCE TO RELATED
`APPLICATIONS
`
`[0001] This application claims benefit ofprovisional patent
`application No. 61/419,386, entitled “Non-Contact Surface
`Characterization Using Modulated Illumination,” filed on
`Dec. 3, 2010, the entire contents of which are hereby incor-
`porated by reference.
`
`TECHNICAL FIELD
`
`[0002] This disclosure relates to methods of non-contact
`surface characterization using modulated illumination and
`devices configured to perform the same.
`
`BACKGROUND
`
`[0003] The field of non-contact surface characterization
`includes characterization and/or measurement of information
`
`about an object’s surface including, for example, shape and
`surface finish measurements. Tools capable of non-contact
`surface characterization can be useful in manufacturing dur-
`ing stages such as process development, quality control and
`process control on the production floor. Non-contact surface
`characterization methods can include obtaining information
`about the three-dimensional (3D) shape and location of an
`object by analyzing the frequency content of an image of the
`object in an optical system.
`
`SUMMARY
`
`In general, the subject matter described in this speci-
`[0004]
`fication relates to methods and systems for obtaining infor-
`mation about a surface of an object. The methods disclosed
`can include, for example, directing a structured illumination
`pattern to a surface of best-focus of an imaging optic while
`imaging a field in the surface of best-focus onto a multi-
`element detector. A structured illumination pattern is a non-
`uniform illumination pattem that contains some form of
`encoding based on light intensity. Examples of structured
`illumination patterns include periodic intensity patterns (e.g.,
`modulated in one or two dimensions). An object is scanned
`through the surface of best-focus while a series of images of
`the illumination patterns on the object surface are sequen-
`tially acquired by the multi-element detector. For each suc-
`cessive image obtained, the illumination pattern is modified
`so that its intensity at each point in the field corresponding to
`a detector element is modulated from one frame to the next.
`
`The resulting signal acquired at each detector element has a
`modulation, with the maximum amplitude occurring at or
`near the position of the object surface as it intersects the
`surface ofbest-focus ofthe objective. Based on an analysis of
`the signals provided by the sequence of images, measure-
`ments of topography and surface texture of the object can be
`obtained and, in some implementations, a three-dimensional
`image of the object’s surface can be produced.
`[0005]
`In some embodiments, a detector plane and pattem-
`generating plane are mapped onto a surface in object space
`that is nominally conformal to the shape of the object to be
`characterized including, for example, a plane, a sphere, a
`parabola, a cylinder, a cone, or an aspheric optic.
`[0006]
`In certain embodiments, the signal recorded at a
`pixel (i.e., element) on the detector has a modulation envelope
`
`that emulates some of the characteristics of a scanning low-
`coherence interferometer (SWLI) signal. Conversion of scan
`data to topography or reflectivity data may therefore be
`accomplished by application of envelope-detection algo-
`rithms developed for SWLI such as a frequency domain
`analysis (“FDA”) or least squares analysis (“LSQ”).
`[0007]
`In some embodiments, the device used for creating
`the projected illumination pattern is programmable, which
`allows adapting the frequency content, orientation and spatial
`intensity distribution to optimize the measurement capability
`for a given object.
`[0008] The structured illumination modulation scheme can
`be optimized to enable a fast autofocus scan of an optical
`system. An application in the context of laser eye surgery is
`the localization of the position of laser optics with respect to
`a critical component that makes contact with the cornea.
`Another application is rapid focusing of a low-coherence
`interferometer,
`in which case the autofocus scan rate is
`selected to average out the interference signal from the
`detected signal.
`[0009]
`In certain embodiments, the apparatus presents to
`the user an enhanced image ofthe object that combines height
`information with additional surface information, such as
`color, absorption, texture, etc.
`[0010] Various aspects of the subject matter described in
`this specification are summarized as follows.
`[0011]
`In general, one aspect of the subject matter
`described in this specification can be embodied in methods
`for forming a three-dimensional image of a test object, in
`which the methods include directing light to a surface of
`best-focus of an imaging optic, where the light has an inten-
`sity modulation in at least one direction in the surface of
`best-focus, scanning a test object relative to the imaging optic
`so that a surface ofthe measurement object passes through the
`surface of best-focus of the imaging optic as the test object is
`scanned, acquiring, for each of a series ofpositions ofthe test
`object during the scan, a single image of the measurement
`object using the imaging optic, in which the intensity modu-
`lation of the light in the surface of best-focus is different for
`successive images, and forming a three-dimensional image of
`the test object based on the acquired images.
`[0012] These and other embodiments can each optionally
`include one or more ofthe following features. For example, in
`some implementations, directing the light to the surface of
`best-focus includes imaging a spatial light modulator (SLM)
`to the surface of best-focus. The intensity modulation of the
`light in the surface of best-focus can be varied using the
`spatial light modulator.
`[0013]
`In some implementations, directing the light to the
`surface of best-focus includes imaging a pattem-generating
`plane onto a surface in object space. The surface can be
`conformal to a shape of the test object. The shape of the test
`object can be planar, spherical, parabolic, cylindrical, coni-
`cal, or aspheric.
`[0014]
`In some implementations, the intensity modulation
`is a periodic modulation. The periodic modulation can be a
`sinusoidal modulation. The phase of the periodic modulation
`can be varied by less than 231 between each successive image.
`The phase of the periodic modulation can be varied by at or
`less between each successive image. The phase of the peri-
`odic modulation can be varied by 313/2 between each succes-
`sive image.
`[0015]
`In some implementations, the intensity modulation
`is a two-dimensional intensity modulation. The scan posi-
`
`0020
`
`0020
`
`

`

`US 2012/0140243 A1
`
`Jun. 7, 2012
`
`tions can be evenly spaced. The intensity modulation can be
`selected based on a slope of the test object surface.
`[0016]
`In some implementations, forming the three-dimen-
`sional image includes identifying, for multiple different loca-
`tions of the test object surface, the scan position correspond-
`ing to where each location intersects the surface ofbest-focus.
`An intensity of the acquired images as a function of scan
`position at each of the different locations can include an
`amplitude modulation. Identifying the scan position corre-
`sponding to where each location intersects the surface of
`best-focus can include identifying the scan position where a
`modulation amplitude is largest.
`[0017]
`In some implementations, forming the three-dimen-
`sional image includes deriving an intensity signal for each of
`multiple different locations of the test object surface, each
`intensity signal corresponding to the intensity of the acquired
`images at the corresponding location as a function of scan
`position. Forming the three-dimensional image can include
`identifying a scan position corresponding to a maximum
`amplitude of a modulation of the intensity signal for each
`location. Identifying the scan position can include transform-
`ing each intensity signal into a frequency domain.
`[0018]
`In some implementations, the test object is a lens
`element. The three-dimensional image can be, for example, a
`monochrome image. Alternatively,
`the three-dimensional
`image can be a color image.
`[0019]
`In general, another aspect of the subject matter
`described in this specification can be embodied in methods
`for forming a three-dimensional image of a test object, in
`which the methods each include forming an image of a spatial
`light modulator at a surface ofbest-focus of an imaging optic,
`scanning a test object relative to the imaging optic so that a
`surface ofthe measurement object passes through a surface of
`best-focus of the imaging optic as the test object is scanned,
`acquiring, for each of a series of positions of the test object
`during the scan, a single image of the test object using the
`imaging optic, in which the spatial light modulator varies an
`intensity modulation in the light forming the image so that the
`modulation ofthe light at the surface ofbest-focus is different
`for successive images, and forming a three-dimensional
`image of the test object based on the acquired images.
`[0020]
`In general, another aspect of the subject matter
`described in this specification can be embodied in methods
`for forming a three-dimensional image of a test object, in
`which the methods each include directing light to a surface of
`best-focus of an imaging optic, where the light has an inten-
`sity modulation in at least one direction in the surface of
`best-focus, scanning a test object relative to the imaging optic
`so that a surface of the measurement object passes through a
`surface of best-focus of the imaging optic as the test object is
`scanned, imaging the surface ofbest-focus to a multi-element
`detector, acquiring, for each of a series of positions of the test
`object during the scan, a single intensity measurement at one
`or more elements of the multi-element detector, in which the
`intensity modulation ofthe light in the surface ofbest-focus is
`different for successive positions of the test object during the
`scan, and forming a three-dimensional image of the test
`object based on the acquired intensity measurements.
`[0021]
`In general, another aspect of the subject matter
`described in this specification can be embodied in systems for
`forming a three-dimensional image of a test object, in which
`the systems each include a microscope including an imaging
`optic, the imaging optic having a surface of best-focus, a
`spatial
`light modulator, one or more optical elements
`
`arranged to direct light from the spatial light modulator to
`form an image of the SLM at the surface of best-focus during
`operation ofthe system, a scanning stage arranged to scan the
`test object relative to the microscope object during operation
`of the system so that a surface of the test object intersects the
`surface of best-focus, a multi-element detector positioned
`relative to the microscope such that the microscope forms an
`image of a field at the surface of best-focus on the multi-
`element detector during operation of the system, and an elec-
`tronic control module in communication with the scanning
`stage,
`the spatial
`light modulator, and the multi-element
`detector, in which during operation, the system causes the
`multi-element detector to acquire a single image of the test
`object for each of multiple scan positions of the test object
`relative to the imaging optic, causes the SLM to variably
`modulate the intensity of light at the surface of best-focus in
`at least one direction so that the intensity modulation of the
`light is different for successive images, and forms a three-
`dimensional image of the test object based on the acquired
`images.
`[0022] This and other embodiments can each optionally
`include one or more of the following features. In some imple-
`mentations, the system includes, for example, a light source
`arranged to direct light to the SLM during operation of the
`system. The SLM can be a reflective SLM or a transmissive
`SLM.
`
`In some implementations, the SLM can be a liquid
`[0023]
`crystal panel, or include a micro-mirror array.
`[0024]
`In some implementations, the imaging optic has a
`numerical aperture greater than 0.6, a numerical aperture
`greater than 0.8, a numerical aperture greater than 0.9, or a
`numerical aperture of 0.95.
`the system further
`[0025]
`In some implementations,
`includes color filters arranged to filter the wavelength of light
`forming the images at the multi-element detector.
`[0026] The one or more optical elements can include, for
`example, a fisheye lens, an endoscope, and/or a zoom lens.
`[0027]
`Implementations disclosed herein can offer several
`advantages. For example,
`in some implementations,
`the
`methods and apparatus can be used to provide non-contact
`three-dimensional imaging of an object. The imaging can be
`performed using short measurement times and/or to achieve
`high resolution images of objects’ surfaces. In some imple-
`mentations, the imaging is performed with reduced sensitiv-
`ity to environmental perturbations such as, for example,
`vibration or acoustic noise.
`
`Implementations offer other advantages as well. For
`[0028]
`example, the methods and structures disclosed herein can, in
`some implementations, provide similar depth sectioning
`capabilities as a conventional confocal microscope. The use
`of structured illumination profiling can offer certain benefits
`compared to confocal microscopy (e.g., reduced source
`brightness requirements) and interference microscopy. For
`example, in some implementations, structured illumination
`has reduced source brightness requirements relative to con-
`ventional interference microscopes. Alternatively, or in addi-
`tion, restrictions on which non-interferometric microscope
`objectives can be used can be reduced. For example, in some
`implementations, higher numerical aperture objectives are
`available relative to the objectives available in Michelson or
`Mirau interferometers.
`
`[0029] The details of one or more embodiments are set
`forth in the accompanying drawings and the description
`
`0021
`
`0021
`
`

`

`US 2012/0140243 A1
`
`Jun. 7, 2012
`
`below. Other features and advantages will be apparent from
`the description, drawings, and from the claims.
`
`BRIEF DESCRIPTION OF DRAWINGS
`
`FIG. 1 is a schematic diagram of an exemplary sys-
`[0030]
`tem for providing a 3D image of a sample object.
`[0031]
`FIG. 2 is a schematic diagram of an exemplary sys-
`tem for providing a 3D image of a sample object.
`[0032]
`FIG. 3 is a schematic illustrating various scan posi-
`tions of a sample object.
`[0033]
`FIG. 4 is a graph ofa simulated light intensity signal.
`[0034]
`FIG. 5 is a graph of an experimental intensity signal
`recorded by a pixel of an image-recording device.
`[0035]
`FIG. 6 is an exemplary topography map ofa l3-um
`step height step generated by FDA processing of light inten-
`sity signals.
`[0036]
`FIG. 7 is a graph of a simulated intensity signal.
`[0037]
`FIGS. 8A-8B are plots of a simulated analog inten-
`sity signal overlaid with a corresponding digitized sampling
`of the intensity signal.
`[0038]
`FIG. 8C is a plot of the simulated analog intensity
`signal of FIG. 8B overlaid with the corresponding digitized
`intensity signal of FIG. 8B, where zeros have been inserted
`into the digitized intensity signal.
`[0039]
`FIG. 9A is a plot of the magnitude of the digitized
`sample signal from FIG. 8A and the phase of the digitized
`intensity signal from FIG. 8C.
`[0040]
`FIG. 9B is a plot ofthe phase ofthe digitized sample
`signal from FIG. 8A and the magnitude of the digitized inten-
`sity signal from FIG. 8C.
`[0041]
`FIG. 10A is a plot of a simulated analog intensity
`signal overlaid with a corresponding digitized sampling ofthe
`simulated intensity signal.
`[0042]
`FIG. 10B is a plot of the simulated analog intensity
`signal of FIG. 10A overlaid with the digitized intensity signal
`of FIG. 10A, where zero values have been inserted into the
`digitized intensity signal.
`[0043]
`FIGS. 11-13 are exemplary images of a microlens
`onto which an illumination pattern is projected.
`[0044]
`FIG. 14 is a schematic diagram of an exemplary
`structured illumination imaging system.
`[0045]
`FIG. 15 is a schematic diagram of an exemplary
`imaging system.
`[0046]
`FIG. 16 is a schematic diagram of an exemplary
`imaging system.
`[0047]
`FIGS. 17A-17B are schematic diagrams of exem-
`plary imaging systems.
`[0048]
`FIGS. 18A-18B are schematic diagrams of exem-
`plary imaging systems.
`[0049]
`FIG. 19A is a 3D graph of experimental topography
`data collected on a sample object.
`[0050]
`FIG. 19B is a color image of the surface of the
`sample object of FIG. 19A.
`[0051]
`FIG. 19C is an exemplary color image produced by
`combining the topographical data of FIG. 19A with the color
`image of FIG. 19B.
`[0052]
`FIG. 20 is a schematic of an exemplary system for
`performing laser eye surgery.
`
`DETAILED DESCRIPTION
`
`Acquiring the Light Intensity Signal
`[0053]
`FIG. 1 is a schematic diagram of an exemplary
`structured illumination system 100 for providing a 3D image
`
`of a sample object 102. An illumination portion 103 of the
`system 100 includes an illumination source 104, a spatial
`light modulator (SLM) 106, one or more lenses (108a-108c),
`an aperture 110 and a beam combiner 112. An imaging por-
`tion 105 of the system 100 includes an image-recording
`device 114, a beam splitter 116, an objective lens 118 and a
`tube lens 120. For the purposes of this disclosure, the coor-
`dinate system is defined such that the z-axis is parallel to the
`optical axis of the objective lens 118 and the x and y-axes are
`parallel to the lens’ plane ofbest-focus such that xyz forms an
`orthogonal coordinate system.
`[0054] During operation of the system 100, light 101 gen-
`erated from the illumination source 104 passes through the
`lenses 108a, 1081) and the aperture 110, where the light 101
`then is incident on the beam splitter 112. Beam splitter 112
`reflects a portion of the incident light 101 onto the SLM 106.
`In the present example, the SLM 106 is configured to modify
`and reflect the light incident on its surface so as to produce an
`illumination pattern characterized as either binary (i.e., the
`imaging pattern has regions where light is present and regions
`where light is absent) or quasi-continuous (i.e., the imaging
`pattern can be approximated by continuous varying levels of
`light intensity).
`[0055] The illumination pattern reflected from the SLM
`106 then passes through the beam splitter 112 and lens 108c
`to a second beam splitter 116 where the illumination pattern
`then is directed through the objective lens 118 and preferen-
`tially fills the back pupil of the objective 118. The illumina-
`tion pattern is re-imaged in object space to the plane of best-
`focus (focal plane) of the objective lens 118. Light reflecting
`and/or scattering off a surface of the sample object 102 then
`proceeds through the objective lens 118, the beam splitter
`116, and the tube lens 120 onto a light-detecting surface/plane
`of the image-recording device 114 where it is recorded. The
`recorded light thus acquired can be stored in digital format as
`an array of light intensity signals, with each light intensity
`signal being acquired from a corresponding pixel of the
`image-recording device 114.
`[0056] During imaging, the object 102 is translated verti-
`cally with respect to the objective lens 118 (i.e., toward or
`away from the objective lens 118). The sample object 102 can
`be displaced or actuated by an electromechanical transducer
`(not shown) and associated drive electronics controlled by a
`computer 122 so as to enable precise scans along a direction
`of translation of the object 102. Examples of transducers
`include, but are not
`limited to, piezoelectric transducers
`(PZT), stepper motors and voice coils. Alternatively, or in
`addition, the objective lens 118 may be translated vertically
`with respect to a position of the sample object 102. Again, an
`electromechanical transducer and associated drive electron-
`
`ics may be used to control the translation. The image-record-
`ing device 114 simultaneously records imaging data as the
`object 102 is scanned through the plane of best-focus of the
`objective lens 118 such that multiple light intensity signals
`will be recorded over time. That is, an image ofthe object and
`illumination pattern is captured by the image-recording
`device in the form of the light intensity signal at correspond-
`ing scan positions. For example, if the image-recording
`device 114 includes a 128x128 array of pixels and if 64
`images are stored during a scan, then there will be approxi-
`mately 16,000 light intensity signals each 64 data points in
`length. In some implementations,
`the scan positions are
`
`0022
`
`0022
`
`

`

`US 2012/0140243 A1
`
`Jun. 7, 2012
`
`evenly spaced, i.e., the translation distance between succes-
`sive images captured by the image-recording device is the
`same.
`
`Furthermore, the SLM 106 spatially modulates the
`[0057]
`illumination pattern as the sample object 102 is translated,
`such that the object 102 is illuminated with a different pattern
`at different corresponding vertical positions of the scan. The
`continuous modification of the illumination pattern during
`the scan results in a modulated light intensity signal at each
`pixel of the image-recording device 114. After the data has
`been acquired, the computer 122 can process the light inten-
`sity signals in accordance with, for example, pattern match-
`ing techniques, and output data indicative of a surface topog-
`raphy of the sample object 102.
`[0058] When a surface of the sample object 102 is located
`at the plane of best-focus of the objective 118, an in-focus
`image ofthe illumination pattern is formed on a surface ofthe
`object 102. The in-focus image exhibits the highest contrast
`achievable for a given illumination numerical aperture (NA)
`of the system 100. The intensity of the image of the object
`surface that is reflected back to the image-recording device
`114 at each pixel is proportional to the product of the local
`object surface reflectivity and the local intensity of the pro-
`jected illumination pattern. In contrast, when the surface of
`the object 102 is located away from the plane ofbest-focus of
`the objective lens 118, the image of the illumination pattern
`on the surface of the object 102 is blurred and thus exhibits
`reduced contrast. The image of the surface of the object 102
`passing back through the objective lens 118 towards the
`image-recording device then is blurred again. The result is
`that the contrast of the projected illumination pattern, as seen
`on the image-recording device, is a function of the vertical
`displacement ofthe object surface with respect to the plane of
`best-focus of the objective. Accordingly, the depth profiling
`capability of the system 100 arises from the detection of the
`position of the object 102 for which the pattern contrast is a
`maximum at each pixel ofthe image-recording device 114, as
`the object 102 is scanned through the best-focus plane.
`[0059]
`In some implementations, the SLM 106 can be used
`in a transmission arrangement as opposed to reflection. For
`example, FIG. 2 shows an exemplary system 200 that
`includes the same components as system 100, except that a
`single beam splitter 216 is used in the arrangement and a SLM
`is configured to transmit incident light rather than reflect the
`incident light (computer 122 also is omitted for ease of view-
`ing). In the example shown in FIG. 2, the transmission SLM
`206 modifies incident light 201 generated by an illumination
`source 204 to produce a binary or quasi-continuous illumina-
`tion pattern. The illumination pattern produced by the SLM
`206 passes through a lens 208c and is incident on the beam
`splitter 216, where the pattern then is directed through an
`objective lens 218 and imaged onto a plane ofbest-focus for
`the objective. The illumination pattern reflects off a sample
`object 202 and passes back through the objective lens 218, the
`beam splitter 216, a tube lens 220 and is finally detected by a
`image-recording device 214. Similar to the system 100, the
`illumination pattern produced by the SLM 206 can be spa-
`tially modulated as the image data is acquired, such that
`different light intensity patterns are produced at correspond-
`ing different scan positions of the object.
`[0060]
`In each of the foregoing examples, the illumination
`sources 104, 204 can include, but are not limited to, spec-
`trally-broadband sources or narrow band sources. Examples
`of broadband sources include: an incandescent source, such
`
`as a halogen bulb or metal halide lamp, with or without
`spectral bandpass filters; a broadband laser diode; a light-
`emitting diode; a combination of several light sources of the
`same or different types; an arc lamp; any source in the visible
`spectral region (about 400 to 700 nm), any source in the IR
`spectral region (about 0.7 to 300 um), any source in the UV
`spectral region (about 10 to 400 nm). For broadband applica-
`tions, the source preferably has a net spectral bandwidth
`broader than 5% of the mean wavelength, or more preferably
`greater than 10%, 20%, 30%, or even 50% ofthe mean wave-
`length. The source may also include one or more diffuser
`elements to increase the spatial extent of the input light being
`emitted from the source. Examples of narrow band sources
`include a laser or a broadband source in combination with a
`narrowband filter.
`
`[0061] The source can be a spatially-extended source or a
`point source. In some implementations, it is preferable to use
`a spatially-extended source (e.g., when the surface of an
`object being imaged is smooth) to avoid observing a high-
`contrast pattern regardless of the position of the object with
`respect to the plane of best focus. The image-recording
`devices 114, 214 can include a plurality of detector elements,
`e.g., pixels, arranged in at least one and more generally two
`dimensions. Examples of image-recording devices include
`digital cameras, multi-element charge coupled devices
`(CCDs) and complementary metal oxide semiconductor
`(CMOS) detectors. Other image-recording devices may be
`used as well. In some implementations, one or more color
`filters can be included in the system to filter the wavelength of
`light forming the images at the image-recording device. The
`one or more color filters can be arranged to be an integral
`component ofthe image-recording device or as separate from
`the image-recording device.
`[0062]
`In some implementations, the objective lenses 118,
`218 can be incorporated as components of any standard
`microscope. For example, the systems can include a micro-
`scope configured for use with one or more different objective
`lenses, each providing a different magnification. The NA of
`the objective lenses 118, 218 can be about 0.1 or greater (e.g.,
`about 0.2 or greater, about 0.3 or greater, about 0.4 or greater,
`about 0.5 or greater, about 0.6 or greater, about 0.7 or greater,
`about 0.8 or greater, about 0.9 or greater, or about 0.95 or
`greater). Other NA values are possible as well.
`[0063] Examples of SLMs that can be used to reflect light,
`similar to the arrangement shown in FIG. 1, include liquid
`crystal on silicon (LCOS) devices or micromirror array
`devices, e.g., a digital micromirror device (DMD). To spa-
`tially modulate illumination patterns produced by a LCOS
`device, a user can electronically control, using the appropriate
`hardware and software, the direction and amount of light
`reflected by each pixel of the LCOS device. Likewise, to
`spatially modulate the illumination pattern produced by a
`DMD, a user can electronically control the direction and/or
`orientation of each individual mirror to vary the amount of
`light reflected at each mirror. Examples of SLMs that can be
`used to transmit light, similar to the arrangement shown in
`FIG. 2 include liquid crystal device (LCD) modulators that
`can be electronically controlled to vary the amount of light
`transmitted. Alternatively, in some implementations, trans-
`mission SLMs can include intensity masks, such as gratings,
`Ronchi rulings, or other patterned surfaces that have regions
`of varying degrees of absorption for the wavelength of inci-
`dent light. To spatially modulate the illumination patterns
`produced by the transmission SLMs formed from intensity
`
`0023
`
`0023
`
`

`

`US 2012/0140243 A1
`
`Jun. 7, 2012
`
`masks, the SLMs can be mounted on a mechanical frame that
`provides controlled in-plane motion of the SLM. The motion
`ofthe frame can be provided using one or more mechanical or
`electronic actuators, including, for example, piezoelectric
`actuators, stepper-motors, or voice coils. Other SLMs can be
`used in transmission

This document is available on Docket Alarm but you must sign up to view it.


Or .

Accessing this document will incur an additional charge of $.

After purchase, you can access this document again without charge.

Accept $ Charge
throbber

Still Working On It

This document is taking longer than usual to download. This can happen if we need to contact the court directly to obtain the document and their servers are running slowly.

Give it another minute or two to complete, and then try the refresh button.

throbber

A few More Minutes ... Still Working

It can take up to 5 minutes for us to download a document if the court servers are running slowly.

Thank you for your continued patience.

This document could not be displayed.

We could not find this document within its docket. Please go back to the docket page and check the link. If that does not work, go back to the docket and refresh it to pull the newest information.

Your account does not support viewing this document.

You need a Paid Account to view this document. Click here to change your account type.

Your account does not support viewing this document.

Set your membership status to view this document.

With a Docket Alarm membership, you'll get a whole lot more, including:

  • Up-to-date information for this case.
  • Email alerts whenever there is an update.
  • Full text search for other cases.
  • Get email alerts whenever a new case matches your search.

Become a Member

One Moment Please

The filing “” is large (MB) and is being downloaded.

Please refresh this page in a few minutes to see if the filing has been downloaded. The filing will also be emailed to you when the download completes.

Your document is on its way!

If you do not receive the document in five minutes, contact support at support@docketalarm.com.

Sealed Document

We are unable to display this document, it may be under a court ordered seal.

If you have proper credentials to access the file, you may proceed directly to the court's system using your government issued username and password.


Access Government Site

We are redirecting you
to a mobile optimized page.





Document Unreadable or Corrupt

Refresh this Document
Go to the Docket

We are unable to display this document.

Refresh this Document
Go to the Docket