throbber
EXHIBIT 1064
`
`J. dela TORRE, Jr., “Earth Reference Imager Experiment for Satellite Attitude
`
`Determination,” Arizona State University (1995)
`
`TRW Automotive U.S. LLC: EXHIBIT 1064
`PETITION FOR INTER PARTES REVIEW
`OF U.S. PATENT NUMBER 8,599,001
`IPR2015-00436
`
`

`
`Earth Reference Imager Experiment for
`Satellite Attitude Determination
`
`Student Author: Jose' L. delaTorre, Jr.1
`Faculty Advisors: Dr. Jordi Puig-Suari2 and Dr. Helen L. Reed3
`Arizona State University
`Tempe, Arizona 85287-8006
`
`Abstract
`
`1. Introduction
`
`The Earth Reference Imager (ERI)
`experiment on board the ASUSat I student satellite
`will provide valuable attitude information to validate
`the new attitude sensors along with the GPS data.
`Using the concept of triangulation, the satellite's tilt,
`swing, height, azimuth, and coordinates will be
`attained from an image of the Earth. This
`information will then be compared with the attitude
`sensors to confirm the data received from the satellite.
`
`Nomenclature
`
`1.1 ASUSat 1
`
`Since its inception, Arizona State
`University's first small satellite, ASUSat 1, has been
`a student-designed project. Many members of local
`industry and the faculty of ASU have been very
`supportive of the project, and have sponsored and
`advised several aspects of the program.
`Weighing ten pounds (4.5 kg), the satellite
`will be the lightest ever put into orbit with meaning-
`ful science experiments' . The payload will be
`launched by Orbital Sciences Corporation on a
`Pegasus launch vehicle and will be placed at a 325-
`km sun-synchronous orbit. It is expected to remain in
`orbit for approximately 22 days.
`ASUSat I will allow researchers to measure
`the properties of the ionosphere at a low cost.
`Furthermore, it will provide an audio transponder for
`amateur radio operators.
`There have been an average of forty students
`ranging from high-school through Ph.D level
`working in ten different sub-systems. These are
`science, structures & materials, dynamics & controls,
`communica-tions, power, thermal, commands, ground
`support equipment (GSE), software & data analysis,
`and systems integration.
`
`+X
`
`L
`N
`n
`o
`R (cid:9)
`Rv (cid:9)
`
`S C1 (cid:9)
`S v (cid:9)
`
`A, B, C (cid:9)
`ground control points
`a, b, c (cid:9)
`image ground control points
`f (cid:9)
`focal length
`II (cid:9)
`altitude of satellite
`hA,hn,hc elevation at point
`Lp (cid:9)
`distance from nodal point of lens to
`place of photograph
`space position of front nodal point
`ground nadir point
`nadir point
`principal point
`ratio between AB and AC on the ground
`ratio between All and AC on a vertical
`photograph
`ratio between AB and BC on the ground
`ratio between AB and AC on a vertical
`photograph
`swing of photograph
`s (cid:9)
`* (cid:9)
`introduced tilt of arbitrary amount
`t
`XL, YL, 4 space coordinates of exposure station
`(x, y) (cid:9)
`photographic coordinates of image
`x', y' (cid:9)
`ground coordinates with arbitrary tilt
`calculated about the y axis
`ground coordinates with arbitrary tilt
`calculate about the x axis
`ground survey azimuth
`azimuth of principal plane
`azimuth of control line
`amount of rotation
`
`x", y" (cid:9)
`
`OCG (cid:9)
`ano (cid:9)
`ap (cid:9)
`0 (cid:9)
`
`1Junior, Mathematics, ASUSat 1 Science Team
`2 Assistant Professor, Mechanical & Aerospace
`Engineering; AIAA Member
`3Director, Aerospace Research Center, Associate
`Director, ASU NASA Space Grant Program;
`Professor, Mechanical & Aerospace Engineering;
`Associate Fellow, AIAA
`
`Figure 1 Pullaway view of ASUSat 1. Drawing by
`Shea Ferring and Chris Michaelis, Structures Team.
`1064-001
`
`1064-001
`
`

`
`The science team is developing a Hall-
`Current Ion Accelerator, a Magnetohydrodynamic
`Generator, Ion Thrusters, and the Earth Reference
`Imager.
`
`The structures team designed a fourteen-sided
`graphite / epoxy bus and overlooks cabling and
`mounting on the satellite.
`Dynamics is providing stabilization with an
`aerodynamic boom and a fluid damper. Moreover,
`they are providing positioning information through a
`GPS receiver and an array of photodiodes.
`Communications has student-built the
`modem and designed a deployable antenna.
`The power sub-system will be providing
`power to the satellite through two six-packs of Sanyo
`NiCd batteries, re-charged with GaAs solar cells
`around the fourteen sides of the bus. The satellite
`will be powered up with a mechanical switch when
`the satellite is deployed.
`The thermal team keeps track of the
`temperatures inside the bus, and stabilizes them with
`paints and highly emissive coatings. The
`temperatures expected outside the satellite are in the
`range of -24° C to 90° C.
`Commands is designing the processor around
`the INTEL 80C188EC chip, running it at 10 MHz.
`Two I/O ports provide interfaces with the
`science/dynamics board and with the camera interface
`board. There is also an Error Detection and
`Correction (EDAC) facility built in for the 1 Mbyte
`of RAM available for the satellite's data. One
`EPROM will be used for the default operating
`system, while another will contain the bootloader for
`the satellite's system.
`Ground support is providing the testing
`equipment and facilities. Orbital Sciences has
`allowed the ASUSat 1 team to use several of its
`facilities.
`The software for ASUSat 1 is being written
`in C and compiled to run in the BekTek Real-time
`Spacecraft Operating System designed for micro-
`satellites.
`Finally, the systems and integration team is
`coordinating all the designing, integration, testing and
`documentation of all the sub-systems of the satellite.
`The pointing requirements have been defined
`to be + 10°, with orbital positioning being within ±
`1 km1 to attain full coverage of the Earth. It is hoped
`that the drag coefficient can be calculated. To achieve
`these goals, a new attitude-sensors array has been
`designed and developed to give the satellite's
`orientation.
`The coordinate axis is defined as shown in
`figure 1. The x axis lies along the axis of symmetry,
`y axis runs parallel to the inside panels and
`perpendicular to the x axis, and the z axis runs
`perpendicular to the panels, through the center.
`
`2
`
`1.2 Earth and Sun Sensors
`
`To keep control of the satellite, its
`orientation information will be recorded through a
`complex network of photo sensors. With a total of
`sixty of these photodiodes positioned around the
`satellite, there will be almost complete coverage of
`the sky. Of the sixty sensors, forty-six will be sun
`sensors which will be keyed in to visible light, and
`the remaining fourteen will be Earth sensors, filtered
`to receive in the infra-red?
`There will be blocks of three sun sensors on
`each of the fourteen sides of the satellite, one being
`normal to the surface of the side, and the other two at
`45° from the normal. Earth sensors will be
`positioned on every other side of the satellite, and the
`remaining sensors will be placed on the front and
`back plates of the satellite.2
`The sensors themselves are Motorola Photo
`Detectors, MRD510, with filters being placed on the
`Earth sensors. These diodes will produce voltages
`ranging from 0-5 V, proportional to the amount of
`light received. The readings from each sensor will be
`recorded every five minutes, and the data will later be
`processed on the ground?
`However, this dynamics sensor array has
`never been flown, and other data is necessary to
`confirm the accuracy of the data from the satellite's
`sensors. To confirm the information received from
`the sensors, a camera will be flown to image the
`Earth vertically. Then, by performing the methods of
`triangulation on the ground images from the satellite,
`it is hoped to confirm the positioning data given by
`the sensors array. To ensure that the camera is
`perpendicular to the Earth, the sensors will be used to
`approximate where the Earth is relative to the
`satellite.
`
`1 3 GPS
`
`The Global Positioning System (GPS) was
`developed primarily for military purposes by the
`Department of Defense. However, the technology has
`become available to the public.
`Consisting of a global network of twenty-
`four satellites at high altitudes, GPS provides its
`users with accurate position and velocity data. GPS
`receivers have been used on Lower Earth Orbit
`satellites in the past for determining the satellite's
`orbital parameters.3 By a method of triangulation,
`the satellites are able to practically pinpoint the
`receiver's position.
`Several factors have introduced errors in the
`GPS measurements such as the ionosphere which
`causes some interference. On the average, GPS may
`be off by about sixty feet, and in the worst case, GPS
`data may be off by as much as 350 ft. 3
`
`1064-002
`
`1064-002
`
`

`
`ASUSat 1 will carry a Trimble GPS receiver
`on board. The SVee Six-CM2 is a terrestrial board
`with 6 channels and has never been flown in space
`before, but it has been rigorously tested by ASU
`students to ensure that it will survive in space,4.8
`The accuracy of this particular unit is 25m. Readings
`will periodically be measured and recorded from the
`GPS to determine the satellite's position and
`velocity.4
`From the GPS data included in the
`telemetry, it will be known from where the images
`were taken. This will prove useful in identifying the
`precise location of the image, and in verifying the
`position calculations from the images.
`
`1.4 Earth Reference Imager
`
`Once the images have been received and
`processed on the ground, it will be possible to
`perform several measurements on the images, and do
`calculations on them with a certain degree of error.
`The error will depend on the resolution of the image
`and on the scale of the photograph.
`From the image, three points will have to be
`identified, the distances between each of them
`calculated, and the elevations from sea level of each
`point attained to perform the desired calculations6.
`
`2 Earth Reference Imager
`
`2.1 Imager
`
`The cameras being used for the experiments
`on ASUSat 1 are single-chip CMOS digital cameras.
`The cameras, VVL-1070, have a pixel resolution of
`160 x 160 pixels, and put out 8-bit digital parallel
`output. The pixel size is 10.5 tun x 10.5 ttm
`giving a total effective array size of 1.68 mm x 1.68
`mm. The camera also provides analog output, and an
`exposure control (2,000:1), gain control, and variable
`frame rate (0.5 - 24 frames/sec). At 5V, the chip
`consumes less than 100 mW.
`
`6 cm
`
`0.65 t.'r1
`
`0
`1.5 cm
`
`+5
`P7
`
`PO
`
`PBus
`
`Figure 2 Dimensions and basic layout of camera
`board. There are five control lines and eight data lines
`on each board. The lens is mounted in the corner.
`
`3
`
`2 2 Optics
`
`The lenses being used for the cameras are V-
`4308 multi-element lens with a diameter of 15 mm
`across the top and focal length of 8mm with an f-stop
`of 2.0. The lens weighs about 7 grams. The field
`of view given in the specifications is 31° x 23° (H x
`V). 7 Assuming the witPtlite is at a 325 km orbit and
`that the camera is facing vertically down to Earth, the
`range of the image would be 195.3km x 138km. 8
`giving a resolution of about 1 km per pixel (Fig. 3).
`Furthermore, the scale is calculated as the focal length
`over the altitude of the camera, which is
`approximately 1: 40,625,000 in this case. 7 By
`keeping the field of view and ground coverage limited,
`the resolution should be optimized for the digital
`camera chips.
`
`2.3 Placement
`
`The lens will be fixed over the camera on the
`camera board using a lens holder, with the appropriate
`focal length to focus the image. The board will
`measure approximately 4 cm x 6 cm (Fig. 2),
`allowing the camera to be placed along the side of one
`of the satellite's side panels. A hole of 0.65 cm in
`diameter will be drilled to allow the lens to be
`mounted up to it to view the Earth at a right angle to
`the satellite's axis of symmetry (Fig. 4) close to the z
`axis (Fig. 1).
`
`-41
`
`velocity
`vector
`
`.1111)
`8mm
`
`32 5 km
`
`138 km
`
`Figure 3 With a field of view of 31°-23°, altitude of
`325 km, and focal length of 8 mm, the ground
`coverage should be about 138 km x 195.3 km.
`
`1064-003
`
`1064-003
`
`

`
`2,4 otection
`
`The circuitry for the camera board will be
`surface mounted and conformal coated to protect from
`vibration and shock. Dampers will be placed around
`the board and lens as well. A gold foil will be put
`around the outside camera board to reflect the sun's
`radiation, to keep it cool, and to protect the circuitry
`from cosmic particle hits.
`
`Camera
`
`8 bits
`
`Camera
`
`3 eitmera Boards
`
`+5 V
`
`Power
`
`Commands
`Interface
`
`Main
`CPU
`
`The VVL-1070 chip has a built-in analog-to-
`digital converter (A/D) circuit, so the output is
`already digital.9 The supporting electronics, then,
`drive the camera at real-time clock speeds, and transfer
`the images to the main CPU bus through a 25-pin
`interface.
`ASUSat 1 will carry two CMOS cameras,
`one for determining the satellite's roll rate, and the
`other for the ER1 experiment. Both, will have the
`supporting circuit board (Fig, 2). The boards consist
`of their own microprocessors and memory with built-
`in commands to run the camera. Running at real
`time rates, the image will be dumped into the local
`memory, where the main CPU can then read the
`memory at its own rate and store it to be downloaded
`at a later time.10 There will also be an interface board
`between the two camera boards and the main CPU
`which handles the control of the two cameras from a
`chip select in the interface. The interface consists of
`five control lines for each camera, and eight data lines
`which will be shared by both cameras (Fig. 5). The
`camera boards will be designed by Gordon Minns &
`Associates in Wyoming.
`The software to control all three micro-
`processors will be written in ANSI C and will be
`compiled for a BekTek operating system.I1 The
`
`Earth
`Reference
`Imager
`
`1.5°
`
`Figure 4 The ERI camera will be mounted to one of
`the 14 side panels of the satellite to view the Earth
`vertically with a range of 23°.
`
`4
`
`Figure 5 The two cameras on ASUSat 1 will go to a
`camera interface board which will connect to the main
`processor's control interface. The 8-bit data lines will
`be shared by both cameras.
`
`supporting software will be included with the camera-
`board product to be modified by the software team.
`The commands will be uploaded to the satellite,
`where the cameras will be allotted ten minutes per
`orbit to take images.'2 The camera boards will
`consume less than 100 ntW as well, and the size of
`each image will be about 25 k13. At this size, it will
`be possible to take about ten consecutive images in
`one orbit.I°
`
`4 Image Processing
`
`Once the images have been taken, and
`transferred to the main CPU bus, the data will be
`downloaded to the ground station at Arizona State
`University several times a day. When the images
`have been successfully transferred, they will be
`moved over the internet to a Silicon Graphics
`machine for image processing. Using an image
`processing package such as XV, several analyses will
`be done on the images such as filterings, histograms,
`and averaging.I3 Software such as 1DL is also
`available along with faculty experienced in
`professional image processing.
`Once the images have been optimized,
`whether taking the average of consecutive frames or
`filtering high frequencies, the image will be ready for
`the data analysis.8 From the image, it will be crucial
`to identify the objects and locations on Earth, or in
`the sky, for determining the satellite's attitude. This
`can be done by comparing daily weather maps, terrain
`maps, and GPS data along with the dynamics-sensors
`data to approximate the location at which the image
`was taken.2 Weather maps will reveal cloud covers
`where an image may be completely white, in which
`case the image data will not be very helpful.
`Otherwise, the image can then be used to perform the
`triangulation.
`
`1064-004
`
`1064-004
`
`

`
`(2)
`
`where hAB is the average elevation of points A and B,
`Hvp is the approximate altitude, f is the focal length
`of the lens, and the distances between the points are
`ab and AB. The scale of the photograph is calculated
`from6
`
`scale = —f (cid:9)
`H
`
`(3)
`
`If the camera were truly vertical to the ground, then
`the ground coordinates would simply beta
`
`X (cid:9)
`A (cid:9)
`
`h
`f (cid:9)
`
`x a , y
`
`H — hA
`
`Ya
`
`H hB (cid:9)
`
`f (cid:9)
`H — hc (cid:9)
`f (cid:9)
`
`x, (cid:9)
`
`xc (cid:9)
`
`XB = (cid:9)
`
` C = (cid:9)
`
`H hi;
`B = f (cid:9)
`
`Yb (4)
`
`H h c
`
`Yc
`
`5.1 Assumptions
`
`The principle of triangulation is based on the
`fact that the ratio of two ground lengths computed
`from the ground coordinates of the tilted photograph
`must be the same as the ratio of the corresponding
`known ground lengths. To obtain two ratios, three
`lines are needed, so at least three points must be
`identified. This is assuming, of course, that three
`distinct points can be distinguished on the image, the
`distances between the three points on the ground can
`be calculated, and their respective elevations from sea
`level can be found as well. From these three ground
`points, A, 13 and C (Fig 6), the length between two
`pairs of the points can be determined analytically.
`Furthermore, the location of the three points on the
`globe will be known, so the position of the satellite
`can be determined if the tilt of the image is known.
`
`5.2 Determining Tilt
`
`A tilt t occurs when the optical axis is
`unintentionally deviated from the vertical axis
`coinciding with the direction of gravity." This is
`given by the angle oLn as shown in figures 6 and 7,
`where L is the position of the satellite moving in the
`+x direction, o is the center or principal point of the
`image, and n is the nadir point. The nadir point is
`the point on the photograph vertically beneath the
`satellites imager, or exposure station.
`From the three points on the image, two
`control lines can be formed to find two independent
`ratios such as those given in (8). From these ratios,
`the two components of the tilt on the photograph, tA
`and ty, can be determined, as well as the swing of the
`camera.? The x component relates to a change in
`pitch of the satellite, while the y component of the
`tilt is related to roll in the satellite. The swing is the
`direction of the tilt with respect to the photographic
`axes.
`
`The first step is to calculate the ground
`coordinates based on the photographic coordinates.
`The photographic coordinates are measured with
`respect to the lines joining the opposite fiducial
`marks, giving the three photographic coordinates
`corresponding to the ground points,
`
`(xa,Ya)
`
`( jcb' Yb)
`
`(xc , yc)
`
`C
`
`he
`
`(1)
`
`The altitude of the satellite can be approximated from
`the equation6
`
`5
`
`Figure 6 Three points on the ground represented by
`their corresponding points on a tilted image. The +x
`axis lies in the direction of the velocity vector."
`
`1064-005
`
`1064-005
`
`(cid:9)
`

`
`image plane
`
`ground plane
`
`Figure 7 Tilt is the displacement of the pitch or roll
`on the image plane relative to the ground plane14.
`
`However, taking a swing 9 about the y-axis with an
`induced tilt of t*, the ground coordinates are adjusted
`byt4
`
`x' =xcosO+ysin0
`
`
`y' = —x sin 0+ ycos0+ f tan t*
`
`(5)
`
`Repeating a similar rotational tilt about the x-axis
`yields x" and y". Then, substituting (5) into (4), the
`new equations for the ground coordinates become, in
`genera1,15
`
`Since the ground-control coordinates have now been
`determined, the ground-control lengths of the vertical
`photograph can be calculated by the distance formula
`to yield
`
`2
`
`AB =1((X B — X A) + (Y B — Y A )2)
`
`2
`
`AC = 4(X — X A) + (cid:9)
`
`— Y A)2 ) (7)
`
`BC=
`
`2
`C —XB)2 +( YC —YB)2
`Y C(X
`
`)
`
`Using the three lines above for the triangulation, the
`two ratios are then defined as14
`
`A GBG
`R (cid:9)
`G AGOG
`AGBG
`SG B GCG
`
`(8)
`
`If the camera were vertical or perpendicular to the
`Earth's surface, then the following ground ratios for
`the vertical image could be defined as
`
`A
`R =B
`AC (cid:9)
`V
`AB
`SV=T
`c
`
`(9)
`
`(6)
`
`If equations (8) do not equal the respective ratios in
`(9), then the photograph has some tilt since the ratios
`of the two lines' lengths are not the same. That is,
`the ratios of the lengths from the calculated ground
`coordinates on the photograph do not equal the ratios
`of the same control lengths on the ground.
`Assuming the satellite image will have some tilt,
`either in the pitch or roll, then 14
`
`R'
`
`— A'B'
`A'C' (cid:9)
`A'B'
`S' —
`B'C'
`A"B"
`
`A" C"
`A"B"
`B„ C"
`
`R"
`
`S" —
`
`(10)
`
`1064-006
`
`X — (cid:9)
`
`_ (cid:9)
`
`X" — (cid:9)
`
`Y"
`
`H h
`*
`f sec t • y' sin t
`H h
`* Y'cost*
`f sect — • y' sin t
`H — h
`f sec — y' sin
`H h
`* (cid:9) y, cos f
`fsect
`y'sint
`
`* x'
`
`*
`
`These X and Y coordinates are calculated for each
`point, A, B & C. X' and Y' represent the ground
`coordinates when the tilt is induced about the +y axis,
`while X" and Y" represent the ground coordinates
`when an arbitrary tilt is induced about the +x axis in
`a similar fashion. The lower case x and y in (6)
`represent the photographic x and y coordinates. See
`figure 11 for an illustration of X' and Y'. These are
`the calculated ground coordinates that will be used.
`
`6
`
`1064-006
`
`(cid:9)
`(cid:9)
`(cid:9)
`(cid:9)
`(cid:9)
`(cid:9)
`(cid:9)
`

`
`0
`
`Y'
`
`x
`
`s
`
`n
`
`-
`
`x'
`
`Figure 8 The swing is the displacement of the
`image's coordinate system to the ground's coordinate
`system, ie. a rotation of axes.14
`
`53 Determining Swing
`
`The swing of the photograph is the angle
`measured in the plane of the image from the positive
`y-axis clockwise to a line from the principal point o
`to the nadir point n.14 This is illustrated in figure 8.
`The swing of the camera, s, is easily
`calculated as a function of the components of the tilt
`of the camera, as found when the equations in (15)
`were solved. It is defined as, 14
`tv (cid:9)
`
`(17)
`
`tan s =
`
`tx
`
`It' and S' are the ratios of the lines calculated when a
`tilt was introduced about the +y axis. Similarly, It"
`and S" are the ratios of the lines when an arbitrary tilt
`was introduced about the +x axis. These are found by
`substituting the equations in (6) into the lines defined
`by (7), and taking the ratios as in (8).
`From the information given thus far, the
`change in the ratios can be calculated as the change
`due to the lilt. That is, the change in the ratios due
`to the tilt ty of the camera for the y axis is
`
`R' —RV
`
`(12)
`
`The change in the ratios due to the tilt tx can likewise
`be defined as S'-S„. The rate of change per minute for
`the y axis is then14
`
`R' — RV *- ty, (cid:9) * ty(13)
`
`SV
`
`and
`
`(14)
`
`S" —
`
`SV (cid:9)
`
`t
`
`(cid:9)t
`
`x.
`
`R" —RV
`V (cid:9)
`*
`t (cid:9)
`
`for the tilt in the x axis.
`Since the equations in (8) represent the true
`ground ratios, the changes due to the components of
`the tilt are added. Adding (13) and (14) to the
`equations in (8) yields14
`
`R' — (cid:9)
`
`RG . RV +
`S'
`
`SG = SV + (cid:9)
`
`R" — RV
`
`
`tx
`
`£15)
`
`S"—
`
`+ (cid:9)
`
`' tx
`
`where s can easily be solved if the components of t
`are known. The amount of rotation 0 is defined to be
`
`0 = 1800 — s
`
`(18)
`
`The equations in (15) can then be
`simultaneously solved using linear algebra to find t,
`and ty, the two components of the camera's tilt.
`From the solutions to (15), the pitch and roll of the
`image, and consequently of the satellite, can be
`determined. Furthermore, once the two components
`of the tilt have been found due to the pitch and roll of
`the satellite, the angle t can be found. To solve for
`t,14
`
`t =1(tx2 + ty2 (cid:9)
`
`(16)
`
`Is used. Consequently, the tilt of the camera can be
`determined from the tilt of the image and the
`position of the satellite relative to the Earth can also
`be determined,
`
`Figure 9 The ground survey azimuth is simply the
`sum of the azimuth of the principal plane to the
`azimuth of the given control line 14
`
`7
`
`1064-007
`
`1064-007
`
`(cid:9)
`(cid:9)
`(cid:9)
`(cid:9)
`(cid:9)
`(cid:9)
`

`
`(cid:9) (cid:9)
`
`5,4 Determining Azimuth (cid:9)
`
`The ground-survey azimuth from the North
`is given by'4
`
`then"
`
`allo = (74G — c(!) (cid:9)
`
`(22)
`
`as illustrated in figures 9 and 10. The azimuth, as
`defined in figure 10, will give information on where
`the satellite was located when the image was taken.
`
`(19) (cid:9)
`
`tan aG =
`
` — X
`X
`G2
`G1 (cid:9)
`YG —YG
`1
`2 (cid:9)
`
`5.5 Determining Ground Coordinates
`
`The ground-survey coordinates of the
`exposure station, or the camera, can also be
`calculated, provided that the azimuth of the
`photograph is known from equation (22). Let the
`ground-survey coordinates of the exposure station be
`XL and YL (Fig. 11). The ground-survey axes are
`defined as being X0 and Yo, and the ground axes
`based on the nadir point N and the principal plane are
`X and Y. The ground-survey coordinates of point A
`are then (Xon, YeA) and the ground coordinates are
`(XA, YA). If the X and Y coordinate axes are rotated
`through the positive angle allo, then the transformed
`coordinates of A become (X'A, Y'A). This
`transformation is achieved by'4
`
`X'. XcosaNo + Ysin aNo
`= -XsinaNo +Y cos aNo
`
`(23)
`
`Then the ground coordinates are determined by'4
`
`XL = XGA X' A
`Y = Y -3"
`L GA A
`
`(24)
`
`where ao is the ground-survey azimuth, and points 1
`and 2 are the endpoints of the control lines that were
`used for the calculations of the tilt and swing. If
`
`tan
`
`-X
`1
`2 (cid:9)
`P Y -Y
`2 1
`
`(20)
`
`where ap is the azimuth of the control line based on
`the set of ground coordinates where
`
`XL =0
`YL = 0 (cid:9)
`aNo .0e
`
`L
`
`(21)
`
`+X
`
`Direction
`of
`Flight
`
`Survey
`North
`
`Figure 10 The azimuth of the principal plane is the
`clockwise horizontal angle measured about the ground
`nadir point from the ground survey north meridian to
`the principal plane of the photograph.14
`
`Figure 11 Ground-survey coordinates of exposure
`station for a point A using determined azimuth.'4
`
`8
`
`1064-008
`
`1064-008
`
`(cid:9)
`

`
`6 Conclusions
`
`Using images from the Earth Reference
`Imager experiment, the attitude of the satellite can be
`determined, and the data from the attitude-sensors
`array confirmed. The GPS receiver will help confirm
`the satellite's relative position as well. But using the
`principles of triangulation, it is possible to determine
`the tilt of the camera, the swing, azimuth, and ground
`survey location at a given time, and consequently, the
`satellite's orbital parameters. With this method and
`with extremely small, lightweight and low-power
`cameras, the attitude of any Earth satellite at a given
`time can easily be determined from an image of the
`Earth.
`
`7 AcknQwiedpments
`
`This project is supported by the National
`Science Foundation, the NASA Space Grant
`Program, the Aerospace Research Center, and Arizona
`State University. Orbital Sciences Corporation is
`providing the launch, advising, and facilities, free of
`charge. Honeywell Space Systems Group, AMSAT
`Organization, Intel's University Support, Motorola's
`Satcom and University Support, Photocomm, Inc.,
`Trimble Navigation, Astro Aerospace, Universal
`Propulsion Company, Inc., ICI Fiberite Composites,
`Simula, Inc., DynAir, BekTek, KinetX, Rockwell,
`Sinclabs, Inc., Equipment Reliability Group, Bell
`Atlantic Cable, Gordon Minns and Associates, AERL
`(Australia), and Jet Propulsion Laboratory are also
`providing technical advising, hardware, and use of
`facilities.
`The author especially thanks Dr. Frank
`Aldrich, Dr. Paul Scowen, Dr. Helen Reed, Dr. Jordi
`Puig-Suari and Charles Hewett for their important
`contributions and support.
`The design and development of ASUSat 1 is
`a conglomeration of all students involved since its
`inception in October 1993 through August 1995.
`
`8 References
`
`1 Rademacher, Joel. "Systems Specifications Docu-
`ment: 1.0 Overview." Arizona State University,
`(1995).
`
`2 Potterveld, Curtis. "Systems Specifications Docu-
`ment: 3.7 Dynamics/Control." Arizona State Univer-
`sity (1995).
`
`3 Hum, Jeff. GPS: A Guide to the Next Utility.
`Sunnyvale: Trimble Navigation, 1989.
`
`9
`
`4 Stillman, Catherine. "Systems Specifications
`Document: 3.7.3.4.1 GPS." Arizona State Univer-
`sity (1995).
`
`5 Gish, David. "Testing Procedures of the GPS
`Receiver." Testing documentation, Arizona State
`University (1995).
`
`6 Baker, Wilfred H. Elements s,f Photogrammetty.
`New York: The Ronald Press Company, 1960.
`
`7 Slama, Chester C., ed. Manual of Fhologmin-
`poetry; Fourth Edition. Falls Church: American
`Society of Photogrammetry, 1980.
`
`8 Kiefer, Ralph W. and Thomas Lillesand. Remote
`Sensing and Image Interpretation. 2nd edition. New
`York: John Wiley & Sons, 1987.
`
`9 VVL Vision Ltd. "VVL-1070 Monochrome Mono-
`tide Image Sensor with Analogue and Digital
`Outputs: Preliminary Data Sheet."
`
`10 Wightman, Mark. "Systems Specifications
`Document: 3.3.3 Commands." Arizona State Univer-
`sity (1995).
`
`11 yo, pe- rry. "Systems Specifications Document:
`3,3.2 Software." Arizona State University (1995).
`
`12 Bowman, Roberta. "Systems Specifications
`Document: 3.4 Power." Arizona State University
`(1995).
`
`13 Cracknell, Arthur and Ladson Hayes. Jntroduction
`14 Remote Sensing. New York: Taylor & Francis,
`1991.
`
`14 Moffitt, Francis H. Photogrammetry. Pennsyl-
`vania: International Textbooks Company, 1967.
`
`15 Crone, D. R. Elementary Photogrammetry, New
`York: Frederick Ungar Publishing Co., 1963.
`
`1064-009
`
`1064-009

This document is available on Docket Alarm but you must sign up to view it.


Or .

Accessing this document will incur an additional charge of $.

After purchase, you can access this document again without charge.

Accept $ Charge
throbber

Still Working On It

This document is taking longer than usual to download. This can happen if we need to contact the court directly to obtain the document and their servers are running slowly.

Give it another minute or two to complete, and then try the refresh button.

throbber

A few More Minutes ... Still Working

It can take up to 5 minutes for us to download a document if the court servers are running slowly.

Thank you for your continued patience.

This document could not be displayed.

We could not find this document within its docket. Please go back to the docket page and check the link. If that does not work, go back to the docket and refresh it to pull the newest information.

Your account does not support viewing this document.

You need a Paid Account to view this document. Click here to change your account type.

Your account does not support viewing this document.

Set your membership status to view this document.

With a Docket Alarm membership, you'll get a whole lot more, including:

  • Up-to-date information for this case.
  • Email alerts whenever there is an update.
  • Full text search for other cases.
  • Get email alerts whenever a new case matches your search.

Become a Member

One Moment Please

The filing “” is large (MB) and is being downloaded.

Please refresh this page in a few minutes to see if the filing has been downloaded. The filing will also be emailed to you when the download completes.

Your document is on its way!

If you do not receive the document in five minutes, contact support at support@docketalarm.com.

Sealed Document

We are unable to display this document, it may be under a court ordered seal.

If you have proper credentials to access the file, you may proceed directly to the court's system using your government issued username and password.


Access Government Site

We are redirecting you
to a mobile optimized page.





Document Unreadable or Corrupt

Refresh this Document
Go to the Docket

We are unable to display this document.

Refresh this Document
Go to the Docket