throbber
PROCEEDINGS
`SPIE-The International Society for Optical Engineering
`,;:f ~GoF;7;;7::,-.,
`( ~ f.'.;\'f 1 fi n?7
`)
`'<~,,~·!zile_y}~ ;:~ ~~;~~m::sr~_,✓,l
`-...
`
`I
`
`25th A/PR Workshop
`Emerging Applications
`of Computer Vision
`
`David Schaefer
`Elmer F. Williams
`Chairs/Editors
`
`16-18 October 1996
`Washington, D.C.
`
`Sponsored by
`SPIE-The International Society for Optical Engineering
`AIPR Executive Committee
`
`Published by
`SPIE-The International Society for Optical Engineering
`
`r!1
`
`Volume 2962
`
`I
`l
`,
`! 1
`!
`
`SPIE is an international technical society dedicated to advancing engineering and scientific
`applications of optical, photonic, imaging, electronic, and optoelectronic technologies.
`
`TX 4-468-318
`\IIIIHIHWIIH
`
`•TX0004468318•
`
`

`

`-JlJ1&3f'
`'//is/
`1996¼
`C6f'/ !l-
`
`tE
`
`The papers appearing in this book comprise the proceedings of the meeting mentioned on the
`cover and title page. They reflect the authors' opinions and are published as presented and
`without change, in the interests of timely dissemination. Their inclusion in this publication does
`not necessarily constitute endorsement by the editors or by SPIE.
`
`Please use the following format to cite material from this book:
`Author(s), "Title of paper," in Emerging Applications of Computer Vision, David Schaefer,
`Elmer F. Williams, Editors, Proc. SPIE 2962, page numbers (1997).
`
`ISSN 0277-786X
`ISBN 0-8194-2366-1
`
`Published by
`SPIE-The International Society for Optical Engineering
`P.O. Box 10, Bellingham, Washington 98227-0010 USA
`Telephone 360/676-3290 (Pacific Time) • Fax 360/647-1445
`
`Copyright 0 1997, The Society of Photo-Optical Instrumentation Engineers.
`
`Copying of material in this book for internal or personal use, or for the internal or personal use
`of specific clients, beyond the fair use provisions granted by the U.S. Copyright Law is
`authorized by SPIE subject to payment of copying fees. The Transactional Reporting Service
`base fee for this volume is $10.00 per article (or portion thereof), which should be paid directly
`to the Copyright Clearance Center (CCQ, 222 Rosewood Drive, Danvers, MA 01923. Payment
`may also be made electronically through CCC Online at http://www.directory.net/copyright/.
`Other copying for republication, resale, advertising or promotion, or any form of systematic or
`multiple reproduction of any material in this book is prohibited except with permission in
`writing from the publisher. The CCC fee code is 0277-786X/97/$10.00.
`
`Printed in the United States of America.
`
`...
`
`

`

`-=•·t
`
`Contents
`
`vii
`ix
`
`A/PR Executive Committee
`Introduction
`
`SESSION 1
`
`2
`
`14
`
`26
`
`MEDICAL APPLICATIONS
`Local force model for cardiac dynamics analysis based on CT volumetric image sequences
`[2962-02]
`J. G. Tamez-Pena, Univ. of Rochester; C. W. Chen, Univ. of Missouri/Columbia; K. J. Parker,
`Univ. of Rochester
`
`Small object detection using morphological filtering and multiresolution analysis with
`application to microcalcification detection in mammograms [2962-03]
`L. Chen, C. W. Chen, K. J. Parker, Univ. of Rochester
`Active contour based on the elliptical Fourier series applied to matrix-array ultrasound of the
`heart [2962-04]
`R. Drezek, G. D. Stetten, T. Ota, C. Fleishman, E. Lily, C. Lewis, C. J. Ohazama, T. Ryan,
`D. Glower, J. Kisslo, 0. T. von Ramm, Duke Univ.
`
`SESSION 2
`
`36
`
`IMAGE MINING
`Wavelet index of texture for artificial neural network classification of Landsat images
`[2962-05]
`H. H. Szu, Naval Surface Warfare Ctr.; J. Le Moigne, NASA Goddard Space Flight Ctr.;
`N. Netanyahu, Univ. of Maryland/College Park; C. C. Hsu, Trident Systems, Inc.; M. Francis,
`Univ. of Southwestern Louisiana
`
`45
`
`55
`
`63
`
`78
`
`Registration of satellite imagery utilizing the low-low components of the wavelet transform
`(2962-06]
`E. Kaymaz, B.-T. Lerner, KT-Tech, Inc.; W. J. Campbell, J. le Moigne, NASA Goddard Space
`Flight Ctr.; J. F. Pierce, U.S. Naval Academy
`
`Spectral unmixing of remotely sensed imagery using maximum entropy [2962-07]
`S. R. Chettri, NASA Goddard Space Flight Ctr.; N. Netanyahu, Univ. of Maryland/College Park
`and NASA Goddard Space Flight Ctr.
`Spectral imaging applications: remote sensing, environmental monitoring, medicine, military
`operations, factory automation, and manufacturing (2962-08]
`N. Gat, S. Subramanian, Opto-Knowledge Systems, Inc.; J. Barhen, Oak Ridge National Lab.;
`N. Toomarian, Jet Propulsion Lab.
`Extracting an image similarity index using meta-data content for image mining applications
`[2962-09]
`S. Raghavan, LNK Corp.; R. F. Cromp, NASA Goddard Space Flight Ctr.; 5. Srinivasan,
`R. Poovendran, LNK Corp.; W. J. Campbell, NASA Goddard Space Flight Ctr.; L. Kanai,
`LN K Corp. Inc.
`
`iii
`
`

`

`SESSION 3
`
`94
`
`97
`
`SOCIAL IMPACT
`Milestones on the road to independence for the blind [2962-10)
`K. Reed, NASA Goddard Space Flight Ctr.
`
`Social impact of computer vision (2962-11]
`H. Baetjer, Loyola College
`
`SESSION 4
`
`AUTO AND FLIGHT SAFETY AIDS
`
`104
`
`111
`
`122
`
`Multiple vehicle detection and tracking [2962-13)
`M. Betke, E. Haritaoglu, L. S. Davis, Univ. of Maryland/College Park
`
`Image processing using acousto-optical tunable filtering [2962-14)
`.
`L. J. Denes, B. Kaminsky, M. S. Gottlieb, P. Metes, Carnegie Mellon Research Institute;
`S. Simizu, R. T. Obermyer, C. J. Thong, M. J. Uschak, S. G. Sankar, Advanced Materials Corp.
`
`Real-time visual processing in support of autonomous driving [2962-16)
`M. Nashman, National Institute of Standards and Technology; H. Schneiderman, Carnegie
`Mellon Univ.
`
`133
`
`Real-time landmark-based optical vehicle self-location [2962-17)
`M. D. Squires, M. P. Whalen, G. Moody, C. J. Jacobus, Cybernet Systems Corp.
`
`SESSION 5
`
`REAL-TIME EVENT UNDERSTANDING
`
`144
`
`Autonomous video surveillance [2962-20)
`B. E. Flinchbaugh, T. J. Olson, Texas Instruments Corporate Research Labs.
`
`SESSION 6
`
`MILITARY APPLICATIONS
`
`154
`
`162
`
`171
`
`180
`
`192
`
`RADIUS testbed system (2962-22]
`D. J. Gerson, S. E. Wood, Jr., CIA Office of Research and Development
`
`IU for military and intelligence applications: how automatic will it geH (2962-23)
`J. L. Mundy, GE Corporate Research and Development Ctr.
`
`User interface representations for image understanding (2962-24)
`M.A. J. Puscar, A. J. Hoogs, Lockheed Martin Corp.
`
`Region of interest identification in unmanned aerial vehicle imagery (2962-25]
`J. L. Solka, D. J. Marchette, G. W. Rogers, E. C. Durling, J. E. Green, D. Talsma, Naval Surface
`Warfare Ctr.
`
`RADIUS testbed database: temporal queries and optimization [2962-26]
`R. Cardenas, A. J. Hoogs, Lockheed Martin Corp.
`
`SESSION 7
`
`INFORMATION FROM TWO-DIMENSIONAL IMAGES
`
`202
`
`Integrated system for automated financial document processing (2962-27]
`K. Hassanein, S. Wesolkowski, R. Higgins, NCR (Canada); R. Crabtree, A. Peng, NCR
`
`iv
`
`

`

`213
`
`226
`
`236
`
`Inspection of surface-mount device images using wavelet processing [2962-28]
`G. Carillo, S. D. Cabrera, A. A. Portillo, Univ. ofTexas/EI Paso
`
`Automated building extraction using dense elevation matrices [2962-29]
`A. A. Bendett, U. A. Rauhala, J. J. Pearson, GOE Systems, Inc.
`
`Fusing mainstream and media processors to solve embedded imaging applications affordably
`(2962-30]
`R. Rinn, C. Fleischer, Parsytec Inc.
`
`SESSION 8
`
`FACIAL AND GESTURE RECOGNITION
`
`244
`
`253
`
`Computing 3D head orientation from a monocular image sequence [2962-31]
`T. Horprasert, Y. Yacoob, L. S. Davis, Univ. of Maryland/College Park
`
`FERET (Face Recognition Technology) program [2962-32]
`P. J. Rauss, P. J. Phillips, M. K. Hamilton, U.S. Army Research Lab.; A. T. DePersia, National
`Institute of Justice
`
`264
`
`Face recognition using hybrid systems [2962-34]
`S. Gutta, J. Huang, H. Wechsler, George Mason Univ.; B. Takacs, Physics Optics Corp.
`
`SESSION 9
`
`LAW ENFORCEMENT APPLICATIONS
`
`276
`
`287
`
`Infrared facial recognition technology being pushed toward emerging applications [2962-35]
`D. C. Evans, Technology Recognition Systems, Inc.
`
`Hyperspeed data acquisition for 3D computer vision metrology as applied to law enforcement
`(2962-36]
`B. R. Altschuler, Walter Reed Army Medical Ctr.
`
`295
`
`Commercialization of the weapons team engagement trainer: update of process [2962-37]
`J. W. Healy, J. Horey, R. T. McCormack, R. S. Wolff, E. E. Purvis 111, Naval Air Warfare Center
`
`302
`
`Author Index
`
`/
`
`V
`
`

`

`+
`!
`
`I
`4
`
`i
`
`\
`
`Autonomous Video Surveillance
`
`Bruce E. Flinchbaugh and Thomas J. Olson
`
`Texas Instruments Corporate Research Laboratories
`P.O. Box 655303, MIS 8374, Dallas, TX 75265
`
`ABSTRACT
`
`fl,
`\
`I
`I
`,tr!
`
`..
`
`i
`
`•
`
`This presentation highlights needs for autonomous video surveillance in the context of physical secu(cid:173)
`rity for office buildings and surrounding areas. Physical security is described from an operational
`perspective, defining the principal responsibilities and concerns of a physical security system. Capabilities
`and limitations of current video surveillance technology are described, followed by examples of how com(cid:173)
`puter vision techniques are being used and advanced for autonomous video surveillance systems.
`
`Keywords: video surveillance, scene monitoring, computer vision, security
`
`1. Physical Security Systems and Operations
`
`Major activities of physical security for office buildings today are: access control, intrusion detection,
`guard patrols, CCTV surveillance, alarm monitoring, response dispatching, and investigations. Autono(cid:173)
`mous video surveillance technology will ultimately improve productivity and effectiveness in all of these
`activities. The primary responsibilities of physical security are described below.
`
`1.1 Access Control
`
`The most common approach to physical security is to control who may enter a building or an area at
`the perimeter. By restricting access to trusted individuals (e.g., employees), many opportunities for securi(cid:173)
`ty breaches are eliminated. Door locks and keys provide much of this security, and many buildings deploy
`guards at building entrances to control who enters. The primary automatic access control technology in use
`today is provided by electronic badge reader systems. In this approach, an electronically readable badge is
`issued to each person with access privileges, and the person may use the badge like a key to open doors
`where guards are not stationed. A drawback of keys and electronic badges is that they may be used by un(cid:173)
`authorized individuals to gain access. For tighter security in access control, a variety of biometric access
`control technologies are available to measure physical characteristics of people: voice verifcation, retina
`scanners, fingerprint scanners, hand scanners, face recognition, and body weight measurements. The pri(cid:173)
`mary limitation of access control technology is that it does not defeat security breaches by insiders.
`
`1.2 Intrusion Detection
`
`A second line of defense against physical security breaches is to detect situations where people gain
`unauthorized access to a facility. This may occur at regular access control points (e.g., doors) or at other
`places (e.g., windows and fences). For example, an unauthorized person might enter by "piggy backing"
`on access by authorized person (i.e., by slipping through a door before it closes), or by any number of
`physical "breaking and entering" approaches. Various technologies are available to detect intrusions. The
`most commonly used devices are door switches and infrared motion detectors, like those found in many
`home security systems. Infrared motion detectors operate by detecting rapid heat changes in an area that
`
`144
`
`SPIE Vol. 2962 • 0277-786X/97/$10.00
`
`

`

`+
`!
`
`I
`I .,.
`
`I
`
`T
`
`~
`;
`I
`I
`4:";
`
`... :·
`
`...
`i .. I
`!
`
`l
`
`4
`I
`
`~
`i
`
`I
`~
`
`~
`
`I
`i
`
`l
`
`I -~
`
`e.~~
`i
`I
`
`are caused when a person enters the area. Other intrusion detection techniques include sound detectors,
`g_lass-break detectors, light beams, and various other electromagnetic change detectors. Although an intru(cid:173)
`sion detector may be used to signal alarms, it does not provide a description of the specific situation that
`causes an alarm, thereby requiring a subsequent follow-up action to respond to the alarm or investigate the
`incident later.
`1.3 Guard Patrols
`In addition to guards who monitor entrances for access control, a common security practice is for
`guards to periodically patrol sites to visually confirm that the premises are secure and to identify and report
`adverse conditions (e.g., missing property and damage from vandalism). Nearly all of the technology that
`supports these visual surveillance activities in practice today involves mechanical and electronic devices
`that the guards use to prove that they visited specific areas and to report observations. However, some ad(cid:173)
`vanced robotic security systems attempt to accomplish various guard patrol responsibilities. For example,
`mobile robots equipped with sonic sensors for navigation and detection have been installed for experimen(cid:173)
`tal security applications to patrol buildings, but this approach faces many problems to overcome before it
`can compete effectively with guards and more-reliable video surveillance approaches.
`1.4 CCTV Surveillance
`Closed circuit television (CCTV) camera networks that supply video data to security system centers
`are often used to support physical security operations in buildings and surrounding areas. In a small sys(cid:173)
`tem, a few cameras may be cabled to TV monitors for remote viewing by guards or other security personel.
`Another common practice is the use of time-lapse video tape recorders to record video data from one or
`more cameras. Some time-lapse recorders provide a multiplexed recording capability so that several cam(cid:173)
`eras may be recorded on a single tape. In large CCTV security systems, many cameras are cabled to an
`array of TV monitors and video tape recorders, to support live observations as well as after-the-fact inves(cid:173)
`tigations using recorded video data.
`CCTV cameras may be mounted with a fixed field of view, or mounted on a pan and tilt mechanism
`that can be remotely controlled by an observer to view a wider area. Video cameras may also be mounted
`on a small platform that moves along a track, enabling a single camera to scan a much wider area ( e.g., by
`moving along a long wall in a large parking garage).
`The primary autonomous video surveillance systems available today are known as "video motion de(cid:173)
`tectors" or "VMDs". In principle, VMDs can be programmed for various tasks such as intrusion detection
`and to signal alarms for fairly complex situations. For example, some VMDs can be programmed to signal
`an alann when something moves across the field of view from left to right, while not signaling an alarm
`when something moves from right to left. However, in practice available VMDs produce too many false
`alarms in typical environments [6], and they require substantial operator training and experience to pro(cid:173)
`gram the system.
`1.5 Alarm Monitoring and Response Dispatching
`Large physical security systems generally have a centralized control center where security staff moni(cid:173)
`tor alarms, dispatch guards to respond to incidents, and maintain a record of the incidents and their
`resolution. In some cases, control center operators actively monitor remote cameras to detect incidents and
`enoaoe in wide-ranoino visual surveillance tasks. Although people are very good at these tasks, long peri(cid:173)
`0 r inactivity in a scene make this monitoring job tedious. It is also impractical for
`od; ;f routine activity
`one person to reliably and simultaneously monitor hundreds of cameras. Thus control ce_nter operators typ(cid:173)
`ically spend most of their time monitoring and processing alarms from autonomous devices.
`
`0
`
`; l
`
`I
`
`145
`
`

`

`1.6 Investigations
`Although the primary goal of physical security is to prevent security breaches, or to detect and re(cid:173)
`spond in time to minimize the impact, a rapid and accurate capability to investigate incidents after the fact
`is also a key activity of physical security. Unlike retail store security (where a frequent threat is shoplifting
`by customers) office building security often faces the challenge of determining which insider has stolen
`something after it is determined that a theft has occurred. In this matter, differences between physical secu(cid:173)
`rity and information security begin to blur. For example, evidence for theft of trade secrets might be
`visually observable (as when an employee opens a file drawer and reads or copies a document) even
`though the document may be physically replaced in the drawer within a few minutes. For cases where a
`large material property item has been stolen, time-lapse video recordings of building entrances and exits
`provide a useful record of events for investigators to search for suspects. However, the limited video sur(cid:173)
`veillance tool for investigators in this regard remains a video tape player with a fast-forward button, and
`automatic video surveillance systems to assist investigators in detecting theft of intellectual property are
`beyond the state of the art.
`
`2. Computer Vision in Autonomous Video Surveillance Systems
`
`Texas Instruments has demonstrated a variety of autonomous video surveillance capabilities involving
`computer vision. In this section we summarize several TI capabilities for surveillance in office building
`environments.
`
`Our approach to video surveillance has been to exploit data that is readily computed from incoming
`video streams by using domain-specific constraints and contextual information to interpret the data. The
`video processing is primarily a matter of detecting areas of dynamic change via thresholded image differ(cid:173)
`ences and forming connected components for the subsequent analysis. In addition to devising new methods
`for providing more accurate and reliable information about moving entities and their surroundings, we are
`focusing on computer vision for interpreting scene events involving complex spatio-temporal conditions
`and interactions.
`
`2.1 People Tracking and Position Mapping
`
`The most basic function of a surveillance system is to provide situational awareness, i.e., to inform se(cid:173)
`curity personnel of what is going on in the monitored area. TI has demonstrated an end-to-end real-time
`video surveillance system that uses a visual memory paradigm to integrate information from multiple cam(cid:173)
`eras and to provide situational awareness. The visual memory is an object-oriented database that supports a
`variety of spatio-temporal queries [1, 3, 4]. The system detects and tracks people at ten frames per second,
`applies a ground plane constraint to estimate their 3-D positions, and records their positions and other in(cid:173)
`formation in the visual memory. A graphical interface allows users to construct security-related queries
`and view the results.
`
`Figure 1 shows the system being used to monitor activity in a hallway. To detect people, the system
`differences live video frames with a background reference image to estimate regions of change. These re(cid:173)
`gions are grouped into collections that are consistent with the size and shape of a person walking in the
`field of view. This process is repeated until all regions are either interpreted as part of a person, or dis(cid:173)
`missed if sizes and shapes of neighboring regions are inconsistent with such an interpretation. The user
`interface presents video surveillance observations from the visual memory in an interactive map display.
`Users interact with a map of the monitored region, and can pan, scroll and zoom to focus on a region of in(cid:173)
`terest. Queries enable the display of both current positions of objects of interest and historical positions
`(i.e., paths) of particular objects. Users can also specify alarms to be generated when someone enters a par(cid:173)
`ticular region.
`
`146
`
`

`

`♦
`
`•·
`
`FIGC~RE 1. TI people tracking and position mapping system. The system tracks moving objects in video from
`secunty cameras, as at left Square icons in the map-based interface at right represent location histories of humans at
`left. Pop-up window gives a detailed view of one observation record. Shaded rectangles in map are alarm regions.
`
`2.2 Detection of People Carrying Boxes
`After a theft has been reported in an area monitored by time-lapse video recorders, security staff typi(cid:173)
`cally review video tapes for the period during which the theft occurred. Tapes captured at building access
`points often reveal how the stolen material left the facility, or yield a short list of people who may have re(cid:173)
`moved the material. However, the process of reviewing the tapes can be extremely tedious, especially when
`the period of interest spans many days.
`In these experiments, TI demonstrated a system for screening security video tapes to identify seg(cid:173)
`ments that may contain images of people carrying boxes. This task was suggested by experts in physical
`security operations, who often look for boxes as part of theft investigation. The algorithm uses perceptual
`grouping techniques to identify collections of lines that may be projections of rectangular solids. If it finds
`a high-quality grouping of an appropriate size, it flags that video frame for inspection by a human and pro(cid:173)
`ceeds to the next frame. This relieves the investigator of the need to examine and interpret every frame.
`The box detection algorithm was tested on 500 frames taken from time-lapse video sequences show(cid:173)
`ing humans passing through a revolving door. Figure 2 shows a typical frame. The system exhibited a 93%
`detection rate at a false alarm rate of 13%. Details of the experiments and algorithm are given in a forth(cid:173)
`coming paper [5].
`2.3 Event Description for Video Indexing Using Motion Graphs
`Future autonomous video surveillance systems will need to be able to classify motions and interac(cid:173)
`tions of objects into events that are meaningful and important to security staff. TI has developed an
`Automatic Video Indexing (AVI) system ([2], figure 3), which performs event recognition on surveillance
`video tapes. As in the case of the box detection work described above, the immediate goal of this project
`was to assist human investigators in finding relevant segments of security video recordings; however, the
`event recogntion algorithm is general and can be applied to many video understanding tasks.
`In the AVI system, object motions and interactions are described by a directed acyclic graph called a
`motion graph. Each node of the graph is an observation of an object, which is simply a region of change
`detected by image differencing. In the motion graph, each object is tracked and is linked to its predecessor
`and successor in time. Forks and joins in the graph represent complex interac~ions. For example, if a per(cid:173)
`son enters a scene, puts down an objects and leaves, the graph will contain a chain of nodes representing
`the person, with a fork node whose successors are the continuation of the person track and a chain of ob-
`servations of a stationary object.
`
`147
`
`

`

`f
`
`t
`
`t
`
`,
`
`,
`
`r
`
`FIGURE 2. Box Detection in a security video image. Lines overlaid on the figure were identified by the detection
`algorithm as providing evidence for a box. In testing on 500 frames, the algorithm achieved a 93% detection rate at a
`false alarm rate of 13%.
`
`FIGURE 3. The Automatic Video Indexing system detects significant events in security videotapes. Above, the
`system identifies an instance of removal of an object from the scene, defined as disappearance of a stationary object
`(the briefcase) while in contact with a moving object (the human). An overlaid box highlights the objects involved in
`the removal event. Other removals from other points in the videotape are shown on the clipboard at right.
`
`Figure 4 provides an example of how the motion graph is constructed and interpreted. Vertical lines
`rep~esent frames, here compressed to ID to simplify the diagram. Each node of the graph consists of an ob(cid:173)
`servation of an object at a particular place in a particular frame. Links between objects constitute
`hypotheses about identity, constructed by tracking objects over time. Nodes with two or more successors
`correspond to places where the tracking algorithm determined that an object split into multiple objects.
`Similarly, nodes with multiple predecessors result from the merging of multiple objects. Events are defined
`
`148
`
`

`

`t
`
`t
`
`7
`
`,
`
`"' f
`
`1
`
`r
`
`Entrance
`
`·\1
`! I -
`I
`
`Remoll,a/ J f
`
`Entrance
`
`FIGURE 4. Automatic Video Indexing motion graph example. The motion graph captures the motions and
`interactions of objects in the scene. Predicates on the graph correspond to event classes such as Entrance, Exit,
`Removal, et cetera.
`by charactenst1c structures m the graph, as shown m the figure. For example, a cham of observations be(cid:173)
`ginning near the edge of the image constitutes an Entrance event, a moving object that splits into a moving
`object and a stationary one constitutes a Deposit event, and so on.
`The AVI video indexing capability allows extremely rapid access to significant events in long time(cid:173)
`lapse security videos. The algorithms were demonstrated live at the 1996 Image Understanding Workshop
`and have been used with both visible and infrared video data.
`
`3. Advanced Proof-of-Concept Demonstrations
`In autonomous video surveillance research, TI is integrating the capabilities described in the preced(cid:173)
`ing section to produce a series of proof-of-concept demonstrations. Together these demonstrations
`illustrate how physical security operations of the future will use autonomous video surveillance systems.
`Key challenges are computing event graphs on-line at ten frames per second, and devising event graph
`analysis methods to exploit contextual information in useful office surveillance scenarios.
`This research is focussed on three overall surveillance scenarios, for monitoring hallway, office, and
`building perimeter areas. In each area, a camera provides live video data of scenes in the field of view,
`while the AVS system monitors the video to analyze events and signal alarms.
`3.1 Hallway Surveillance
`In this scenario, suggested previously in Figure 1, the autonomous video surveillance system detects
`and tracks people as they walk in office building hallways. Alarms are interactively defined for conditions
`such as when someone loiters in a specified area or enters a particular office. This visual assessment pro(cid:173)
`vides information to augment other security system data, such as biometric access control information at
`building entrance points.
`
`149
`
`

`

`, lj
`
`FIGURE 5. Office Surveillance Scenario. Video event recognition can be used to detect unauthorized access to
`documents, file drawers, computers, et cetera.
`
`t
`
`;
`
`.,
`'
`
`t
`
`FIGURE 6. Perimeter Surveillance scenario. The AVI event recognition algorithms can be used with low-cost
`infrared cameras to detect security threats in tptal darkness.
`
`3.2 Office Surveillance
`
`Inside individual offices, the autonomous video surveillance system will maintain a situational aware(cid:173)
`ness record of events and signal alarms for a variety of specified conditions. For example, an alarm may be
`specified for events in which someone enters the office and places a briefcase on the desk, but not if the
`person leaves a document on the desk. Using contextual information such as time of day and access control
`identification, the system will report other alarm conditions that are functions of the number of people in
`the room and what they do. A representative office surveillance scene is shown in Figure 5.
`3.3 Perimeter Surveillance
`
`For perimeter monitoring scenarios, a TI NightSight infrared camera will provide video data for the
`surveillance system to monitor areas outside the building at night. For example, the system could monitor a
`building entrance and signal an alarm if someone walks by and leaves a box outside the door (e.g., as illus(cid:173)
`trated in Figure 6), but not if someone loiters without leaving a box.
`
`150
`
`

`

`•
`
`t
`
`;
`
`.,
`'
`
`t
`
`The ability to reliably discriminate significant and insignificant events is important to reduce false
`alarms for physical security applications, and poses many challenges for computer vision research to ad(cid:173)
`vance the state of the art.
`
`4. Acknowledgements
`
`The autonomous video surveillance capabilities described in this paper represent the research of many
`members of the technical staff at TI over the past five years. Among them the primary contributors are Tom
`Bannon, Frank Brill, Jon Courtney, Chris Kellogg, and Kashi Rao.
`
`5. Bibliography
`
`[1] Bannon, T., "Visual Memory Prototype Demonstration", TI Internal Technical Report CSL-ITR-93-01-
`25, Jan. 1993.

`[2] Courtney, J. "Automatic video indexing via object motion analysis", TI Internal Technical Report, to
`appear in Pattern Recognition.
`[3] Flinchbaugh, B., and T. Bannon, "Autonomous Scene Monitoring System", Proc. 10th Annual Joint
`Government-Industry Security Technology Symposium, American Defense Preparedness Association,
`June 1994.
`[4] Kellogg, C., "Visual Memory", TI Internal Technical Report, also MIT Media Lab MS thesis, 1993.
`
`[5] Rao, K., and P. Sarwal, "A computer vision system to detect 3-D rectangular objects", TI Internal Tech(cid:173)
`nical Report. Accepted for IEEE Workshop on Applications of Computer Vision (WACV 96).
`
`[6] Vigil, J., "An evaluation of exterior video motion detection systems. Volume 1: Intrusion detection
`tests. Volume 2: Nuisance alarm results", Sandia Reports SAND92-0108/l, SAND92-0108/2, Sandia
`National Laboratories, 1992 .
`
`151
`
`

This document is available on Docket Alarm but you must sign up to view it.


Or .

Accessing this document will incur an additional charge of $.

After purchase, you can access this document again without charge.

Accept $ Charge
throbber

Still Working On It

This document is taking longer than usual to download. This can happen if we need to contact the court directly to obtain the document and their servers are running slowly.

Give it another minute or two to complete, and then try the refresh button.

throbber

A few More Minutes ... Still Working

It can take up to 5 minutes for us to download a document if the court servers are running slowly.

Thank you for your continued patience.

This document could not be displayed.

We could not find this document within its docket. Please go back to the docket page and check the link. If that does not work, go back to the docket and refresh it to pull the newest information.

Your account does not support viewing this document.

You need a Paid Account to view this document. Click here to change your account type.

Your account does not support viewing this document.

Set your membership status to view this document.

With a Docket Alarm membership, you'll get a whole lot more, including:

  • Up-to-date information for this case.
  • Email alerts whenever there is an update.
  • Full text search for other cases.
  • Get email alerts whenever a new case matches your search.

Become a Member

One Moment Please

The filing “” is large (MB) and is being downloaded.

Please refresh this page in a few minutes to see if the filing has been downloaded. The filing will also be emailed to you when the download completes.

Your document is on its way!

If you do not receive the document in five minutes, contact support at support@docketalarm.com.

Sealed Document

We are unable to display this document, it may be under a court ordered seal.

If you have proper credentials to access the file, you may proceed directly to the court's system using your government issued username and password.


Access Government Site

We are redirecting you
to a mobile optimized page.





Document Unreadable or Corrupt

Refresh this Document
Go to the Docket

We are unable to display this document.

Refresh this Document
Go to the Docket