throbber
APPLICATION FOR UNITED STATES PATENT
`
`INVENTORS:
`
`ALAN J. LIPTON
`THOMAS M. STRAT
`PETER L. VENETIANER
`MARK C. ALLMEN
`WILLIAM E. SEVERSON
`NIELS HAERING
`ANDREW J. CHOSAK
`ZHONG ZHANG
`MATTHEW F. FRAZIER
`JAMES S. SFEK.AS
`TASUK! HIRATA
`JOHN CLARK
`
`TITLE:
`
`VIDEO SURVEILLANCE SYSTEM EMPLOYING
`VIDEO PRIMITNES
`
`ATTORNEYS' ADDRESS:
`
`VENABLE
`1201 New York Avenue, N.W., Suite 1000
`Washington, D.C. 20005-3917
`Telephone: (202) 962-4800
`Telefax: (202) 962-8300
`
`ADDRESS FOR U.S.P.T.O. CORRESPONDENCE:
`
`VENABLE
`Post Office Box 34385
`Washington, D.C. 20043-9998
`
`ATTORNEY DOCKET NO.:
`
`37112-175340
`
`

`

`VIDEO SURVEILLANCE SYSTEM EMPLOYING VIDEO PRIMITIVES
`
`CROSS-REFERENCE TO RELATED APPLICATIONS
`
`[1]
`
`This application claims the priority of U.S. Patent Application No. 09/694,712
`
`5
`
`filed October 24, 2000, which is incorporated herein by reference.
`
`BACKGROUND OF THE INVENTION
`
`Field of the Invention
`
`[2]
`
`The invention relates to a system for automatic video surveillance employing
`
`~ video primitives.
`
`References
`
`[3]
`
`For the convenience of the reader, the references referred to herein are listed
`
`below. In the specification, the numerals within brackets refer to respective references. The
`
`!il5
`
`listed references are incorporated herein by reference.
`
`[4]
`
`[5]
`
`The following references describe moving target detection:
`
`{1} A. Lipton, H. Fujiyoshi and R. S. Patil, "Moving Target Detection and
`
`Classification from Real-Time Video," Proceedings ofiEEE WACV '98, Princeton, NJ, 1998,
`
`pp. 8-14.
`
`20
`
`[6]
`
`{2} W.E.L. Grimson, et al., ''Using Adaptive Tracking to Classify and Monitor
`
`Activities in a Site", CVPR, pp. 22-29, June 1998.
`
`[7]
`
`{3} A.J. Lipton, H. Fujiyoshi, R.S. Patil, "Moving Target Classification and
`
`Tracking from Real-time Video," IUW, pp. 129-136, 1998.
`
`-1-
`
`(37112-175340)
`
`

`

`[8]
`
`{ 4} T .J. Olson and F .Z. Brill, "Moving Object Detection and Event Recognition
`
`Algorithm for Smart Cameras," IUW, pp. 159-175, May 1997.
`
`[9]
`
`The following references describe detecting and tracking humans:
`
`[10]
`
`{5} A. J. Lipton, "Local Application of Optical Flow to Analyse Rigid Versus
`
`5 Non-Rigid Motion," International Conference on Computer Vision, Corfu, Greece, September
`
`1999.
`
`[11]
`
`{ 6} F. Bartolini, V. Cappellini, and A. Mecocci, "Counting people getting in and
`
`out of a bus by real-time image-sequence processing," IVC, 12(1):36-41, January 1994.
`
`[12]
`
`{7} M. Rossi and A. Bozzoli, "Tracking and counting moving people," ICIP94,
`
`[13]
`
`{8} C.R. Wren, A. Azarbayejani, T. Darrell, and A. Pentland, "Pfinder: Real-
`
`time tracking of the human body," Vismod, 1995.
`
`[14]
`
`{9} L. Khoudour, L. Duvieubourg, J.P. Deparis, "Real-Time Pedestrian Counting
`
`by Active Linear Cameras," JEI, 5(4):452-459, October 1996.
`
`15
`
`[15]
`
`{10} S. loffe, D.A. Forsyth, ~'Probabilistic Methods for Finding People," IJCV,
`
`43(1):45-68, June 2001.
`
`[16]
`
`{ 11} M. Isard and J. MacCormick, "BraMBLe: A Bayesian Multiple-Blob
`
`Tracker," ICCV, 2001.
`
`[17] The following references describe blob analysis:
`
`20
`
`(18]
`
`{12} D.M. Gavrila, "The Visual Analysis of Human Movement: A Survey,"
`
`CVIU, 73(1):82-98, January 1999.
`
`[19]
`
`{ 13} Niels Haering and Niels da Vitoria Lobo, "Visual Event Detection," Video
`
`Computing Series, Editor Mubarak Shah, 2001.
`
`- 2-
`
`(37112-175340)
`
`

`

`[20] The following references describe blob analysis for trucks, cars, and people:
`
`[21]
`
`{14} Collins, Lipton, Kanade, Fujiyoshi, Duggins, Tsin, Tolliver, Enomoto, and
`
`Hasegawa, "A System for Video Surveillance and Monitoring: VSAM Final Report," Technical
`
`Report CMU-Rl-TR-00-12, Robotics Institute, Carnegie Mellon University, May 2000.
`
`5
`
`[22]
`
`{15} Lipton, Fujiyoshi, and Patil, "Moving Target Classification and Tracking
`
`from Real-time Video," 98 Darpa IUW, Nov. 20-23, 1998.
`
`[23] The following reference describes analyzing a single-person blob and its contours:
`
`[24]
`
`{16} C.R. Wren, A. Azarbayejani, T. Darrell, and A.P. Pentland. "Pfinder: Real-
`
`Time Tracking of the Human Body," PAMI, vol19, pp. 780-784, 1997.
`
`{~
`
`[25] The following reference describes internal motion of blobs, including any motion-
`
`based segmentation:
`
`[26]
`
`{17} M. Allmen and C. Dyer, "Long--Range Spatiotemporal Motion
`
`Understanding Using Spatiotemporal Flow Curves," Proc. IEEE CVPR, Lahaina, Maui, Hawaii,
`
`pp. 303-309, 1991.
`
`15
`
`[27]
`
`{18} L. Wixson, "Detecting Salient Motion by Accumulating Directionally
`
`Consistent Flow", IEEE Trans. Pattern Anal. Mach. Intell., vol. 22, pp. 774-781, Aug, 2000.
`
`Background of the Invention
`
`[28] Video surveillance of public spaces has become extremely widespread and
`
`20
`
`accepted by the general public. Unfortunately, conventional video surveillance systems produce
`
`such prodigious volumes of data that an intractable problem results in the analysis of video
`
`surveillance data.
`
`- 3-
`
`(37112-175340)
`
`

`

`[29] A need exists to reduce the amount of video surveillance data so analysis ofthe
`
`video surveillance data can be conducted.
`
`[30] A need exists to filter video surveillance data to identify desired portions of the
`
`video surveillance data.
`
`5
`
`'rb
`
`~~~:o~
`
`SUMMARY OF THE INVENTION
`
`[31] An object of the invention is to reduce the amount of video surveillance data so
`
`analysis of the video surveillance data can be conducted.
`
`[32] An object of the invention is to filter video surveillance data to identify desired
`
`portions of the video surveillance data.
`
`[33] An object of the invention is to produce a real time alarm based on an automatic
`
`detection of an event from video surveillance data.
`
`[34] An object of the invention is to integrate data from surveillance sensors other than
`
`video for improved searching capabilities.
`
`15
`
`[35] An object of the invention is to integrate data from surveillance sensors other than
`
`video for improved event detection capabilities
`
`[36] The invention includes an article of manufacture, a method, a system, and an
`
`apparatus for video surveillance.
`
`[37]
`
`The article of manufacture of the invention includes a computer-readable medium
`
`20
`
`comprising software for a video surveillance system, comprising code segments for operating the
`
`video surveillance system based on video primitives.
`
`[38]
`
`The article of manufacture of the invention includes a computer-readable medium
`
`comprising software for a video surveillance system, comprising code segments for accessing
`
`-4-
`
`(37112-175340)
`
`

`

`archived video primitives, and code segments for extracting event occurrences from accessed
`
`archived video primitives.
`
`[39] The system of the invention includes a computer system including a computer-
`
`readable medium having software to operate a computer in accordance with the invention.
`
`5
`
`[ 40] The apparatus of the invention includes a computer including a computer-readable
`
`medium having software to operate the computer in accordance with the invention.
`
`[41] The article of manufacture ofthe invention includes a computer-readable medium
`
`having software to operate a computer in accordance with the invention.
`
`[ 42] Moreover, the above objects and advantages of the invention are illustrative, and
`
`not exhaustive, of those that can be achieved by the invention. Thus, these and other objects and
`
`advantages of the invention will be apparent from the description herein, both as embodied
`
`"'
`
`herein and as modified in view of any variations which will be apparent to those skilled in the
`
`art.
`
`l5 Definitions
`
`[43] A "video" refers to motion pictures represented in analog and/or digital form.
`
`Examples of video include: television, movies, image sequences from a video camera or other
`
`observer, and computer-generated image sequences.
`
`[44] A "frame" refers to a particular image or other discrete unit within a video.
`
`20
`
`[ 45] An "object" refers to an item of interest in a video. Examples of an object
`
`include: a person, a vehicle, an animal, and a physical subject.
`
`- 5-
`
`(37112-175340)
`
`

`

`[46] An "activity" refers to one or more actions and/or one or more composites of
`
`actions of one or more objects. Examples of an activity include: entering; exiting; stopping;
`
`moving; raising; lowering; growing; and shrinking.
`
`[ 4 7] A "location" refers to a space where an activity may occur. A location can be, for
`
`5
`
`example, scene-based or image-based. Examples of a scene-based location include: a public
`
`space; a store; a retail space; an office; a warehouse; a hotel room; a hotel lobby; a lobby of a
`
`building; a casino; a bus station; a train station; an airport; a port; a bus; a train; an airplane; and
`
`a ship. Examples of an image-based location include: a video image; a line in a video image; an
`
`area in a video image; a rectangular section of a video image; and a polygonal section of a video
`
`l-<9
`
`image.
`
`[48] An "event" refers to one or more objects engaged in an activity. The event may
`
`be referenced with respect to a location and/or a time.
`
`[ 49] A "computer" refers to any apparatus that is capable of accepting a structured
`
`input, processing the structured input according to prescribed rules, and producing results of the
`
`15
`
`processing as output. Examples of a computer include: a computer; a general purpose computer;
`
`a supercomputer; a mainframe; a super mini-computer; a mini-computer; a workstation; a micro(cid:173)
`
`computer; a server; an interactive television; a hybrid combination of a computer and an
`
`interactive television; and application-specific hardware to emulate a computer and/or software.
`
`A computer can have a single processor or multiple processors, which can operate in parallel
`
`20
`
`and/or not in parallel. A computer also refers to two or more computers connected together via a
`
`network for transmitting or receiving information between the computers. An example of such a
`
`computer includes a distributed computer system for processing information via computers
`
`linked by a network.
`
`- 6-
`
`(37112-175340)
`
`

`

`[50] A "computer-readable medium" refers to any storage device used for storing data
`
`accessible by a computer. Examples of a computer-readable medium include: a magnetic hard
`
`disk; a floppy disk; an optical disk, such as a CD-ROM and a DVD; a magnetic tape; a memory
`
`chip; and a carrier wave used to carry computer-readable electronic data, such as those used in
`
`5
`
`transmitting and receiving e-mail or in accessing a network.
`
`[51]
`
`"Software" refers to prescribed rules to operate a computer. Examples of
`
`software include: software; code segments; instructions; computer programs; and programmed
`
`logic.
`
`[52] A "computer system" refers to a system having a computer, where the computer
`
`comprises a computer-readable medium embodying software to operate the computer.
`
`[53] A "network" refers to a number of computers and associated devices that are
`
`"'
`
`connected by communication facilities. A network involves permanent connections such as
`
`cables or temporary connections such as those made through telephone or other communication
`
`links. Examples of a network include: an internet, such as the Internet; an intranet; a local area
`
`~t5
`
`network (LAN); a wide area network (WAN); and a combination of networks, such as an internet
`
`and an intranet.
`
`BRIEF DESCRIPTION OF THE DRAWINGS
`
`[54] Embodiments ofthe invention are explained in greater detail by way of the
`
`20
`
`drawings, where the same reference numerals refer to the same features.
`
`[55]
`
`Figure 1 illustrates a plan view of the video surveillance system of the invention.
`
`[56]
`
`Figure 2 illustrates a flow diagram for the video surveillance system of the
`
`invention.
`
`-7-
`
`(37112-175340)
`
`

`

`[57]
`
`Figure 3 illustrates a flow diagram for tasking the video surveillance system.
`
`[58]
`
`Figure 4 illustrates a flow diagram for operating the video surveillance system.
`
`[59]
`
`Figure 5 illustrates a flow diagram for extracting video primitives for the video
`
`surveillance system.
`
`5
`
`[60]
`
`Figure 6 illustrates a flow diagram for taking action with the video surveillance
`
`system.
`
`[61]
`
`Figure 7 illustrates a flow diagram for semi-automatic calibration of the video
`
`surveillance system.
`
`[62]
`
`Figure 8 illustrates a flow diagram for automatic calibration of the video
`
`f~
`. ')~.:.:;~g
`
`surveillance system .
`
`[63]
`
`Figure 9 illustrates an additional flow diagram for the video surveillance system
`
`of the invention.
`
`[64]
`
`Figures 10-15 illustrate examples of the video surveillance system of the
`
`invention applied to monitoring a grocery store.
`
`15
`
`DETAILED DESCRIPTION OF THE INVENTION
`
`[ 65]
`
`The automatic video surveillance system of the invention is for monitoring a
`
`location for, for example, market research or security purposes. The system can be a dedicated
`
`video surveillance installation with purpose-built surveillance components, or the system can be
`
`20
`
`a retrofit to existing video surveillance equipment that piggybacks off the surveillance video
`
`feeds. The system is capable of analyzing video data from live sources or from recorded media.
`
`The system can have a prescribed response to the analysis, such as record data, activate an alarm
`
`mechanism, or active another sensor system. The system is also capable of integrating with
`
`- 8-
`
`(37112-175340)
`
`

`

`other surveillance system components. The system produces security or market research reports
`
`that can be tailored according to the needs of an operator and, as an option, can be presented
`
`through an interactive web-based interface, or other reporting mechanism.
`
`[ 66] An operator is provided with maximum flexibility in configuring the system by
`
`5
`
`using event discriminators. Event discriminators are identified with one or more objects (whose
`
`descriptions are based on video primitives), along with one or more optional spatial attributes,
`
`and/or one or more optional temporal attributes. For example, an operator can define an event
`
`discriminator (called a "loitering" event in this example) as a "person" object in the "automatic
`
`teller machine" space for "longer than 15 minutes" and "between 10:00 p.m. and 6:00 a.m."
`
`[67] Although the video surveillance system of the invention draws on well-known
`
`computer vision techniques from the public domain, the inventive video surveillance system has
`
`several unique and novel features that are not currently available. For example, current video
`
`surveillance systems use large volumes of video imagery as the primary commodity of
`
`information interchange. The system of the invention uses video primitives as the primary
`
`15
`
`commodity with representative video imagery being used as collateral evidence. The system of
`
`the invention can also be calibrated (manually, semi-automatically, or automatically) and
`
`thereafter automatically can infer video primitives from video imagery. The system can further
`
`analyze previously processed video without needing to reprocess completely the video. By
`
`analyzing previously processed video, the system can perform inference analysis based on
`
`20
`
`previously recorded video primitives, which greatly improves the analysis speed of the computer
`
`system.
`
`[68] As another example, the system of the invention provides unique system tasking.
`
`Using equipment control directives, current video systems allow a user to position video sensors
`
`- 9-
`
`(37112-175340)
`
`

`

`and, in some sophisticated conventional systems, to mask out regions of interest or disinterest.
`
`Equipment control directives are instructions to control the position, orientation, and focus of
`
`video cameras. Instead of equipment control directives, the system of the invention uses event
`
`discriminators based on video primitives as the primary tasking mechanism. With event
`
`5
`
`discriminators and video primitives, an operator is provided with a much more intuitive approach
`
`over conventional systems for extracting useful information from the system. Rather than
`
`tasking a system with an equipment control directives, such as "camera A pan 45 degrees to the
`
`left," the system of the invention can be tasked in a human-intuitive manner with one or more
`
`event discriminators based on video primitives, such as "a person enters restricted area A."
`
`[69] Using the invention for market research, the following are examples of the type of
`
`video surveillance that can be performed with the invention: counting people in a store; counting
`
`people in a part of a store; counting people who stop in a particular place in a store; measuring
`
`how long people spend in a store; measuring how long people spend in a part of a store; and
`
`measuring the length of a line in a store.
`
`15
`
`[70] Using the invention for security, the following are examples of the type of video
`
`surveillance that can be performed with the invention: determining when anyone enters a
`
`restricted area and storing associated imagery; determining when a person enters an area at
`
`unusual times; determining when changes to shelf space and storage space occur that might be
`
`unauthorized; determining when passengers aboard an aircraft approach the cockpit; determining
`
`20 when people tailgate through a secure portal; determining if there is an unattended bag in an
`
`airport; and determining if there is a theft of an asset.
`
`[71]
`
`Figure 1 illustrates a plan view of the video surveillance system of the invention.
`
`A computer system 11 comprises a computer 12 having a computer-readable medium 13
`
`- 10-
`
`(37112-175340)
`
`

`

`embodying software to operate the computer 12 according to the invention. The computer
`
`system 11 is coupled to one or more video sensors 14, one or more video recorders 15, and one
`
`or more input/output (I/0) devices 16. The video sensors 14 can also be optionally coupled to
`
`the video recorders 15 for direct recording of video surveillance data. The computer system is
`
`5
`
`optionally coupled to other sensors 17.
`
`[72]
`
`The video sensors 14 provide source video to the computer system 11. Each
`
`video sensor 14 can be coupled to the computer system 11 using, for example, a direct
`
`connection (e.g., a firewire digital camera interface) or a network. The video sensors 14 can
`
`exist prior to installation of the invention or can be installed as part of the invention. Examples
`
`7~~=
`~l.5
`'lj)
`
`of a video sensor 14 include: a video camera; a digital video camera; a color camera; a
`
`monochrome camera; a camera; a camcorder, a PC camera; a webcam; an infra-red video
`
`camera; and a CCTV camera.
`
`[73] The video recorders 15 receive video surveillance data from the computer system
`
`11 for recording and/ or provide source video to the computer system 11. Each video recorder 15
`
`can be coupled to the computer system 11 using, for example, a direct connection or a network.
`
`The video recorders 15 can exist prior to installation of the invention or can be installed as part
`
`of the invention. Examples of a video recorder 15 include: a video tape recorder; a digital video
`
`recorder; a video disk; a DVD; and a computer-readable medium.
`
`[7 4]
`
`The I/0 devices 16 provide input to and receive output from the computer system
`
`20
`
`11. The I/0 devices 16 can be used to task the computer system 11 and produce reports from the
`
`computer system 11. Examples of I/0 devices 16 include: a keyboard; a mouse; a stylus; a
`
`monitor; a printer; another computer system; a network; and an alarm.
`
`- 11 -
`
`(37112-175340)
`
`

`

`[75] The other sensors 17 provide additional input to the computer system 11. Each
`
`other sensor 17 can be coupled to the computer system 11 using, for example, a direct connection
`
`or a network. The other sensors 17 can exit prior to installation of the invention or can be
`
`installed as part of the invention. Examples of another sensor 17 include: a motion sensor; an
`
`5
`
`optical tripwire; a biometric sensor; and a card-based or keypad-based authorization system. The
`
`outputs of the other sensors 17 can be recorded by the computer system 11, recording devices,
`
`and/or recording systems.
`
`(76]
`
`Figure 2 illustrates a flow diagram for the video surveillance system of the
`
`invention. Various aspects of the invention are exemplified with reference to Figures 10-15,
`
`which illustrate examples of the video surveillance system of the invention applied to monitoring
`
`a grocery store.
`
`[77]
`
`In block 21, the video surveillance system is set up as discussed for Figure 1.
`
`Each video sensor 14 is orientated to a location for video surveillance. The computer system 11
`
`is connected to the video feeds from the video equipment 14 and 15. The video surveillance
`
`~f5
`
`system can be implemented using existing equipment or newly installed equipment for the
`
`location.
`
`[78]
`
`In block 22, the video surveillance system is calibrated. Once the video
`
`surveillance system is in place from block 21, calibration occurs. The result of block 22 is the
`
`ability of the video surveillance system to determine an approximate absolute size and speed of a
`
`20
`
`particular object (e.g., a person) at various places in the video image provided by the video
`
`sensor. The system can be calibrated using manual calibration, semi-automatic calibration, and
`
`automatic calibration. Calibration is further described after the discussion of block 24.
`
`- 12-
`
`(37112-175340)
`
`

`

`[79]
`
`In block 23 of Figure 2, the video surveillance system is tasked. Tasking occurs
`
`after calibration in block 22 and is optional. Tasking the video surveillance system involves
`
`specifying one or more event discriminators. Without tasking, the video surveillance system
`
`operates by detecting and archiving video primitives and associated video imagery without
`
`5
`
`taking any action, as in block 45 in Figure 4.
`
`[80]
`
`Figure 3 illustrates a flow diagram for tasking the video surveillance system to
`
`determine event discriminators. An event discriminator refers to one or more objects optionally
`
`interacting with one or more spatial attributes and/or one or more temporal attributes. An event
`
`discriminator is described in terms of video primitives. A video primitive refers to an observable
`
`attribute of an object viewed in a video feed. Examples of video primitives include the
`
`following: a classification; a size; a shape; a color; a texture; a position; a velocity; a speed; an
`
`internal motion; a motion; a salient motion; a feature of a salient motion; a scene change; a
`
`feature of a scene change; and a pre-defined model.
`
`[81] A classification refers to an identification of an object as belonging to a particular
`
`15
`
`category or class. Examples of a classification include: a person; a dog; a vehicle; a police car;
`
`an individual person; and a specific type of object.
`
`[82] A size refers to a dimensional attribute of an object. Examples of a size include:
`
`large; medium; small; :flat; taller than 6 feet; shorter than 1 foot; wider than 3 feet; thinner than 4
`
`feet; about human size; bigger than a human; smaller than a human; about the size of a car; a
`
`20
`
`rectangle in an image with approximate dimensions in pixels; and a number of image pixels.
`
`[83] A color refers to a chromatic attribute of an object. Examples of a color include:
`
`white; black; grey; red; a range ofHSV values; a range ofYUV values; a range ofRGB values;
`
`an average RGB value; an average YUV value; and a histogram ofRGB values.
`
`- 13-
`
`(37112-175340)
`
`

`

`[84] A texture refers to a pattern attribute of an object. Examples of texture features
`
`include: self-similarity; spectral power; linearity; and coarseness.
`
`[85] An internal motion refers to a measure of the rigidity of an object. An example of
`
`a fairly rigid object is a car, which does not exhibit a great amount of internal motion. An
`
`5
`
`example of a fairly non-rigid object is a person having swinging arms and legs, which exhibits a
`
`great amount of internal motion.
`
`[86] A motion refers to any motion that can be automatically detected. Examples of a
`
`motion include: appearance of an object; disappearance of an object; a vertical movement of an
`
`object; a horizontal movement of an object; and a periodic movement of an object.
`
`[87] A salient motion refers to any motion that can be automatically detected and can
`
`be tracked for some period of time. Such a moving object exhibits apparently purposeful
`
`motion. Examples of a salient motion include: moving from one place to another; and moving to
`
`interact with another object.
`
`[88] A feature of a salient motion refers to a property of a salient motion. Examples of
`
`15
`
`a feature of a salient motion include: a trajectory; a length of a trajectory in image space; an
`
`approximate length of a trajectory in a three-dimensional representation of the environment; a
`
`position of an object in image space as a function of time; an approximate position of an object
`
`in a three-dimensional representation of the environment as a function of time; a duration of a
`
`trajectory; a velocity (e.g., speed and direction) in image space; an approximate velocity (e.g.,
`
`20
`
`speed and direction) in a three-dimensional representation of the environment; a duration of time
`
`at a velocity; a change of velocity in image space; an approximate change of velocity in a three-
`
`dimensional representation of the environment; a duration of a change of velocity; cessation of
`
`motion; and a duration of cessation of motion. A velocity refers to the speed and direction of an
`
`- 14-
`
`(37112-175340)
`
`

`

`object at a particular time. A trajectory refers a set of (position, velocity) pairs for an object for
`
`as long as the object can be tracked or for a time period.
`
`[89] A scene change refers to any region of a scene that can be detected as changing
`
`over a period of time. Examples of a scene change include: an stationary object leaving a scene;
`
`5
`
`an object entering a scene and becoming stationary; an object changing position in a scene; and
`
`an object changing appearance (e.g. color, shape, or size).
`
`[90] A feature of a scene change refers to a property of a scene change. Examples of a
`
`feature of a scene change include: a size of a scene change in image space; an approximate size
`
`of a scene change in a three-dimensional representation of the environment; a time at which a
`
`scene change occurred; a location of a scene change in image space; and an approximate location
`
`of a scene change in a three-dimensional representation of the environment.
`
`[91] A pre..:defined model refers to an a priori known model of an object. Examples of
`
`a pre-defined include: an adult; a child; a vehicle; and a semi-trailer.
`
`[92]
`
`In block 31, one or more objects types of interests are identified in terms of video
`
`15
`
`primitives or abstractions thereof. Examples of one or more objects include: an object; a person;
`
`a red object; two objects; two persons; and a vehicle.
`
`[93]
`
`In block 32, one or more spatial areas of interest are identified. An area refers to
`
`one or more portions of an image from a source video or a spatial portion of a scene being
`
`viewed by a video sensor. An area also includes a combination of areas from various scenes
`
`20
`
`and!or images. An area can be an image-based space (e.g., a line, a rectangle, a polygon, or a
`
`circle in a video image) or a three-dimensional space (e.g., a cube, or an area of floor space in a
`
`building).
`
`- 15-
`
`(37112-175340)
`
`

`

`[94]
`
`Figure 12 illustrates identifying areas along an aisle in a grocery store. Four areas
`
`are identified: coffee; soda promotion; chips snacks; and bottled water. The areas are identified
`
`via a point-and-click interface with the system.
`
`[95]
`
`In block 33, one or more temporal attributes of interest are optionally identified.
`
`5
`
`Examples of a temporal attribute include: every 15 minutes; between 9:00p.m. to 6:30 a.m.; less
`
`than 5 minutes; longer than 30 seconds; over the weekend; and within 20 minutes of.
`
`[96]
`
`In block 34, a response is optionally identified. Examples of a response includes
`
`the following: activating a visual and/or audio alert on a system display; activating a visual
`
`and/or audio alarm system at the location; activating a silent alarm; activating a rapid response
`
`~~:Fi:
`
`~:to mechanism; locking a door; contacting a security service; forwarding data (e.g., image data,
`
`video data, video primitives; and/or analyzed data) to another computer system via a network,
`
`such as the Internet; saving such data to a designated computer-readable medium; activating
`
`some other sensor or surveillance system; tasking the computer system 11 and/or another
`
`computer system; and directing the computer system 11 and/or another computer system.
`
`[97]
`
`In block 35, one or more discriminators are identified by describing interactions
`
`between video primitives (or their abstractions), spatial areas of interest, and temporal attributes
`
`of interest. An interaction is determined for a combination of one or more objects identified in
`
`block 31, one or more spatial areas of interest identified in block 32, and one or more temporal
`
`attributes of interest identified in block 33. One or more responses identified in block 34 are
`
`20
`
`optionally associated with each event discriminator.
`
`[98] Examples of an event discriminator for a single object include: an object appears;
`
`a person appears; and a red object moves faster than 1 Ornls.
`
`- 16-
`
`(37112-175340)
`
`

`

`[99] Examples of an event discriminator for multiple objects include: two objects
`
`come together; a person exits a vehicle; and a red object moves next to a blue object.
`
`[1 00] Examples of an event discriminator for an object and a spatial attribute include:
`
`an object crosses a line; an object enters an area; and a person crosses a line from the left.
`
`5
`
`[1 01] Examples of an event discriminator for an object and a temporal attribute include:
`
`an object appears at 10:00 p.m.; a person travels faster then 2m/s between 9:00 a.m. and 5:00
`
`p.m.; and a vehicle appears on the weekend.
`
`[1 02] Examples of an event discriminator for an object, a spatial attribute, and a
`
`temporal attribute include: a person crosses a line between midnight and 6:00a.m.; and a vehicle
`
`~p stops in an area for longer than 1 0 minutes.
`
`[103] An example of an event discriminator for an object, a spatial attribute, and a
`
`temporal attribute associated with a response include: a person enters an area between midnight
`
`and 6:00 a.m., and a security service is notified.
`
`[104]
`
`In block 24 of Figure 2, the video surveillance system is operated. The video
`
`15
`
`surveillance system of the invention operates automatically, detects and archives video
`
`primitives of objects in the scene, and detects event occurrences in real time using event
`
`discriminators. In addition, action is taken in real time, as appropriate, such as activating alarms,
`
`generating reports, and generating output. The reports and output can be displayed and/or stored
`
`locally to the system or elsewhere via a network, such as the Internet. Figure 4 illustrates a flow
`
`20
`
`diagram for operating the video surveillance system.
`
`[105] In block 41, the computer system 11 obtains source video from the video sensors
`
`14 and! or the video recorders 15.
`
`- 17-
`
`(37112-175340)
`
`

`

`[1 06]
`
`In block 42, video primitives are extracted in real time from the source video. As
`
`an option, non-video primitives can be obtained and/or extracted from one or more other sensors
`
`17 and used with the invention. The extraction of video primitives is illustrated with Figure 5.
`
`[1 07] Figure 5 illustrates a flow diagram for extracting video primitives for the video
`
`5
`
`surveillance system. Blocks 51 and 52 operate in parallel and can be performed in any order or
`
`concurrently. In block 51, objects are detected via movement. Any motion detection algorithm
`
`for detecting movement between frames at the pixel level can be used for this block. As an
`
`example, the three frame differencing technique can be used, which is discussed in { 1 } . The
`
`detected objects are forwarded to block 53.
`
`:~t:o
`
`[108]
`
`In block 52, objects are detected via change. Any change detection algorithm for
`
`detecting changes from a background model can be used for this block. An object is detected in
`
`'"
`
`this block if one or more pixels in a frame are deemed to be in the foreground of the frame
`
`because the pixels do not conform to a background model of the frame. As an example, a
`
`stochastic background modeling technique, such as dynamically adaptive background
`
`1~1'5
`
`subtraction, can be used, which is

This document is available on Docket Alarm but you must sign up to view it.


Or .

Accessing this document will incur an additional charge of $.

After purchase, you can access this document again without charge.

Accept $ Charge
throbber

Still Working On It

This document is taking longer than usual to download. This can happen if we need to contact the court directly to obtain the document and their servers are running slowly.

Give it another minute or two to complete, and then try the refresh button.

throbber

A few More Minutes ... Still Working

It can take up to 5 minutes for us to download a document if the court servers are running slowly.

Thank you for your continued patience.

This document could not be displayed.

We could not find this document within its docket. Please go back to the docket page and check the link. If that does not work, go back to the docket and refresh it to pull the newest information.

Your account does not support viewing this document.

You need a Paid Account to view this document. Click here to change your account type.

Your account does not support viewing this document.

Set your membership status to view this document.

With a Docket Alarm membership, you'll get a whole lot more, including:

  • Up-to-date information for this case.
  • Email alerts whenever there is an update.
  • Full text search for other cases.
  • Get email alerts whenever a new case matches your search.

Become a Member

One Moment Please

The filing “” is large (MB) and is being downloaded.

Please refresh this page in a few minutes to see if the filing has been downloaded. The filing will also be emailed to you when the download completes.

Your document is on its way!

If you do not receive the document in five minutes, contact support at support@docketalarm.com.

Sealed Document

We are unable to display this document, it may be under a court ordered seal.

If you have proper credentials to access the file, you may proceed directly to the court's system using your government issued username and password.


Access Government Site

We are redirecting you
to a mobile optimized page.





Document Unreadable or Corrupt

Refresh this Document
Go to the Docket

We are unable to display this document.

Refresh this Document
Go to the Docket