throbber
APPLICATION FOR UNITED STATES PATENT
`
`
`
`INVENTORS:
`
`ALAN J. LIPTON
`THOMASM. STRAT
`PETER L. VENETIANER
`MARK C. ALLMEN
`WILLIAM E. SEVERSON
`NIELS HAERING
`ANDREW J. CHOSAK.
`ZHONG ZHANG
`MATTHEWF. FRAZIER
`JAMESS. SFEKAS
`TASUKI HIRATA
`JOHN CLARK
`
`TITLE:
`
`VIDEO SURVEILLANCE SYSTEM EMPLOYING
`VIDEO PRIMITIVES
`
`ATTORNEYS' ADDRESS:
`
`VENABLE
`1201 New York Avenue, N.W., Suite 1000
`Washington, D.C. 20005-3917
`Telephone: (202) 962-4800
`Telefax: (202) 962-8300
`
`ADDRESSFOR U.S.P.T.O. CORRESPONDENCE:
`
`VENABLE
`Post Office Box 34385
`Washington, D.C. 20043-9998
`
`ATTORNEY DOCKET NO.:
`
`37112-175340
`
`Canon Ex. 1040 Page 1 of 31
`
`Canon Ex. 1040 Page 1 of 31
`
`

`

`VIDEO SURVEILLANCE SYSTEM EMPLOYING VIDEO PRIMITIVES
`
`CROSS-REFERENCE TO RELATED APPLICATIONS
`
`[1]
`
`This application claimsthepriority of U.S. Patent Application No. 09/694,712
`
`filed October 24, 2000, which is incorporated herein by reference.
`
`BACKGROUND OF THE INVENTION
`
`Field of the Invention
`
`[2]
`
`The invention relates to a system for automatic video surveillance employing
`
`video primitives.
`
`References
`
`[3]
`
`For the convenienceofthe reader, the references referred to herein are listed
`
`below. In the specification, the numerals within brackets refer to respective references. The
`
`listed references are incorporated herein by reference.
`
`[4]
`
`[5]
`
`The following references describe movingtarget detection:
`
`{1} A. Lipton, H. Fujiyoshi andR.S. Patil, “Moving Target Detection and
`
`Classification from Real-Time Video,” Proceedings of IEEE WACV °98, Princeton, NJ, 1998,
`
`
`
`
`
`pp. 8-14.
`
`20
`
`[6]
`
`{2} W.ELL. Grimson,et al., “Using Adaptive Tracking to Classify and Monitor
`
`Activities in a Site”, CVPR, pp. 22-29, June 1998.
`
`[7]
`
`{3} A.J. Lipton, H. Fujiyoshi, R.S. Patil, “Moving Target Classification and
`
`Tracking from Real-time Video,” IUW,pp. 129-136, 1998.
`
`-1-
`
`(37112-175340)
`
`Canon Ex. 1040 Page 2 of 31
`
`Canon Ex. 1040 Page 2 of 31
`
`

`

`[8]
`
`{4} T.J. Olson and F.Z.Brill, “Moving Object Detection and Event Recognition
`
`
`Algorithm for Smart Cameras,” IUW,pp. 159-175, May 1997.
`
`[9]
`
`The following references describe detecting and tracking humans:
`
`[10]
`
`{5} A.J. Lipton, “Local Application of Optical Flow to Analyse Rigid Versus
`
`Non-Rigid Motion,” International Conference on Computer Vision, Corfu, Greece, September
`
`1999.
`
`
`
`
`[11]
`
`{6} F. Bartolini, V. Cappellini, and A. Mecocci, “Counting people getting in and
`
`
`out of a bus by real-time image-sequenceprocessing,” IVC, 12(1):36-41, January 1994.
`
`[12]
`
`{7} M.Rossi and A. Bozzoli, “Tracking and counting moving people,” ICIP94,
`
`pp. 212-216, 1994.
`
`[13]
`
`{8} C.R. Wren, A. Azarbayejani, T. Darrell, and A. Pentland, “Pfinder: Real-
`
`time tracking of the human body,” Vismod, 1995.
`
`[14]
`
`{9} L. Khoudour, L. Duvieubourg,J.P. Deparis, “Real-Time Pedestrian Counting
`
`by Active Linear Cameras,” JEI, 5(4):452-459, October 1996.
`
`[15]
`
`{10} S. Ioffe, D.A. Forsyth, “Probabilistic Methods for Finding People,” IJCV,
`
`43(1):45-68, June 2001.
`
`[16]
`
`{11} M. Isard and J. MacCormick, “BraMBLe: A Bayesian Multiple-Blob
`
`Tracker,” ICCV,2001.
`
`[17]
`
`The following references describe blob analysis:
`
`20
`
`[18]
`
`{12} D.M.Gavrila, “The Visual Analysis of Human Movement: A Survey,”
`
`CVIU, 73(1):82-98, January 1999.
`
`[19]
`
`{13} Niels Haering and Niels da Vitoria Lobo, “Visual Event Detection,” Video
`
`Computing Series, Editor Mubarak Shah, 2001.
`
`-2-
`
`(37112-175340)
`
`Canon Ex. 1040 Page3 of 31
`
`Canon Ex. 1040 Page 3 of 31
`
`

`

`[20]
`
`[21]
`
`The following references describe blob analysis for trucks, cars, and people:
`
`{14} Collins, Lipton, Kanade, Fujiyoshi, Duggins, Tsin, Tolliver, Enomoto, and
`
`Hasegawa, "A System for Video Surveillance and Monitoring: VSAM Final Report," Technical
`
`Report CMU-RI-TR-00-12, Robotics Institute, Carnegie Mellon University, May 2000.
`
`[22]
`
`{15} Lipton, Fujiyoshi, and Patil, “Moving Target Classification and Tracking
`
`from Real-time Video,” 98 Darpa IUW,Nov. 20-23, 1998.
`
`[23]
`
`[24]
`
`The following reference describes analyzing a single-person blobandits contours:
`
`{16} C.R. Wren, A. Azarbayejani, T. Darrell, and A.P. Pentland. “Pfinder: Real-
`
`Time Tracking of the Human Body,” PAMI,vol 19, pp. 780-784, 1997.
`
`[25]
`
`The following reference describes internal motion ofblobs, including any motion-
`
`based segmentation:
`
`[26]
`
`{17} M. Allmen and C.Dyer, “Long--Range Spatiotemporal Motion
`
`Understanding Using Spatiotemporal Flow Curves,” Proc. IEEE CVPR, Lahaina, Maui, Hawaii,
`
`pp. 303-309, 1991.
`
`[27]
`
`{18} L. Wixson,"Detecting Salient Motion by Accumulating Directionally
`
`Consistent Flow", IEEE Trans. Pattern Anal. Mach.Intell., vol. 22, pp. 774-781, Aug, 2000.
`
`
`
`
`Backgroundof the Invention
`
`[28] Video surveillance of public spaces has become extremely widespread and
`
`20
`
`accepted by the general public. Unfortunately, conventional video surveillance systems produce
`
`such prodigious volumesofdata that an intractable problem results in the analysis of video
`
`surveillance data.
`
`-3-
`
`(37112-175340)
`
`Canon Ex. 1040 Page 4 of 31
`
`Canon Ex. 1040 Page 4 of 31
`
`

`

`[29]
`
`A need exists to reduce the amountof video surveillance data so analysis of the
`
`video surveillance data can be conducted.
`
`[30] A need exists to filter video surveillance data to identify desired portions of the
`
`video surveillance data.
`
`
`
`SUMMARYOF THE INVENTION
`
`[31] An object of the invention is to reduce the amountof video surveillance data so
`
`analysis of the video surveillance data can be conducted.
`
`[32] An object of the inventionisto filter video surveillance data to identify desired
`
`portions of the video surveillance data.
`
`[33] An object of the invention is to produce a real time alarm based on an automatic
`
`detection of an event from video surveillance data.
`
`[34] An object of the invention is to integrate data from surveillance sensors other than
`
`video for improved searching capabilities.
`
`[35] An object of the inventionis to integrate data from surveillance sensors other than
`
`video for improved event detection capabilities
`
`[36]
`
`The invention includesan article of manufacture, a method, a system, and an
`
`apparatus for video surveillance.
`
`[37]
`
`The article of manufacture of the invention includes a computer-readable medium
`
`20
`
`comprising software for a video surveillance system, comprising code segments for operating the
`
`video surveillance system based on video primitives.
`
`[38]
`
`The article ofmanufacture of the invention includes a computer-readable medium
`
`comprising software for a video surveillance system, comprising code segments for accessing
`
`-4.
`
`(37112-175340)
`
`Canon Ex. 1040 Page 5 of 31
`
`Canon Ex. 1040 Page 5 of 31
`
`

`

`archived video primitives, and code segments for extracting event occurrences from accessed
`
`archived video primitives.
`
`[39]
`
`The system of the invention includes a computer system including a computer-
`
`readable medium having software to operate a computer in accordance with the invention.
`
`[40]
`
`The apparatus of the invention includes a computer including a computer-readable
`
`medium having software to operate the computer in accordance with the invention.
`
`[41]
`
`The article of manufacture of the invention includes a computer-readable medium
`
`having software to operate a computer in accordance with the invention.
`
`[42] Moreover, the above objects and advantages ofthe inventionare illustrative, and
`
`not exhaustive, ofthose that can be achieved by the invention. Thus, these and other objects and
`
`advantages of the invention will be apparent from the description herein, both as embodied
`
`herein and as modified in view of any variations which will be apparent to those skilled in the
`
`art.
`
`Definitions
`
`
`
`
`
`
`[43] A “video”refers to motion pictures represented in analog and/or digital form.
`
`Examplesof video include: television, movies, image sequences from a video cameraor other
`
`observer, and computer-generated image sequences.
`
`[44] A “frame”refers to a particular image or other discrete unit within a video.
`
`20
`
`[45] An “object” refers to an item of interest in a video. Examples of an object
`
`include: a person, a vehicle, an animal, and a physical subject.
`
`-5-
`
`(37112-175340)
`
`Canon Ex. 1040 Page 6 of 31
`
`Canon Ex. 1040 Page 6 of 31
`
`

`

`[46] An “activity” refers to one or more actions and/or one or more composites of
`
`actions of one or more objects. Examples of an activity include: entering; exiting; stopping;
`
`moving; raising; lowering; growing; and shrinking.
`
`[47] A “location”refers to a space where an activity may occur. A location can be, for
`
`example, scene-based or image-based. Examples of a scene-based location include: a public
`
`space; a store; a retail space; an office; a warehouse; a hotel room; a hotel lobby; a lobby of a
`
`building; a casino; a busstation; a train station; an airport; a port; a bus; a train; an airplane; and
`
`a ship. Examples of an image-based location include: a video image;a line in a video image; an
`
`area in a video image; a rectangular section of a video image; and a polygonal section of a video
`
`image.
`
`[48] An “event” refers to one or more objects engaged in an activity. The event may
`
`be referenced with respect to a location and/or a time.
`
`[49] A “computer” refers to any apparatus that is capable of accepting a structured
`
`input, processing the structured input accordingto prescribed rules, and producingresults of the
`
`processing as output. Examples of a computer include: a computer; a general purpose computer;
`
`a supercomputer; a mainframe; a super mini-computer; a mini-computer; a workstation; a micro-
`
`computer; a server; an interactive television; a hybrid combination of a computer and an
`
`interactive television; and application-specific hardware to emulate a computer and/or software.
`
`A computer can have a single processor or multiple processors, which can operatein parallel
`
`
`
`
`
`20
`
`and/or notin parallel. A computer also refers to two or more computers connected together via a
`
`network for transmitting or receiving information between the computers. An example of such a
`
`computer includesa distributed computer system for processing information via computers
`
`linked by a network.
`
`-6-
`
`(37112-175340)
`
`Canon Ex. 1040 Page 7 of 31
`
`Canon Ex. 1040 Page 7 of 31
`
`

`

`[50] A “computer-readable medium”refers to any storage device used for storing data
`
`accessible by a computer. Examples of a computer-readable medium include: a magnetic hard
`
`disk; a floppy disk; an optical disk, such as a CD-ROM and a DVD;a magnetic tape; a memory
`
`chip; and a carrier wave used to carry computer-readable electronic data, such as those used in
`
`transmitting and receiving e-mail or in accessing a network.
`
`[51]
`
`“Software”refers to prescribed rules to operate a computer. Examples of
`
`software include: software; code segments; instructions; computer programs; and programmed
`
`logic.
`
`[52] A “computer system” refers to a system having a computer, where the computer
`
`comprises a computer-readable medium embodying software to operate the computer.
`
`[53] A “network”refers to a number of computers and associated devices that are
`
`connected by communication facilities. A network involves permanent connections such as
`
`cables or temporary connections such as those made through telephone or other communication
`
`links. Examples of a network include: an internet, such as the Internet; an intranet; a local area
`
`network (LAN); a wide area network (WAN); and a combination of networks, such as an internet
`
`and an intranet.
`
`
`
`BRIEF DESCRIPTION OF THE DRAWINGS
`
`[54]
`
`Embodiments ofthe invention are explained in greater detail by way ofthe
`
`20
`
`drawings, where the same reference numerals refer to the same features.
`
`[55]
`
`Figure 1 illustrates a plan view ofthe video surveillance system of the invention.
`
`[56]
`
`Figure 2 illustrates a flow diagram forthe video surveillance system of the
`
`invention.
`
`-7-
`
`(37112-175340)
`
`Canon Ex. 1040 Page 8 of 31
`
`Canon Ex. 1040 Page 8 of 31
`
`

`

`[57]
`
`Figure 3 illustrates a flow diagram for tasking the video surveillance system.
`
`[58]
`
`Figure 4 illustrates a flow diagram for operating the video surveillance system.
`
`[59]
`
`Figure 5 illustrates a flow diagram for extracting video primitives for the video
`
`surveillance system.
`
`[60]
`
`Figure6 illustrates a flow diagram for taking action with the video surveillance
`
`system.
`
`
`
`[61] Figure7illustrates a flow diagram for semi-automatic calibration of the video
`
`surveillance system.
`
`[62]
`
`Figure8 illustrates a flow diagram for automatic calibration of the video
`
`surveillance system.
`
`[63]
`
`Figure 9 illustrates an additional flow diagram for the video surveillance system
`
`of the invention.
`
`[64]
`
`Figures 10-15 illustrate examples of the video surveillance system of the
`
`invention applied to monitoring a grocery store.
`
`
`
`
`DETAILED DESCRIPTION OF THE INVENTION
`
`[65]
`
`The automatic video surveillance system of the invention is for monitoring a
`
`location for, for example, market research or security purposes. The system can be a dedicated
`
`video surveillance installation with purpose-built surveillance components,or the system can be
`
`20
`
`a retrofit to existing video surveillance equipmentthat piggybacksoff the surveillance video
`
`feeds. The system is capable of analyzing video data from live sources or from recorded media.
`
`The system can have a prescribed responseto the analysis, such as record data, activate an alarm
`
`mechanism, or active another sensor system. The system is also capable of integrating with
`
`-8-
`
`(37112-175340)
`
`Canon Ex. 1040 Page 9 of 31
`
`Canon Ex. 1040 Page 9 of 31
`
`

`

`other surveillance system components. The system produces security or market research reports
`
`that can be tailored according to the needs of an operator and, as an option, can be presented
`
`through an interactive web-based interface, or other reporting mechanism.
`
`[66] An operator is provided with maximum flexibility in configuring the system by
`
`using event discriminators. Event discriminators are identified with one or more objects (whose
`
`descriptions are based on video primitives), along with one or more optionalspatial attributes,
`
`and/or one or more optional temporalattributes. For example, an operator can define an event
`
`discriminator(called a "loitering" event in this example) as a “person” object in the “automatic
`
`teller machine” space for “longer than 15 minutes” and “between 10:00 p.m. and 6:00 a.m.”
`
`[67] Although the video surveillance system of the invention draws on well-known
`
`computer vision techniques from the public domain, the inventive video surveillance system has
`
`several unique and novel features that are not currently available. For example, current video
`
`surveillance systems use large volumesof video imagery as the primary commodity of
`
`information interchange. The system of the invention uses video primitives as the primary
`
`commodity with representative video imagery being used as collateral evidence. The system of
`
`the invention can also be calibrated (manually, semi-automatically, or automatically) and
`
`thereafter automatically can infer video primitives from video imagery. The system can further
`
`analyze previously processed video without needing to reprocess completely the video. By
`
`analyzing previously processed video, the system can perform inference analysis based on
`
`
`
`
`
`
`20
`
`previously recorded video primitives, which greatly improvesthe analysis speed of the computer
`
`system.
`
`[68] As another example, the system of the invention provides unique system tasking.
`
`Using equipment control directives, current video systemsallowauserto position video sensors
`
`-9-
`
`(37112-175340)
`
`Canon Ex. 1040 Page 10 of 31
`
`Canon Ex. 1040 Page 10 of 31
`
`

`

`and, in some sophisticated conventional systems, to mask out regionsofinterest or disinterest.
`
`Equipmentcontrol directives are instructions to control the position, orientation, and focus of
`
`video cameras. Instead of equipment control directives, the system of the invention uses event
`
`discriminators based on video primitives as the primary tasking mechanism. With event
`
`discriminators and video primitives, an operator is provided with a much moreintuitive approach
`
`over conventional systems for extracting useful information from the system. Rather than
`
`tasking a system with an equipment control directives, such as “camera A pan 45 degrees to the
`
`left,” the system of the invention can be tasked in a human-intuitive manner with one or more
`
`event discriminators based on video primitives, such as “a person enters restricted area A.”
`
`[69] Using the invention for market research, the following are examples of the type of
`
`video surveillance that can be performed with the invention: counting people in a store; counting
`
`people in a part of a store; counting people whostop in a particular place in a store; measuring
`
`how long people spendin a store; measuring how long people spendin a part of a store; and
`
`[70] Using the invention for security, the following are examplesofthe type of video
`
`surveillance that can be performed with the invention: determining when anyoneenters a
`
`restricted area and storing associated imagery; determining when a person enters an area at
`
`unusual times; determining when changesto shelf space and storage space occur that might be
`
`unauthorized; determining when passengers aboard an aircraft approach the cockpit; determining
`
`
`
`measuring the length of a line inastore.
`
`20
`
`when peopletailgate through a secure portal; determining if there is an unattended bag in an
`
`airport; and determiningif there is a theft of an asset.
`
`[71]
`
`Figure 1 illustrates a plan view ofthe video surveillance system of the invention.
`
`A computer system 11 comprises a computer 12 having a computer-readable medium 13
`
`-10-
`
`(37112-175340)
`
`Canon Ex. 1040 Page 11 of 31
`
`Canon Ex. 1040 Page 11 of 31
`
`

`

`embodying software to operate the computer 12 according to the invention. The computer
`
`system 11 is coupled to one or more video sensors 14, one or more video recorders 15, and one
`
`or more input/output(I/O) devices 16. The video sensors 14 can also be optionally coupled to
`the video recorders 15 for direct recording ofvideo surveillance data. The computer system is
`
`optionally coupled to other sensors 17.
`
`[72]
`
`The video sensors 14 provide source video to the computer system 11. Each
`
`video sensor 14 can be coupled to the computer system 11 using, for example, a direct
`
`connection(e.g., a firewire digital camera interface) or a network. The video sensors 14 can
`
`exist prior to installation of the invention or can beinstalled as part of the invention. Examples
`
`of a video sensor 14 include: a video camera;a digital video camera; a color camera; a
`
`monochrome camera; a camera; a camcorder, a PC camera; a webcam; an infra-red video
`
`camera; and a CCTV camera.
`
`[73]
`
`The video recorders 15 receive video surveillance data from the computer system
`
`11 for recording and/or provide source video to the computer system 11. Each video recorder 15
`
`can be coupled to the computer system 11 using, for example, a direct connection or a network.
`
`The video recorders 15 can exist prior to installation of the invention or can be installed as part
`
`of the invention. Examplesof a video recorder 15 include: a video tape recorder; a digital video
`
`recorder; a video disk; a DVD; and a computer-readable medium.
`
`
`
`
`[74]
`
`The I/O devices 16 provide input to and receive output from the computer system
`
`20
`
`11. The I/O devices 16 can be used to task the computer system 11 and produce reports from the
`
`computer system 11. Examples of I/O devices 16 include: a keyboard; a mouse; a stylus; a
`
`monitor; a printer; another computer system; a network; and an alarm.
`
`-ll-
`
`(37112-175340)
`
`Canon Ex. 1040 Page 12 of 31
`
`Canon Ex. 1040 Page 12 of 31
`
`

`

`[75]
`
`The other sensors 17 provide additional input to the computer system 11. Each
`
`other sensor 17 can be coupled to the computer system 11 using, for example, a direct connection
`
`or anetwork. The other sensors 17 can exit prior to installation of the invention or can be
`
`installed as part of the invention. Examples of another sensor 17 include: a motion sensor; an
`
`optical tripwire; a biometric sensor; and a card-based or keypad-based authorization system. The
`
`outputs of the other sensors 17 can be recorded by the computer system 11, recording devices,
`
`and/or recording systems.
`
`
`
`[76] Figure2illustrates a flow diagram for the video surveillance system of the
`
`
`
`
`invention. Various aspects of the invention are exemplified with reference to Figures 10-15,
`
`whichillustrate examples of the video surveillance system of the invention applied to monitoring
`
`a grocery store.
`
`[77]
`
`In block 21, the video surveillance system is set up as discussed for Figure1.
`
`Each video sensor 14 is orientated to a location for video surveillance. The computer system 11
`
`is connected to the video feeds from the video equipment 14 and 15. The video surveillance
`
`system can be implemented using existing equipment or newly installed equipment for the
`
`location.
`
`[78]
`
`In block 22, the video surveillance system is calibrated. Once the video
`
`surveillance system is in place from block 21, calibration occurs. The result of block 22 is the
`
`ability of the video surveillance system to determine an approximate absolute size and speed of a
`
`20
`
`particular object (e.g., a person) at various places in the video image provided by the video
`
`sensor. The system can be calibrated using manualcalibration, semi-automatic calibration, and
`
`automatic calibration. Calibration is further described after the discussion of block 24.
`
`-12-
`
`(37112-175340)
`
`Canon Ex. 1040 Page 13 of 31
`
`Canon Ex. 1040 Page 13 of 31
`
`

`

`[79]
`
`In block 23 of Figure 2, the video surveillance system is tasked. Tasking occurs
`
`after calibration in block 22 and is optional. Tasking the video surveillance system involves
`
`specifying one or more eventdiscriminators. Without tasking, the video surveillance system
`
`operates by detecting and archiving video primitives and associated video imagery without
`
`taking any action, as in block 45 in Figure 4.
`
`
`
`
`[80]
`
`Figure 3 illustrates a flow diagram for tasking the video surveillance system to
`
`determine event discriminators. An event discriminatorrefers to one or more objects optionally
`
`interacting with one or more spatial attributes and/or one or more temporal attributes. An event
`
`discriminator is described in terms of video primitives. A video primitive refers to an observable
`
`attribute of an object viewed in a video feed. Examples of video primitives include the
`
`following: a classification; a size; a shape; a color; a texture; a position; a velocity; a speed; an
`
`internal motion; a motion; a salient motion; a feature of a salient motion; a scene change; a
`
`feature of a scene change; and a pre-defined model.
`
`[81]
`
` Acclassification refers to an identification of an object as belongingto a particular
`
`category or class. Examples ofa classification include: a person;a dog; a vehicle; a police car;
`
`an individual person; and a specific type of object.
`
`[82] A size refers to a dimensionalattribute of an object. Examples of a size include:
`
`large; medium; small; flat; taller than 6 feet; shorter than 1 foot; wider than 3 feet; thinner than 4
`
`feet; about human size; bigger than a human; smaller than a human; about the size of a car; a
`
`20
`
`rectangle in an image with approximate dimensionsin pixels; and a number of image pixels.
`
`[83]
`
` Acolor refers to a chromatic attribute of an object. Examples of a color include:
`
`white; black; grey; red; a range of HSV values; a range of YUV values; a range of RGB values;
`
`an average RGB value; an average YUV value; and a histogram of RGB values.
`
`-13-
`
`(37112-175340)
`
`Canon Ex. 1040 Page 14 of 31
`
`Canon Ex. 1040 Page 14 of 31
`
`

`

`[84]
`
`A texture refers to a pattern attribute of an object. Examples of texture features
`
`include:self-similarity; spectral power; linearity; and coarseness.
`
`[85] An internal motion refers to a measure ofthe rigidity of an object. An example of
`
`a fairly rigid object is a car, which does not exhibit a great amount of internal motion. An
`
`example of a fairly non-rigid object is a person having swinging arms and legs, which exhibits a
`
`great amountof internal motion.
`
`
`
`15
`
`[86] A motion refers to any motion that can be automatically detected. Examples of a
`
`motion include: appearance of an object; disappearance of an object; a vertical movement of an
`
`object; a horizontal movementof an object; and a periodic movementof an object.
`
`[87] A salient motion refers to any motion that can be automatically detected and can
`
`be tracked for some period of time. Such a moving object exhibits apparently purposeful
`
`motion. Examples ofa salient motion include: moving from oneplace to another; and moving to
`
`interact with another object.
`
`[88] A feature of a salient motion refers to a property of a salient motion. Examples of
`
`a feature of a salient motion include: a trajectory; a length of a trajectory in image space; an
`
`approximate length of a trajectory in a three-dimensional representation of the environment; a
`
`position of an object in image space as a function of time; an approximateposition of an object
`
`in a three-dimensional representation of the environmentas a function of time; a duration of a
`
`trajectory; a velocity (e.g., speed and direction) in image space; an approximatevelocity(e.g.,
`
`20
`
`speed and direction) in a three-dimensional representation of the environment; a duration of time
`
`at a velocity; a change of velocity in image space; an approximate changeof velocity in a three-
`
`dimensional representation of the environment; a duration of a change of velocity; cessation of
`
`motion; and a duration of cessation of motion. A velocity refers to the speed and direction of an
`
`-14-
`
`(37112-175340)
`
`Canon Ex. 1040 Page 15 of 31
`
`Canon Ex. 1040 Page 15 of 31
`
`

`

`object at a particular time. A trajectory refers a set of (position, velocity) pairs for an object for
`
`as long as the object can be tracked or for a time period.
`
`[89] A-scene change refers to any region of a scene that can be detected as changing
`
`over a period of time. Examples of a scene changeinclude:an stationary object leaving a scene;
`
`an object entering a scene and becomingstationary; an object changing position in a scene; and
`
`an object changing appearance(e.g. color, shape, or size).
`
`[90]
`
`A feature of a scene changerefers to a property of a scene change. Examples of a
`
`feature of a scene changeinclude: a size of a scene change in image space; an approximate size
`ofa scene change in a three-dimensional representation ofthe environment: a time at which a
`
`scene change occurred; a location of a scene change in image space; and an approximate location
`
`of a scene change in a three-dimensional representation of the environment.
`
`[91] A pre-defined model refers to an a priori known model of an object. Examples of
`
`a pre-defined include: an adult; a child; a vehicle; and a semi-trailer.
`
`[92]
`
`In block 31, one or more objects types of interests are identified in terms of video
`
`primitives or abstractions thereof. Examples of one or more objects include: an object; a person;
`
`
`
`
`15
`
`a red object; two objects; two persons; and a vehicle.
`
`[93]
`
`In block 32, one or morespatial areas of interest are identified. An area refers to
`
`one or more portions of an image from a source videoor a spatial portion of a scene being
`
`viewed by a video sensor. An area also includes a combination of areas from various scenes
`
`20
`
`and/or images. An area can be an image-based space(e.g., a line, a rectangle, a polygon, or a
`
`circle in a video image) or a three-dimensionalspace(e.g., a cube, or an area of floor space ina
`
`building).
`
`-15-
`
`(37112-175340)
`
`Canon Ex. 1040 Page 16 of 31
`
`Canon Ex. 1040 Page 16 of 31
`
`

`

`[94]
`
`Figure 12 illustrates identifying areas along an aisle in a grocery store. Four areas
`
`are identified: coffee; soda promotion; chips snacks; and bottled water. The areas are identified
`
`via a point-and-click interface with the system.
`
`[95]
`
`In block 33, one or more temporal attributes of interest are optionally identified.
`
`Examples of a temporal attribute include: every 15 minutes; between 9:00 p.m. to 6:30 a.m.; less
`
`than 5 minutes; longer than 30 seconds; over the weekend; and within 20 minutesof.
`
`[96]
`
`In block 34,a responseis optionally identified. Examples of a response includes
`
`the following: activating a visual and/or audio alert on a system display; activating a visual
`
`and/or audio alarm system at the location; activating a silent alarm; activating a rapid response
`
`mechanism;locking a door; contacting a security service; forwarding data (e.g., image data,
`
`video data, video primitives; and/or analyzed data) to another computer system via a network,
`
`such as the Internet; saving such data to a designated computer-readable medium;activating
`
`some other sensoror surveillance system; tasking the computer system 11 and/or another
`
`computer system; and directing the computer system 11 and/or another computer system.
`
`[97]
`
`In block 35, one or more discriminators are identified by describing interactions
`
`between video primitives (or their abstractions), spatial areas of interest, and temporal attributes
`
`of interest. An interaction is determined for a combination of one or more objects identified in
`
`block 31, one or morespatial areas of interest identified in block 32, and one or more temporal
`
`attributes of interest identified in block 33. One or more responsesidentified in block 34 are
`
`
`
`20
`
`optionally associated with each event discriminator.
`
`[98]
`
`Examples of an event discriminator for a single object include: an object appears;
`
`a person appears; and a red object movesfaster than 10m/s.
`
`-16-
`
`(37112-175340)
`
`Canon Ex. 1040 Page 17 of 31
`
`Canon Ex. 1040 Page 17 of 31
`
`

`

`[99]
`
`Examplesof an event discriminator for multiple objects include: two objects
`
`cometogether; a person exits a vehicle; and a red object moves next to a blue object.
`
`[100] Examples of an event discriminator for an object and a spatial attribute include:
`
`an object crosses a line; an object enters an area; and a person crossesa line from theleft.
`
`[101] Examples of an event discriminator for an object and a temporalattribute include:
`
`an object appears at 10:00 p.m.; a person travels faster then 2m/s between 9:00 a.m. and 5:00
`
`p.m.; and a vehicle appears on the weekend.
`
`[102] Examples of an event discriminator for an object, a spatial attribute, and a
`
`temporalattribute include: a personcrossesa line between midnight and 6:00 a.m.; and a vehicle
`
`stops in an area for longer than 10 minutes.
`
`[103] An example of an event discriminator for an object, a spatial attribute, and a
`
`temporalattribute associated with a response include: a person enters an area between midnight
`
`and 6:00 a.m., and a security service is notified.
`
`[104]
`
`In block 24 of Figure 2, the video surveillance system is operated. The video
`
`surveillance system of the invention operates automatically, detects and archives video
`
`primitives of objects in the scene, and detects event occurrencesin real time using event
`
`discriminators. In addition, action is taken in real time, as appropriate, such as activating alarms,
`
`generating reports, and generating output. The reports and output can be displayed and/or stored
`
`locally to the system or elsewhere via a network, such as the Internet. Figure 4 illustrates a flow
`
`
`
`
`20
`
`diagram for operating the video surveillance system.
`
`[105]
`
`In block 41, the computer system 11 obtains source video from the video sensors
`
`14 and/or the video recorders 15.
`
`-17-
`
`(37112-175340)
`
`Canon Ex. 1040 Page 18 of 31
`
`Canon Ex. 1040 Page 18 of 31
`
`

`

`[106]
`
`In block 42, video primitives are extracted in real time from the source video. As
`
`an option, non-video primitives can be obtained and/or extracted from one or more other sensors
`
`17 and used with the invention. The extraction of video primitivesis illustrated with Figure5.
`
`[107] Figure 5 illustrates a flow diagram for extracting video primitives for the video
`
`5—surveillance system. Blocks 51 and 52 operate in parallel and can be performed in any orderor
`
`concurrently. In block 51, objects are detected via movement. Any motion detection algorithm
`
`for detecting movement between framesat the pixel level can be used for this block. As an
`
`example, the three frame differencing technique can be used, whichis discussed in {1}. The
`
`detected objects are forwardedto block 53.
`
` subtraction, can be used, which is described in {1} and U.S. Patent Application No. 09/694,712
`
`[108]
`
`In block 52, objects are detected via change. Any change detection algorithm for
`
`detecting changes from

This document is available on Docket Alarm but you must sign up to view it.


Or .

Accessing this document will incur an additional charge of $.

After purchase, you can access this document again without charge.

Accept $ Charge
throbber

Still Working On It

This document is taking longer than usual to download. This can happen if we need to contact the court directly to obtain the document and their servers are running slowly.

Give it another minute or two to complete, and then try the refresh button.

throbber

A few More Minutes ... Still Working

It can take up to 5 minutes for us to download a document if the court servers are running slowly.

Thank you for your continued patience.

This document could not be displayed.

We could not find this document within its docket. Please go back to the docket page and check the link. If that does not work, go back to the docket and refresh it to pull the newest information.

Your account does not support viewing this document.

You need a Paid Account to view this document. Click here to change your account type.

Your account does not support viewing this document.

Set your membership status to view this document.

With a Docket Alarm membership, you'll get a whole lot more, including:

  • Up-to-date information for this case.
  • Email alerts whenever there is an update.
  • Full text search for other cases.
  • Get email alerts whenever a new case matches your search.

Become a Member

One Moment Please

The filing “” is large (MB) and is being downloaded.

Please refresh this page in a few minutes to see if the filing has been downloaded. The filing will also be emailed to you when the download completes.

Your document is on its way!

If you do not receive the document in five minutes, contact support at support@docketalarm.com.

Sealed Document

We are unable to display this document, it may be under a court ordered seal.

If you have proper credentials to access the file, you may proceed directly to the court's system using your government issued username and password.


Access Government Site

We are redirecting you
to a mobile optimized page.





Document Unreadable or Corrupt

Refresh this Document
Go to the Docket

We are unable to display this document.

Refresh this Document
Go to the Docket