`(12) Patent Application Publication (10) Pub. No.: US 2006/0092043 A1
`Lagassey
`(43) Pub. Date:
`May 4, 2006
`
`US 20060092043A1
`
`(54) ADVANCED AUTOMOBILE ACCIDENT
`DETECTION, DATA RECORDATION AND
`REPORTING SYSTEM
`
`(76) Inventor: Paul J. Lagassey, Vero Beach, FL (US)
`Correspondence Address:
`MILDE & HOFFBERG, LLP
`1O BANK STREET
`SUTE 460
`WHITE PLAINS, NY 10606 (US)
`(21) Appl. No.:
`11/267,732
`
`(22) Filed:
`
`Nov. 3, 2005
`Related U.S. Application Data
`(60) Provisional application No. 60/522,749, filed on Nov.
`3, 2004.
`
`Publication Classification
`
`(51) Int. Cl.
`G08G I/095
`
`(2006.01)
`
`(52) U.S. Cl. .............................................................. 340/907
`
`(57)
`
`ABSTRACT
`
`A system for monitoring a location to detect and report a
`vehicular incident, comprising a transducer for detecting
`acoustic waves at the location, and having an audio output;
`a processor for determining a probable occurrence or
`impending occurrence of a vehicular incident, based at least
`upon said audio output; an imaging system for capturing
`images of the location, and having an image output; a buffer,
`receiving said image output, and storing at least a portion of
`said images commencing at or before said determination;
`and a communication link, for selectively communicating
`said portion of said images stored in said buffer with a
`remote location and at least information identifying the
`location, wherein information stored in said buffer is pre
`served at least until an acknowledgement of receipt is
`received representing Successful transmission through said
`communication link with the remote location.
`
`100 N
`
`40
`
`
`
`45
`
`
`
`SEE : B
`
`
`
`Exhibit 1011
`Page 01 of 29
`
`
`
`Patent Application Publication May 4, 2006 Sheet 1 of 6
`
`US 2006/0092043 A1
`
`
`
`S.
`
`f
`s
`
`Exhibit 1011
`Page 02 of 29
`
`
`
`Patent Application Publication May 4, 2006 Sheet 2 of 6
`
`US 2006/0092043 A1
`
`(N
`S. N
`
`
`
`O
`O
`ve
`
`Exhibit 1011
`Page 03 of 29
`
`
`
`Patent Application Publication May 4, 2006 Sheet 3 of 6
`
`US 2006/0092043 A1
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`Control Unit Active and Receiving incoming Audio and Video Signals 51
`
`Store. At Least Video Data and Other Desired Accident Related Data in a
`Circular Buffer White 52
`
`NO
`
`
`
`incoming Audio Signals for Match to
`Stored Acoustic Signature of Qualifying Sound to Determine if
`Qualifying Sound is Present?
`
`Yes
`Stop Overwriting and Preserve Data in Circular Buffer 54
`
`Continue to Save at Least Subsequent Video Data and Other Desired
`ACCident-Related Data 55
`
`Initiate Contact With Monitoring Center 75
`
`Contact Established 76
`
`
`
`No
`
`Yes
`Transmit Location Data and At Least One image to Monitoring Center 77
`
`Continue Saving Desired Accident Related Data 78
`Yes
`
`ime Limit Passed,
`emory Limit Been Reached, or Termination Signal Received
`79
`
`Yes
`Stop Saving Accident-Related Data 80
`
`
`
`Transmit Accident-Related Data 81
`
`Uccessful Transmission
`Verified 82
`Yes
`End Transmission 85
`
`Flush and Reuse allocated memory 90
`
`Fig. 3
`
`
`
`
`
`O
`
`No
`
`No
`
`Exhibit 1011
`Page 04 of 29
`
`
`
`Patent Application Publication May 4, 2006 Sheet 4 of 6
`
`US 2006/0092043 A1
`
`START
`
`NO
`
`Receive Audio And Video Signals, Time And
`LOCation Data 50
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`NO
`
`
`
`Audio Signals For Match
`Stored Acoustic Signature
`
`
`
`
`
`
`
`YES
`
`Continue Saving And Analyze
`Subsequent Audio Signals For
`Match To Qualifying Sound 61
`
`Sound Detected Within Fif
`Predetermined Time?
`
`NO
`
`Transmit Location Data And At
`Least One Image To Monitoring
`Center 77
`
`econd Predetermined
`Time Passed, Storage
`Capacity Been Reached, Or
`Termination Signal
`
`Stop Saving Accident-Related Data
`80
`
`Transmit Or Upload Accident
`Related Data 81
`
`Successful
`Transmission Or Upload
`Verified 82
`
`YES
`Flush Buffer 90
`
`Reinitialize (Optional) 99
`Fig. 4
`
`Exhibit 1011
`Page 05 of 29
`
`
`
`Patent Application Publication May 4, 2006 Sheet 5 of 6
`
`US 2006/0092043 A1
`
`
`
`27 A/S
`
`A /
`
`s
`
`Exhibit 1011
`Page 06 of 29
`
`
`
`Patent Application Publication May 4, 2006 Sheet 6 of 6
`
`US 2006/0092043 A1
`
`Detect acoustic waves at the location 301
`
`Analyze conditions at location 302
`Determine a likely occurrence or an imminent occurrence of a vehicular accident or
`Other incident 303
`y
`Optionally determine compliance with traffic control regulations 304
`Capture initial images of the location along with audio, timecode, state of traffic signal,
`GPS code, optionally polling a plurality of cameras 305
`- Y - -
`Store data, starting no later than the determination of likely occurrence 306
`v
`Optionally Communicate Location and At Least One image 307
`Continue to capture stream of images of the location along with audio, timecode, state
`of traffic signal, GPS code until cease condition 308
`Optionally use sensor data to model location 309
`Optionally communicate to or from traffic signal control device 310
`
`-
`
`Establish communication pathway and communicate the stored
`images and incident related data to a remote location 311
`
`
`
`
`
`
`
`erification of Successfu
`Communication?
`312
`
`
`
`
`
`Retry and/or try
`alternate Communication
`pathway 313
`
`Yes
`Preserve stored information at least until verification of successful Communication, and
`then delete 314
`
`Receive and display information from a plurality of locations at a remote monitoring
`center on a map 315
`
`Route information to an available live agent at a remote monitoring Center,
`coordinate multiple communications 316
`Preserve information in a forensically reliable record 317
`
`
`
`Communicate from remote monitoring center to location with audio Communications,
`to control and program traffic signal control device, control and program the
`components of the system, and to activate visual alert 318
`Fig. 6
`
`Exhibit 1011
`Page 07 of 29
`
`
`
`US 2006/0092043 A1
`
`May 4, 2006
`
`ADVANCED AUTOMOBILE ACCIDENT
`DETECTION, DATA RECORDATION AND
`REPORTING SYSTEM
`
`CROSS REFERENCE TO RELATED
`APPLICATION
`0001. The present application claims benefit of priority
`from U.S. Provisional Patent Application 60/522,749 filed
`Nov. 3, 2004.
`
`BACKGROUND OF THE INVENTION
`0002 The invention generally relates to an automobile
`accident detection and data recordation and reporting sys
`tem, and in particular to a system which detects accidents
`based on a set of characteristic sounds or other cues.
`0003 Traffic accidents cause significant costs in terms of
`direct loss, consequential loss, and Societal loss due to
`obstruction of the roadway in the aftermath of an accident.
`Another issue is the allocation of direct costs, for example
`when more than one vehicle is involved, the vehicle at fault
`is generally held liable for the damages.
`0004. It is possible to monitor locations that are likely
`places for accidents to occur, however, without intelligence,
`this process may be inefficient and unproductive. Likewise,
`without immediate and efficient communication of the infor
`mation obtained, benefits of the monitoring are quite limited.
`0005 Since cellular telephone technology has become so
`widely adopted, the most common means by which motor
`vehicle accidents are reported to agencies in the U.S. is
`through cellular telephones. However, this is not always
`reliable or immediate if the victims are unable to use their
`cellular phones or if there are no witnesses with cellular
`phones to report the accident, and it fails to record an actual
`record of the accident which can later be used as evidence.
`0006 Automobile accident detection systems are com
`mon in the art. Upon the occurrence of an automobile
`accident, it may be desirable to obtain video images and
`sounds of the accident and to record the time of the accident
`and the status of the traffic lights at the time the accident
`occurred. This information can then be sent to a remote
`location where emergency crews can be dispatched and the
`information further examined and forwarded to authorities
`in order to determine fault and liability.
`0007. A number of prior art techniques are available for
`predicting the occurrence of an accident. Some of these
`require an extended period of time for an automated system
`to analyze the data, and thus any report generated is Sub
`stantially delayed. In others, the accuracy of the system
`depends on environmental conditions, such as lighting or
`time of day. Therefore, in order to provide an immediate and
`reliable response to a predicted occurrence of an accident,
`Such techniques are Suboptimal.
`0008 For example, Japanese Patent Application No.
`8-162911 entitled “Motor Vehicle Accident Monitoring
`Device” (“the Japanese reference”), expressly incorporated
`herein by reference in its entirety, discloses a system for
`monitoring traffic accidents including a plurality of micro
`phones and video cameras disposed at an intersection.
`Collision sounds are chosen from among the typical Sounds
`at an intersection. The Source of the collision Sounds is
`
`determined by comparing the time differences of the sounds
`received by each of the microphones. Image data from the
`cameras is recorded upon the occurrence of the collision.
`However, the Japanese reference discloses a system that is
`constantly photographing the accident scene thereby wast
`ing video resources.
`0009 U.S. Pat. No. 6,141,611 issued to Mackey et al.
`entitled “Mobile Vehicle Accident Data System” (“the
`Mackey reference”), expressly incorporated herein by ref
`erence in its entirety, discloses an on-board vehicle accident
`detection system including one or more video cameras that
`continuously record events occurring at a given scene.
`Camera images of the scene are digitally stored after com
`pression. An accident detector on-board the vehicle deter
`mines if an accident has occurred, and if so, the stored
`images are transmitted to a remote site for observation.
`However, the Mackey reference includes video cameras
`on-board the vehicles themselves, increasing the likelihood
`that the cameras would become damaged during an accident
`thereby rendering them impractical for accident-recording
`systems. Further, the on-board cameras image-capturing
`ability is severely limited due to the constraints of the
`vehicle themselves. Additionally, the Mackey reference dis
`closes a system that determines if an accident is present by
`the sudden acceleration or deceleration of the vehicle, with
`out the use of fixed microphones. The invention claimed by
`Mackey is on board the vehicle, it does nothing to solve the
`problem or record an accident in two vehicles which are not
`so equipped. Equipping every vehicle with this system is
`impractical and therefore not feasible.
`0010 U.S. Pat. No. 6,111,523 issued to Mee entitled
`“Method and Apparatus for Photographing Traffic in an
`Intersection', expressly incorporated herein by reference in
`its entirety, describes a system for taking photographs of
`vehicles at a traffic intersection by triggering a video camera
`to capture images wherein the triggering mechanism of the
`Video camera is based upon certain vehicle parameters
`including the speed of the vehicle prior to its entrance into
`the traffic intersection.
`0011 U.S. Pat. No. 6,088,635 issued to Cox et al. entitled
`“Railroad Vehicle Accident Video Recorder, expressly
`incorporated herein by reference in its entirety, discloses a
`system for monitoring the status of a railroad vehicle prior
`to a potential accident. The system employs a video camera
`mounted within the railroad car that continuously views the
`status of a given scene, and continuously stores the images
`of the scene. Like Mackey, it is impractical and therefore not
`feasible to equip every vehicle with this system.
`0012 U.S. Pat. No. 5,717.391 issued to Rodriguez
`entitled “Traffic Event Recording Method and Apparatus',
`expressly incorporated herein by reference in its entirety,
`describes a system for determining the condition of a traffic
`light and includes an audio sensor which monitors sound at
`all times. Sound detected above a certain decibel level
`triggers the recordation of Sounds, the time of day and the
`status of the traffic lights. However, Rodriguez fails to
`disclose video cameras or any image-capturing means.
`0013 U.S. Pat. No. 5,677,684 issued to McArthur
`entitled “Emergency Vehicle Sound-Actuated Traffic Con
`troller, expressly incorporated herein by reference in its
`entirety, describes a traffic controller system utilizing Sound
`detection means connected to a control box which contains
`
`Exhibit 1011
`Page 08 of 29
`
`
`
`US 2006/0092043 A1
`
`May 4, 2006
`
`a Switching mechanism that, in a first orientation, allows
`normal operation of traffic light control and a second ori
`entation that, upon the detection of an approaching siren,
`sets all traffic signals at an intersection to red to prohibit the
`entrance into the intersection of additional vehicles.
`0014 U.S. Pat. No. 5,539,398 issued to Halletal. entitled
`“GPS-based Traffic Control Preemption System”, expressly
`incorporated herein by reference in its entirety, discloses a
`system for determining if a vehicle issuing a preemption
`request to an emergency vehicle or police car is within an
`allowed approach of a traffic intersection, utilizing a GPS
`system.
`0015 U.S. Pat. No. 6,690,294 issued to Zierden entitled
`“System and method for detecting and identifying traffic law
`violators and issuing citations', expressly incorporated
`herein by reference, discloses a mobile or stationary traffic
`monitoring system for detecting violations of speed limits or
`other traffic laws by vehicle operators and issuing citations
`to an operator and/or vehicle owner Suspected of a violation
`using a digital camera to capture images of the operator
`and/or the vehicle, transmitting the captured images and
`other relevant data to an analysis center where the images
`and data are analyzed to determine whether to issue a
`citation and, if so, to issue the citation or take other appro
`priate law enforcement measures. The system captures
`images of a vehicle and/or vehicle operator Suspected of a
`traffic violation, determines the time and geographic loca
`tion of the Suspected violation, transmits the images and
`other data to an analysis center, issues citations to violators
`and derives revenue therefrom.
`0016 U.S. Pat. No. 5,938,717 to Dunne et al., expressly
`incorporated herein by reference, discloses a traffic control
`system that automatically captures an image of a vehicle and
`speed information associated with the vehicle and stores the
`image and information on a hard disk drive. The system uses
`a laser gun to determine whether a vehicle is speeding. The
`hard drive is later connected to a base station computer
`which is, in turn, connected to a LAN at which the infor
`mation from the hard drive is compared with databases
`containing data such as vehicle registration information and
`the like. The system automatically prints a speeding citation
`and an envelope for mailing to the registered owner of the
`vehicle
`0017 U.S. Pat. No. 5,734,337 to Kupersmit, expressly
`incorporated herein by reference, discloses a stationary
`traffic control method and system for determining the speed
`of a vehicle by generating two images of a moving vehicle
`and calculating the vehicle speed by determining the dis
`tance traveled by the vehicle and the time interval between
`the two images. The system is capable of automatically
`looking up vehicle ownership information and issuing cita
`tions to the owner of a vehicle determined to be speeding.
`0018 U.S. Pat. No. 5,948,038 to Daly et al., expressly
`incorporated herein by reference, discloses a method for
`processing traffic violation citations. The method includes
`the steps of determining whether a vehicle is violating a
`traffic law, recording an image of the vehicle committing the
`violation, recording deployment data corresponding to the
`violation, matching the vehicle information with vehicle
`registration information to identify the owner, and providing
`a traffic violation citation with an image of the vehicle, and
`the identity of the registered owner of the vehicle.
`
`0019. The I-95 Corridor Coalition, Surveillance Require
`ments/Technology, Ch. 4. Technology Assessment,
`expressly incorporated herein by reference, describes a
`number of different technologies suitable for incident detec
`tion. For example, AutoAlert: Automated Acoustic Detec
`tion of Traffic Incidents, was an IVHS-IDEA project which
`uses military acoustic sensor technologies, e.g., AT&T IVHS
`NET-2000TM. The AutoAlert system monitors background
`traffic noise and compares it with the acoustic signatures of
`previously recorded accidents and incidents for detection.
`See, David A. Whitney and Joseph J. Pisano (TASC, Inc.,
`Reading, Mass.), “AutoAlert: Automated Acoustic Detec
`tion of Incidents'. IDEA Project Final Report, Contract
`ITS-19, IDEA Program, Transportation Research Board,
`National Research Council, Dec. 26, 1995, expressly incor
`porated herein by reference. The AutoAlert system employs
`algorithms which provide rapid incident detection and high
`reliability by applying statistical models, including Hidden
`Markov Models (HMM) and Canonical Variates Analysis
`(CVA). These are used to analyze both short-term and
`time-varying signals that characterize incidents.
`0020. The Smart Call Box project (in San Diego, Calif.)
`evaluated the use of the existing motorist aid call box system
`for other traffic management strategies. The system tests the
`conversion of existing cellular-based call boxes to multi
`functional IVHS system components, to transmit the data
`necessary for traffic monitoring, incident detection, hazard
`ous weather detection, changeable message sign control, and
`CCTV control.
`0021. In 1992 the French Toll Motorway Companies
`Union initiated testing an Automatic Incident Detection
`(AID) technique proposed by the French National Institute
`for Research on Transportation and Security (INRETS). The
`technique consists of utilizing computers to analyze video
`images received by television cameras placed along the
`roadway. A “mask’ frames the significant part of the image,
`which typically is a three or four-lane roadway and the
`emergency shoulder. The computer processes five pictures a
`second, compares them two at a time, and analyzes them
`looking for points that have moved between two Successive
`pictures. These points are treated as objects moving along
`the roadway. If a moving object stops and remains stopped
`within the mask for over 15 seconds, the computer considers
`this an anomaly and sets off an alarm. In 1993, as part of the
`European MELYSSA project, the AREA Company con
`ducted a full scale test over an urban section of the A43
`motorway located east of Lyons. The roadway was equipped
`with 16 cameras on 10 meter masts or bridges with focal
`distances varying from 16 to 100 km, and fields of detection
`oscillating between 150 and 600 meters. Image Processing
`and Automatic Computer Traffic Surveillance (IMPACTS) is
`a computer system for automatic traffic Surveillance and
`incident detection using output from CCTV cameras. The
`algorithm utilized by the IMPACTS system takes a different
`approach from most other image processing techniques that
`have been applied to traffic monitoring. Road space and how
`it is being utilized by traffic is considered instead of iden
`tifying individual vehicles. This leads to a qualitative
`description of how the road, within a CCTV image, is
`occupied in terms of regions of empty road or moving or
`stationary traffic. The Paris London Evaluation of Integrated
`ATT and DRIVE Experimental Systems (PLEIADES) is
`part of the DRIVE Research Programme. The Automatic
`Traffic Surveillance (ATS) system has been installed into
`
`Exhibit 1011
`Page 09 of 29
`
`
`
`US 2006/0092043 A1
`
`May 4, 2006
`
`Maidstone Traffic Control Center and provides information
`on four separate CCTV images. This information will be
`used both in the Control Center and passed onto the Traffic
`Information Center via the PLEIADES Information Con
`troller (PIC) and data communications link. Instead of
`remote PCs there is a duplicate display of the Engineers
`workstation that is shown in the Control Office on a single
`computer monitor. The ATS system communicates data at
`regular intervals to the PIC. Any alarms that get raised or
`cleared during normal processing will get communicated to
`the PIC as they occur. The PIC uses the information received
`to display a concise picture of a variety of information about
`the highway region. The ATS system uses video from CCTV
`cameras taken from the existing Control Office Camera
`Multiplex matrix, while not interfering with its normal
`operation. When a camera is taken under manual control, the
`processing of the data for that image is Suspended until the
`camera is returned to its preset position.
`0022 Navaneethakrishnan Balraj, “Automated Accident
`Detection. In Intersections Via Digital Audio Signal Process
`ing' (Thesis, Mississippi State University, December 2003),
`expressly incorporated herein by reference, discusses, inter
`alia, feature extraction from audio signals for accident
`detection. The basic idea of feature extraction is to represent
`the important and unique characteristics of each signal in the
`form of a feature vector, which can be further classified as
`crash or non-crash using a statistical classifier or a neural
`network. Others have tried using wavelet and cepstral trans
`forms to extract features from audio signals such as speech
`signals. S. Kadambe, G. F. Boudreaux-Bartels, “Application
`of the wavelet transform for pitch detection of speech
`signals.” IEEE Trans. on Information Theory, vol. 38, no. 2,
`part 2, pp. 917-924, 1992: C. Harlow and Y. Wang, “Auto
`mated Accident Detection.” Proc. Transportation Research
`Board 80th Annual Meeting, pp. 90-93, 2001. Kadambe etal
`developed a pitch detector using a wavelet transform. One of
`the main properties of the dyadic wavelet transform is that
`it is linear and shift-variant. Another important property of
`the dyadic wavelet transform is that its coefficients have
`local maxima at a particular time when the signal has sharp
`changes or discontinuities. These two important properties
`of the dyadic wavelet transform help to extract the unique
`features of a particular audio signal. Kadambe et al made a
`comparison of the results obtained from using dyadic wave
`let transforms, autocorrelation, and cepstral transforms. The
`investigation showed that the dyadic wavelet transform pitch
`detector gave 100% accurate results. One reason for the
`difference in the results was that the other two methods
`assume stationarity within the signal and measure the aver
`age period, where as the dyadic wavelet transform takes into
`account the non-stationarities in the signal. Hence, the
`dyadic wavelet transform method would be the best to
`extract feature when the signals are non-stationary. Harlow
`et al developed an algorithm to detect traffic accidents at
`intersections, using an audio signal as the input to the
`system. The algorithm uses the Real Cepstral Transform
`(RCT) as a method to extract features. The signals recorded
`at intersections include brake, pile drive, construction and
`normal traffic Sounds. These signals are segmented into
`three-second sections. Each of these three second segmented
`signals is analyzed using RCT. RCT is a method where the
`signal is windowed for every 100 msec using a hamming
`window with an overlap of 50 msec. Thus, for a given
`three-second signal, there will be almost 60 segments of 100
`
`msec duration each. RCT is applied to each of these seg
`ments, and the first 12 coefficients are used as the features.
`The features obtained using the RCT are then classified as
`“crash” or “non-crash” using a neural network.
`0023 Balraj's experimental results showed that among
`the three different statistical classifiers investigated, maxi
`mum likelihood and nearest neighbor performed best,
`although this had high computational costs. Haar,
`Daubechies, and Coiflets provided the best classification
`accuracies for a two-class system. Among the five different
`feature extraction methods analyzed on the basis of the
`overall accuracy, RCT performed best. The second-genera
`tion wavelet method, the lifting scheme, was also investi
`gated. It proved computationally efficient when compared to
`DWT. Thus, it was concluded that the optimum design for
`an automated system would be a wavelet-based feature
`extractor with a maximum likelihood classifier. Thus the
`choice of DWT or the lifting scheme would be preferred for
`a real-time system.
`0024. In any and/or all of the embodiments described
`herein, the systems, equipment systems, Subsystems,
`devices, components, and/or appliances, of and/or utilized in
`any of the respective embodiments, can include and/or can
`utilize the teachings and/or the subject matter of the follow
`ing U.S. Patents, the subject matter and teachings of which
`are hereby incorporated by reference herein and form a part
`of the disclosure of this patent application: U.S. Pat. No.
`6,009,356 (Monroe, Dec. 28, 1999); U.S. Pat. No. 5,890,079
`(Beemer, II, et al., Sep. 7, 1999); U.S. Pat. No. 5,845,240
`(Fielder, Dec. 1, 1998): U.S. Pat. No. 5.948,026 (Levine,
`Mar. 30, 1999); U.S. Pat. No. 5,446,659 (Yamawaki, Aug.
`29, 1995); U.S. Pat. No. 5,056,056 (Gustin, Oct. 8, 1991);
`U.S. Pat. No. 6,718.239 (Rayner, Apr. 6, 2004); U.S. Pat.
`No. 6,449,540 (Rayner, Sep. 10, 2002); U.S. Pat. No.
`6,684,137 (Takagi, et al., Jan. 27, 2004); U.S. Pat. No.
`6,633.238 (Lemelson, et al., Oct. 14, 2003); U.S. Pat. No.
`6,226,389 (Lemelson, et al., May 1, 2001): U.S. Pat. No.
`6,630,884 (Shanmugham, Oct. 7, 2003); U.S. Pat. No.
`6,600,417 (Lerg, et al., Jul. 29, 2003); U.S. Pat. No. 6,288,
`643 (Lerg, et al., Sep. 11, 2001): U.S. Pat. No. 6,281,792
`(Lerg, et al., Aug. 28, 2001): U.S. Pat. No. 6,580,373
`(Ohashi, Jun. 17, 2003); U.S. Pat. No. 6,574,548 (DeKock,
`et al., Jun. 3, 2003); U.S. Pat. No. 6,472,982 (Eida, et al.,
`Oct. 29, 2002); U.S. Pat. No. 5,784,007 (Pepper, Jul. 21,
`1998): U.S. Pat. No. 6,466.260 (Hatae, et al., Oct. 15, 2002);
`U.S. Pat. No. 6,429,812 (Hoffberg, Aug. 6, 2002); U.S. Pat.
`No. 6,252,544 (Hoffberg, Jun. 26, 2001): U.S. Pat. No.
`6,389,340 (Rayner, May 14, 2002); U.S. Pat. No. 6,324.450
`(Iwama, Nov. 27, 2001): U.S. Pat. No. 6,211.907 (Scaman,
`et al., Apr. 3, 2001); and U.S. Pat. No. 5,689,442 (Swanson,
`et al., Nov. 18, 1997): U.S. Pat. No. 6,647,270 (Himmel
`stein, Nov. 11, 2003); U.S. Pat. No. 6,574,538 (Sasaki, Jun.
`3, 2003); U.S. Pat. No. 6,573,831 (Ikeda, et al., Jun. 3,
`2003); U.S. Pat. No. 6,542,077 (Joao, Apr. 1, 2003); U.S.
`Pat. No. 6,404,352 (Ichikawa, et al., Jun. 11, 2002); U.S. Pat.
`No. 6,401,027 (Xu, et al., Jun. 4, 2002); U.S. Pat. No.
`6,392,692 (Monroe, May 21, 2002); U.S. Pat. No. 6,339,370
`(Ruhl, et al., Jan. 15, 2002); U.S. Pat. No. 6,314,364
`(Nakamura, Nov. 6, 2001): U.S. Pat. No. 6,163,338
`(Johnson, et al., Dec. 19, 2000); U.S. Pat. No. 6,154,658
`(Caci, Nov. 28, 2000); U.S. Pat. No. 6,091,956 (Hollenberg,
`Jul. 18, 2000); U.S. Pat. No. 6,087,960 (Kyouno, et al., Jul.
`11, 2000); U.S. Pat. No. 6,075,466 (Cohen, et al., Jun. 13,
`2000); U.S. Pat. No. 5,990,801 (Kyouno, et al., Nov. 23,
`
`Exhibit 1011
`Page 10 of 29
`
`
`
`US 2006/0092043 A1
`
`May 4, 2006
`
`1999); U.S. Pat. No. 5,943,428 (Seri, et al., Aug. 24, 1999);
`U.S. Pat. No. 5,699,056 (Yoshida, Dec. 16, 1997); U.S. Pat.
`No. 5,353,023 (Mitsugi, Oct. 4, 1994); U.S. Pat. No. 5,025,
`324 (Hashimoto, Jun. 18, 1991); U.S. 2004.0022416 (Lemel
`son, et al., Feb. 5, 2004); U.S. 20020008.637 (Lemelson, et
`al., Jan. 24, 2002); U.S. 20030225516 (DeKock, et al., Dec.
`4, 2003); U.S. 20020193938 (DeKock, et al., Dec. 19,
`2002); U.S. 20030222981 (Kisak, et al., Dec. 4, 2003); U.S.
`20030214405 (Lerg, et al., Nov. 20, 2003); U.S.
`20020008619 (Lerg, et al., Jan. 24, 2002); U.S.
`2003.0125853 (Takagi, et al., Jul. 3, 2003); U.S.
`20030081935, 20030081934, 20030081128, 20030081127,
`20030081122, 20030081121, and 20030080878 (Kirmuss,
`May 1, 2003); U.S. 2002012 1969 (Joao, Sep. 5, 2002); U.S.
`20020147982 (Naidoo, et al., Oct. 10, 2002); U.S.
`2003.0062997 (Naidoo, et al., Apr. 3, 2003); U.S.
`20010005804 (Rayner, Jun. 28, 2001): U.S. 20020163579
`(Patel, et al., Nov. 7, 2002); U.S. 20020170685 (Weik, et al.,
`Nov. 21, 2002); U.S. 20030011684 (Narayanaswami, et al.,
`Jan. 16, 2003); U.S. 20030041329 (Bassett, Feb. 27, 2003);
`U.S. 2003.0053536 (Ebrami, Mar. 20, 2003); and U.S.
`20030067542 (Monroe, Apr. 10, 2003).
`0.025 The following references are incorporated herein
`by references as if explicitly set forth herein:
`0026 www.itsdocs.fhwa.dot.gov/JPODOCS/REPT.
`S TE/36D01.PDF;
`0027 www.ndsu.nodak.edu/ndsu/ugpti/MPC Pubs/
`html/MPCO1-122.html;
`0028) www.intelligenthighway.com/ITS/IMITS.pdf,
`0029 stat-www.berkeley.edu/users/kwon/papers/inc de
`tection.pdf
`0030) www-users.cs.umn.edu/~masoud/publications/
`harini-intersection-itsc-2002.pdf.
`0031) The I-80 Experiment: Real-Time Algorithms for
`Travel Time Estimates and Incident Detection—Alex Ska
`bardonis;
`0032) Dougherty, M. S., Chen, H. and Montgomery, F.
`(1998), Principle components based incident detection Proc.
`5th International Conference on Applications of Advanced
`Technologies in Transportation, Newport Beach, Calif., iee
`explore.ieee.org/xpl/tocresult.jsp?isNumber=14013;
`0033 Karim, A. and Adeli, H. (2003), “Fast Automatic
`Incident Detection on Urban and Rural Freeways. Using the
`Wavelet Energy Algorithm,” Journal of Transportation Engi
`neering, ASCE, Vol. 129, No. 1, pp. 57-68. www.pubs.as
`ce.org/WWWaisplay.cgi?0300116;
`0034) Chen, H: Boyle, R D; Montgomery, FO: Kirby, H
`R. Novelty detection for motorway incident management in:
`Proceedings of the Workshop on Engineering for Intelligent
`Systems. 1998. portal.acm.org/citation.cfm?id=643217&dl=
`ACM&col=GUIDE;
`0035) Karim, A. and Adeli, H. (2002), “Comparison of
`Fuzzy Wavelet Radial Basis Function Neural Network Free
`way Incident Detection Model with California Algorithm.”
`Journal of Transportation Engineering, ASCE, Vol. 128, No.
`1,
`pp.
`21-30,
`www.pubs.asce.org/WWW display
`.cgi?02001 17:
`
`0036 Chien-Hua Hsiao, Ching-Teng Lin, and Michael
`Cassidy, “Application of Fuzzy Logic and Neural Networks
`to Automatically Detect Freeway Traffic Incidents”, Journal
`of Transportation Engineering, Vol. 120, No. 5, September/
`October 1994, pp. 753-772:
`0037 A. Samant & H. Adeli, “Feature Extraction for
`Traffic Incident Detection Using Wavelet Transform and
`Linear Discriminant Analysis’, Computer-Aided Civil and
`Infrastructure Engineering, Volume 15 Issue 4 Page 241
`(July 2000);
`0038 Hojat Adelil and Asim Karim, “Fuzzy-Wavelet
`RBFFN Model For Freeway Incident Detection”, Journal of
`Transportation Engineering, Vol. 126, No. 6, November/
`December, 2000;
`0.039
`H. Veeraraghavan, O. Masoud, N. P. Papanikol
`opoulos, "Computer vision algorithms for intersection moni
`toring, IEEE Trans. on Intelligent Transportation Systems,
`vol. 4, no. 2, pp. 78-89, June 2003;
`0040 K. Stubbs, H. Arumugam, O. Masoud, C.
`McMillen, H. Veeraraghavan, R. Janardan, N. P. Papanikol
`opoulos, “A real-time collision warning system for intersec
`tions', in Proc. ITS America 13th Annual Meeting, Minne
`apolis, Minn., May 2003.:
`0041
`H. Veeraraghavan, O. Masoud, N. P. Papanikol
`opoulos, “Vision-based monitoring of intersections', in
`Proc. IEEE 5th International Conference on Intelligent
`Transportation Systems, pp. 7-12. Singapore, September
`2002.:
`0042 Petty, K., Ostland, M., Kwon, J., Rice, J. and
`Bickel, P. (2002) “A New Methodology for Evaluating
`Incident Detection Algorithms. Transportation Research,
`Part C, Volume 10, pp. 189-204.
`0.043
`Peter T. Martin, Joseph Perrin, Blake Hansen, Ryan
`Kump, Dan Moore, “Incident Detection Algorithm Evalua
`tion' (University of Utah, Prepared for Utah Department of
`Transportation) March 2001; and
`0044 Dastidar, Samanwoy & Adeli, Hojat (2003),
`“Wavelet-Clustering-Neural Network Model for Freeway
`Incident Detection.”. Computer-Aided Civil and Infrastruc
`ture Engineering 18 (5), 325-338.
`
`SUMMARY OF THE INVENTION
`0045. Many of the known vehicle accident detection
`systems are limited in their ability to capture or process
`accurate data or to accurately and timely send the processed
`data to the proper location to enable authorities to properly
`assess accident damage and liability. Further, these systems
`generally do not incorporate advance