throbber
ON OF DRIVER FATIGUE
`
`EYE-TRACKING FOR DETECT
`Martin Eriksson
`Nikolaos P. Papanikolopoulos
`Artificial Intelligence, Robotics, and Vision Laboratory
`Department of Computer Science, University of Minnesota,
`
`Minneapolis, MN 55455
`E-mail: {eriksson, npapas}@cs.umn.edu
`Keywords: driver fatigue, eye-tracking, template matching.
`identification and database retrieval [ 3 ] , There
`Abstract
`are
`also many
`real-time
`systems, being
`In this paper, we describe a system that locates
`developed
`in order
`to
`track
`face
`features
`and tracks the eyes of a driver. The purpose of
`[15,13,17]. These kinds of real-time systems
`such a system is to perform detection of driver
`generally consist of three components:
`fatigue. By mounting a small camera inside the
`a) Localization of the eyes (in the first frame),
`car, we can monitor the face of the driver and
`b) Tracking the eyes in the subsequent frames,
`look for eye-movements which indicate that the
`c) Detection of failure in tracking.
`driver is no longer in condition to drive. In such
`a case, a warning signal should be issued. This
`paper describes how to find and track the eyes.
`We also describe a method that can determine if
`the eyes are open or closed. The primary
`criterion for the successful implementation of
`it must be highly non-
`this system is that
`intrusive. The system should start when
`the
`ignition is turned on without having the driver
`initiate the system. Nor should the driver be
`responsible for providing any feedback to the
`system. The system must also operate regardless
`of the texture and the color of the face. It must
`also be able to handle diverse conditions, such as
`changes in light, shadows, reflections, etc.
`
`Localization of the eyes involves looking at the
`entire image of the face and determining the
`eye-envelopes
`(the areas around
`the eyes).
`During
`tracking
`in subsequent
`frames,
`the
`search-space
`reduced
`to
`the
`area
`is
`corresponding
`to
`the eye-envelopes
`the
`in
`current frame. This tracking can be done at
`relatively
`low computational effort, since the
`search-space is significantly reduced. In order to
`detect failure in the tracking, general constraints
`such as distance between the eyes and horizontal
`alignment of the two eyes can be used.
`the next
`This paper is organized as follows: In
`section, we describe some of the previous work
`in
`this area. Afterwards, we describe
`the
`experimental
`setup, and how
`the
`system
`operates. Then, we proceed to the description of
`the algorithm for
`the detection of
`fatigue.
`Finally, we present results and future work.
`PREVIOUS W
`for
`Many methods
`have been proposed
`localizing
`facial
`features
`in
`images
`[2,4,5,6,8,10,11,12,19]. These methods
`can
`roughly
`be divided
`into
`two categories:
`Template-based matching and Feature-based
`matching. These techniques are compared by
`Poggio and Brunelli [ 11. One popular template-
`face
`matching
`technique
`for extraction of
`features is to use deformnhle templates
`[5,16],
`which are similar to the active snakes introduced
`by Kass [7], in the sense that they apply energy
`
`INTRODUCTION
`Driver fatigue is an important factor in a large
`number of accidents. Lowering the number of
`fatigue-related accidents would not only save
`society a significant amount financially, but also
`reduce personal suffering. We believe that by
`monitoring the eyes, the symptoms of driver
`fatigue in our proposed system can be detected
`early enough to avoid several of these accidents.
`Detection of fatigue involves a sequence of
`images of a face, and observation of eye-
`movements and blink patterns.
`The analysis of face images is a popular research
`area with applications such as face recognition,
`virtual tools and handicap aids [9,14], human
`
`314
`
`0-7803~4269-0/97/~10.00 0 1998 IEEE
`
`Page 1 of 6
`
`SAMSUNG EXHIBIT 1008
`Samsung v. Image Processing Techs.
`
`

`

`the computation of
`minimization based on
`image-forces. In feature-based matching,
`the
`system uses knowledge about some geometrical
`constraints. For example, a face has two eyes,
`one mouth and one nose in specific relative
`locations.
`One interesting application for face recognition
`was developed by Stringa [12]. He used the
`observation that the eyes are regions of rapidly
`changing intensity. We use a similar approach
`on a reduced version of the image. Another
`approach, developed by Steifelhagen et al. [ 131
`uses connected regions in order to extract the
`dark disks corresponding to the pupils. Rather
`than looking for the pupils, we used the fact that
`the entire eye-regions are darker
`than
`their
`surroundings, again allowing us to use
`the
`reduced image in order to extract these rough
`regions at a reduced computational cost. The
`systems described in [15] and [13] use color
`information in order to extract the head from
`the background. In order to avoid dependence
`on a fairly colorless background, we decided to
`again use the reduced image and localize the
`symmetry axis [18]. Since the driver will be
`looking almost straight ahead, there will be a
`well defined vertical symmetry line between the
`eyes.
`Many different templates have been described
`for finding the shape of an eye. Xie et al. [ 161
`developed a deformable template consisting of
`10 cost equations, based on image intensity,
`image gradient and
`internal forces of
`the
`template. Since we are greatly concerned about
`computational speed, we decided to use only the
`two cost equations dealing with image intensity.
`Once the eyes are found, the search-space in the
`subsequent
`frames
`is
`limited
`to
`the area
`surrounding
`the found eye-regions.
`In
`the
`system by Stiefelhagen et al., the darkest pixel
`(which is likely to be a pixel inside the pupil) is
`used for tracking, allowing high computational
`speed. Another approach [5] is to perform edge-
`detection on the region of interest and then track
`the region with a high concentration of edges.
`THE SYSTEM
`When the system starts, frames are continuously
`fed from the camera to the computer. We use the
`initial frame in order
`to
`localize
`the eye-
`positions. Once the eyes are localized, we start
`the tracking process by using information in
`previous frames in order to achieve localization
`in subsequent frames. During tracking, error-
`detection is performed in order to recover from
`possible tracking failure. When a tracking failure
`is detected, the eyes are relocalized. During
`
`the detection of
`tracking, we also perform
`fatigue. At this point, we count consecutive
`frames during which the eyes are closed. If this
`number gets too large, we issue a warning signal.
`Experimental setup
`The final system will consist of a camera
`pointing at the driver. The camera is to be
`mounted on the dashboard inside the vehicle.
`For the system we are developing, the camera is
`stationary and will not adjust its position or
`zoom during operation. For experimentation, we
`are using a JVC color video camera, sending the
`frames
`to a Silicon Graphics
`Indigo. The
`grabbed frames are represented in RGB-space
`with 8-bit pixels (256 colors). We do not use any
`specialized hardware for image processing.
`Localization of the eyes
`We localize the eyes in a top-down manner,
`reducing the search-space at each step. The steps
`are:
`1 . Localization of the face.
`2. Computation of the vertical location of the
`eyes.
`3. Computation of the exact location of the
`eyes.
`4. Estimation of the position of the iris.
`
`Localization of the face. Since the face of a
`driver is symmetric, we use a symmetry-based
`approach, similar to [ 181. We found that in order
`for this method to work, it is enough to use a
`subsampled, gray-scale version of the image. A
`symmetry-value is then computed for every
`
`Figure 1. The symmetry histogram.
`
`315
`
`SAMSUNG EXHIBIT 1008
`Page 2 of 6
`
`

`

`,"=I
`
`, = I
`
`Figure 2. The original image, the edges and the histogram of projected edges.
`Find the exact location of the eyes. In
`pixel-column in the reduced image. If the image
`order
`to
`find
`the eye-regions
`given
`the
`is represented as I ( x , y ) then the symmetry-
`proceeding processing, we rely on the fact that
`value for a pixel-column i s given by
`the eyes correspond to intensity-valleys in the
`image. Given that, we can threshold the image
`and then extract the connected regions. We used
`a raster-scan algorithm on the reduced image in
`order to extract these regions. In general, our
`raster-scan algorithm found 4-5
`regions. In
`order
`to
`resolve which of
`these
`regions
`correspond to the eyes, we use the information
`in H(y). We try to find a peak corresponding to
`a row in the image with two connected regions
`on. l h e three best peaks in H ( y ) are considered.
`We also use general constraints, such chat both
`eyes must be located "fairly close" to the centei-
`of the face.
`is to find a
`The difficulty with this method
`threshold that will generate the correct eye-
`regions. We used a method called adaptive
`thresholding [13] that starts out with a low
`threshold. If two good eye-regions are found,
`that threshold is stored, and used the next time
`the eyes have to be localized. If no good eye-
`regions are found,
`the system automatically
`attempts with a higher
`threshold, until
`the
`regions are found.
`Estimation of the position of the iris. Once
`the eye-regions are localized, we can apply a
`very simple template in order to localize the iris.
`
`S(x) is computed for x E [k,xsize - k ] where k
`is the maximum distance from the pixel-column
`that symmetry is measured, and xsize
`is the
`width of the image. The x corresponding to the
`lowest value of S ( x ) is the center of the face. The
`result from this process is shown in Figure 1 .
`The search-space is now limited to the area
`around this line, which reduces the probability
`of having distracting features in the background.
`Computation of the vertical
`location of
`the eyes. As suggested by Stringa [12], we use
`the observation that eye-regions correspond to
`regions of high spatial frequency. Again we are
`working with the reduced image. We create the
`gradient-map, G(x,y), by applying an edge
`detection algorithm on the reduced image. Any
`edge-detection method could be used. We
`choose to use a very simple and fast method
`called pixel-differentiation, that assigns G(x,y) =
`I(x,y) - l(x-1,y). We selected this method since it
`does not involve any convolution. G(x,y) will
`now reveal areas of high spatial frequency. By
`projecting G(x,y) onto its vertical axis, we get a
`histogram H ( y ) :
`
`H ( y ) = E G ( i , y ) .
`r=l
`Since both eyes are likely to be positioned at the
`same row, H ( y ) will have a strong peak on that
`row. However, in order to reduce the risk of
`error, we consider the best three peaks in H ( y )
`for further search rather than just the maximum.
`This process is illustrated in Figure 2.
`
`316
`
`Figure 3. The eye-template.
`We constructed a template consisting of two
`circles, one inside the other. A good match
`
`SAMSUNG EXHIBIT 1008
`Page 3 of 6
`
`

`

`Figure 4. Snapshots from the system during tracking. Note that in the second image, the system
`missed tracking of one eye.
`
`would result in many dark pixels in the area
`inside the inner circle, and many bright pixels in
`the area between the two circles. The template is
`shown in Figure 3. This match occurs when the
`inner circle is centered on the iris and the
`outside circle covers the sclera.
`The match M(al,a2) is computed as
`
`( ~ . q ) ~ a 2
`(p,q)-l
`A low value for M(al,a2) corresponds to a good
`match. The template is matched across
`the
`predicted eye-region, and the best match
`is
`reported.
`
`Tracking the eyes
`We track the eye by looking for the darkest
`pixel in the predicted region [13]. In order to
`recover from tracking errors, we make sure that
`none of the geometrical constraints are violated.
`If they are, we relocalize the eyes in the next
`frame. To find the best match for the eye-
`template, we initially center it at the darkest
`
`in
`pixel, and then perform a gradient descent
`order to find a local minimum. In Figure 4, we
`show a few snapshots during tracking.
`
`DETECTION OF FATIGUE
`As the driver becomes more fatigued, we expect
`the eye-blinks to last longer. We count the
`number of consecutive frames that the eyes are
`closed in order to decide the condition of the
`driver. For this, we need a robust way
`to
`determine if the eyes are open or closed; so we
`developed a method that looks at the horizontal
`histogram across the pupil.
`During initialization (the first frames after the
`driver has settled down), an average match over a
`number of frames is calculated. When the match
`in a frame is
`“significantly”
`lower than the
`average, we call that frame a closed frame. If the
`match is close to the average, we call that a n
`open frame. After C consecutive closed frames,
`we issue a warning signal, where C
`is
`the
`corresponding
`of
`number
`frames
`to
`approximately 2 to 2.5 seconds (the time when
`the eyes have been closed for too long).
`
`317
`
`SAMSUNG EXHIBIT 1008
`Page 4 of 6
`
`

`

`the
`
`Horizontal histogram across
`pupil
`We use the characteristic curve generated by
`plotting
`the image-intensities along
`the
`line
`going through the pupil from left to right, as
`shown in Figure 5. The pupil is always the
`darkest point. Surrounding the pupil, we have
`the iris, which is also very dark. To the right and
`left of the iris is the white sclera. In Figure S we
`show two curves, one corresponding to an open
`eye, and one corresponding to a closed eye.
`Note that the curve corresponding to the closed
`eye is very flat.
`We compute the matching function M(x,y) as
`y> / min{I(x - r, Y>, I ( x + r, y> 1
`M ( x , y> =
`where (x,y) is the computed center of the pupil
`and r is the radius of the iris. Z(x,y) is the image
`intensity at (x,~). When the eye is open, the
`valley in the intensity-curve corresponding
`to
`the pupil will be surrounded by two large peaks
`corresponding
`to the sclera. When the eye is
`closed, this curve is usually very flat in the
`center. However, in the latter case there is no
`pupil to center the curve on, which can lead to a
`very unpredictable shape. In order to minimize
`the risk of having one big peak nearby (due to
`noise), we always use the minimum peak at the
`distance r
`from the pupil. This will lead to a
`good match when the eye is open, and very
`likely to a bad match when the eye is closed.
`RESULTS AND FUTURE WORK
`We simulated
`three “test-drives” where we
`
`Figure 5. Histograms corresponding
`open and a closed eye, respectively.
`
`to an
`
`318
`
`the accuracy of
`measured
`the detection of
`opened/closed eyes. In each
`test-drive, we
`simulated 10 long eye-blinks, and recorded how
`many were computed by the system. In each
`test-drive, the driver had the head turned in a
`different angle. The results are shown in Table
`1.
`For this test, we did not allow for any rapid head
`movements, since we wanted to simulate the
`situation when the driver is tired. For small head-
`movements, the system rarely loses track of the
`eyes, as we can see from the results. We can also
`see that when the head is turned
`too much
`sideways, we had some false alarms. However, in
`the case where the head is tilted forward (which
`is the most likely posture when the driver is
`tired), the system operated perfectly.
`When we perform the detection of driver fatigue,
`we operate on frames of size 640 by 320. This
`frame-size allows us to operate at approximately
`5 frames per second. In order to track the eyes,
`without detecting fatigue,
`it is enough to use
`frames of size 320 by 160, which allows a
`frame-rate of approximately 15 frames / second.
`At this point, the system has problems localizing
`eyes when the person is wearing glasses, or has a
`large amount of facial hair. We believe that by
`using a small set of face templates, similar to
`[lS], we will be able to avoid this problem,
`without losing anything in performance. Also,
`we are not using any color-information in the
`image. By using techniques described in [13],
`we can further enhance robustness.
`Currently, we do not adjust zoom or direction of
`the camera during operation. Future work may
`be to automatically zoom in on the eyes, once
`they are localized. This would avoid the trade-
`off between having a wide field of view in order
`to locate the eyes, and a narrow field of view In
`order to detect fatigue.
`the number of
`We are only
`looking at
`consecutive frames where the eyes are closed. At
`that point, it may be too late to issue the signal.
`By study the eye-movement patterns, we are
`hoping to find a method to generate the alert
`signal at an earlier stage.
`
`SAMSUNG EXHIBIT 1008
`Page 5 of 6
`
`

`

`CONCLUSIONS
`We have developed a system that localizes and
`tracks the eyes of a driver in order to detect
`fatigue. The system uses a combination of
`template-based matching
`and
`feature-based
`matching in order to localize the eyes. During
`tracking, the system is able to decide if the eyes
`are open or closed. When the eyes have been
`closed too long, a warning signal is issued.
`Several experimental results are presented.
`
`ACKNOWLEDGMENTS
`the ITS
`This work has been supported by
`Institute at the University of Minnesota, the
`Minnesota Department
`of
`Transportation
`through Contracts #71789-72983-169
`and
`#71789-72447-159,
`the National
`Science
`Foundation
`through Contracts #IRI-94 10003
`and #IRI-9502245,
`and
`the Center
`for
`Transportation Studies.
`
`REFERENCES
`
`“Face
`and T. Poggio,
`R. Brunelli
`recognition: Features versus Templates, ”
`IEEE Transactions on Pattern Analysis and
`Machine Intelligence, Vol. 15, No. 10, pp.
`1042-1052, 1993.
`G. Chow and X. Li, “Towards a System for
`Automatic Feature Detection, ” Pattern
`Recognition, Vol. 26, No. 12, pp. 1739-
`1755, 1993.
`I.J. Cox, J. Ghosn, P.N. Yianilos, “Feature-
`Based
`Recognition
`Using Mixture-
`NEC
`Research
`Institute,
`Distance,“
`Technical Report 95 - 09, 1995.
`J.R. Lishman,
`I. Craw, H. Ellis and
`“Automatic Extraction of Face-Features,”
`Pattern Recognition Letters, 5, pp. 183-1 87,
`1987.
`L.C. De Silva, K. Aizawa and M. Hatori,
`“Detection and Tracking of Facial Features
`by Using Edge Pixel Counting
`and
`Deformable Circular Template Matching, ”
`IEICE Transaction on Information and
`Systems, Vol. E78-D No. 9, pp. 1195-1207,
`September 1995.
`C. Huang and C. Chen, “Human Facial
`Feature Extraction for Face Interpretation
`and Recognition,” Pattern Recognition,
`Vol. 25, NO. 12 pp. 1435-1444, 1992.
`M. Kass, A. Witkin and D. Terzopoulos,
`“Snakes:
`Active
`Contour Models, ”
`
`International Journal of Computer Vision,
`pp. 321-331, 1988.
`[8] X. Li and N. Roeder, “Face Contour
`Extraction
`from Front-View
`Images,”
`Pattern Recognition, Vol. 28, No. 8, pp.
`1167-1 179, 1995.
`[9] K.P. White, T.E. Hutchinson and J.M.
`Carley, “Spatially Dynamic Calibration of
`IEEE
`an
`Eye-Tracking
`System, ”
`Transactions on Systems, Man,
`and
`Cybernetics, Vol. 23, No. 4, pp. 1 162-1 168,
`1993.
`[ l o N. Roeder and Xiabo Li, “Accuracy
`Analysis for Facial Feature Detection, ”
`Pattern Recognition, Vol. 29, No. 1, pp.
`143-157, 1996.
`[11 Y. Segawa, H. Sakai, T. Endoh, K.
`Murakami, T. Toriu and H. Koshimizu,
`“Face
`Recognition
`Through Hough
`Transform
`for
`Irises Extraction
`and
`Projection
`Procedures
`for
`Parts
`Localization, ” Pacific Rim International
`Conference on Artificial Intelligence, pp.
`625-636, 1996.
`for Face
`[12] L. Stringa, “Eyes Detection
`A rt$cial
`Applied
`Recognition,”
`Intelligence, No. 7, pp. 365-382, 1993.
`[13] R. Stiefelhagen, J. Yang and A. Waibel, “ A
`Model-Based Gaze-Tracking
`System,”
`International
`IEEE Joint Symposia on
`Intelligence and Systems, pp. 304-3 I O ,
`1996.
`[14] E.R. Tello, “Between Man and Machine,”
`BYTE, September, pp. 288-293, 1983.
`and
`1.51 D. Tock and I. Craw, “Tracking
`Real- Time
`Measuring Drivers Eyes,”
`Computer Vision, pp. 71-89, 1995.
`161 X. Xie, R. Sudhakar and H. Zhuang, “ O n
`Improving Eye Features Extraction Using
`Pattern
`Deformable
`Templates,”
`Recognition, Vol. 27, No. 6, pp. 791-799,
`1994.
`171 X. Xie, R. Sudhakar and H. Zhuang, “Real-
`Time Eye Feature Tracking from a Video
`Image Sequence Using Kalman Filter,”
`IEEE Transactions on Systems, Man, and
`Cybernetics, Vol. 25, No. 12, pp. 1568-
`1577, 1995.
`181 T. Yo0 and I. Oh, “Extraction of Face
`Region and Features Based on Chromatic
`Properties of Human Faces,“ Pacific Rim
`International Conference on Artificial
`Intelligence, pp. 637-645, 1996.
`[19] H. Wu, Q. Chen, and M. Yachida, “Facial
`Feature Extraction and Face Verification,”
`International Conference on Pattern
`Recognition, pp. 484-488, 1996.
`
`319
`
`SAMSUNG EXHIBIT 1008
`Page 6 of 6
`
`

This document is available on Docket Alarm but you must sign up to view it.


Or .

Accessing this document will incur an additional charge of $.

After purchase, you can access this document again without charge.

Accept $ Charge
throbber

Still Working On It

This document is taking longer than usual to download. This can happen if we need to contact the court directly to obtain the document and their servers are running slowly.

Give it another minute or two to complete, and then try the refresh button.

throbber

A few More Minutes ... Still Working

It can take up to 5 minutes for us to download a document if the court servers are running slowly.

Thank you for your continued patience.

This document could not be displayed.

We could not find this document within its docket. Please go back to the docket page and check the link. If that does not work, go back to the docket and refresh it to pull the newest information.

Your account does not support viewing this document.

You need a Paid Account to view this document. Click here to change your account type.

Your account does not support viewing this document.

Set your membership status to view this document.

With a Docket Alarm membership, you'll get a whole lot more, including:

  • Up-to-date information for this case.
  • Email alerts whenever there is an update.
  • Full text search for other cases.
  • Get email alerts whenever a new case matches your search.

Become a Member

One Moment Please

The filing “” is large (MB) and is being downloaded.

Please refresh this page in a few minutes to see if the filing has been downloaded. The filing will also be emailed to you when the download completes.

Your document is on its way!

If you do not receive the document in five minutes, contact support at support@docketalarm.com.

Sealed Document

We are unable to display this document, it may be under a court ordered seal.

If you have proper credentials to access the file, you may proceed directly to the court's system using your government issued username and password.


Access Government Site

We are redirecting you
to a mobile optimized page.





Document Unreadable or Corrupt

Refresh this Document
Go to the Docket

We are unable to display this document.

Refresh this Document
Go to the Docket