throbber
Case 1:14-cv-02396-PGG-MHD Document 153-14 Filed 06/28/19 Page 1 of 21
`Case 1:14-cv-02396—PGG-MHD Document 153-14 Filed 06/28/19 Page 1 of 21
`
`EXHIBIT M
`
`EXHIBIT M
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`

`

`Case 1:14-cv-02396-PGG-MHD Document 153-14 Filed 06/28/19 Page 2 of 21
`
`Pattern Classification
`and Scene Analysis
`Richard O. Duda and
`Peter E. Hart
`
`It
`
`r’"
`
`e3
`
`t~
`
`Ct~
`
`0 Wiley-
`
`i nterscience
`
`

`

`Case 1:14-cv-02396-PGG-MHD Document 153-14 Filed 06/28/19 Page 3 of 21
`
`PATTERN
`CLASSIFICATION
`AND SCENE
`ANALYSIS
`
`RICHARD O. DUDA
`PETER E. HART
`
`Stanford Research Institute,
`Menlo Park, California
`
`A WILEY-INTERSCIENCE PUBLICATION
`
`John Wiley & Sons
`
`New York London Sydney Toronto
`
`

`

`Case 1:14-cv-02396-PGG-MHD Document 153-14 Filed 06/28/19 Page 4 of 21
`
`Copyright © 1973, by John Wiley & Sons, Inc.
`
`All rights reserved. Published simultaneously in Canada.
`
`No part of this book may be reproduced by any means,
`nor transmitted, nor translated into a machine language
`without the written permission of the publisher.
`
`Library of Congress Cataloging in Publication Data
`
`Duda, Richard O.
`Pattern classification and scene analysis.
`
`"A Wiley-interscience publication."
`Includes bibliographical references.
`1. Perceptrons. 2. Statistical decision.
`2. Hart, Peter E., joint author. II. Title.
`Q327.D83 001.53’3 72-7008
`
`ISBN 0-471-22361-1
`
`Printed in the United States of America
`
`109876543
`
`

`

`Case 1:14-cv-02396-PGG-MHD Document 153-14 Filed 06/28/19 Page 5 of 21
`
`CONTENTS
`
`Part I PATTERN CLASSIFICATION
`
`1 INTRODUCTION
`
`1.1 Machine Perception
`1.2 An Example
`1.3 The Classification Model
`1.4 The Descriptive Approach
`1.5 Summary of the Book by Chapters
`1.6 Bibliographical Remarks
`
`2 BAYES DECISION THEORY
`
`Introduction
`2.1
`2.2 Bayes Decision Theory--The Continuous Case
`2.3 Two-Category Classification
`2.4 Minimum-Error-Rate Classification
`2.5 Classifiers, Discriminant Functions and Decision Surfaces
`2.5.1 The Multicategory Case
`2.5.2 The Two-Category Case
`2.6 Error Probabilities and Integrals
`2.7 The Normal Density
`2.7.1 The Univariate Normal Density
`2.7.2 The Multivariate Normal Density
`2.8 Discriminant Functions for the Normal Density
`2.8.1 Case 1: ~’i = O21
`2.8.2 Case 2: ~"i---- ~
`2.8.3 Case 3: ~’i Arbitrary
`2.9 Bayesian Decision Theory--The Discrete Case
`2.10 Independent Binary Features
`2.11 Compound Bayes Decision Theory and Context
`
`1
`2
`4
`5
`6
`7
`
`10
`
`10
`13
`15
`16
`17
`17
`20
`20
`22
`22
`23
`24
`26
`27
`3O
`31
`32
`34
`
`xi
`
`

`

`Case 1:14-cv-02396-PGG-MHD Document 153-14 Filed 06/28/19 Page 6 of 21
`
`xii
`
`CONTENTS
`
`2.12 Remarks
`2.13 Bibliographical and Historical Remarks
`Problems
`
`3 PARAMETER ESTIMATION AND SUPERVISED
`LEARNING
`
`3.1 Parameter Estimation and Supervised Learning
`3.2 Maximum Likelihood Estimation
`3.2.1 The General Principle
`3.2.2 The Multivariate Normal Case: Unknown Mean
`3.2.3 The General Multivariate Normal Case
`3.3 The Bayes Classifier
`3.3.1 The Class-Conditional Densities
`3.3.2 The Parameter Distribution
`3.4 Learning the Mean of a Normal Density
`3.4.1 The Univariate Case: p(~ I~c)
`3.4.2 The Univariate Case: p(x IX)
`3.4.3 The Multivariate Case
`3.5 General Bayesian Learning
`3.6 Sufficient Statistics
`3.7 Sufficient Statistics and the Exponential Family
`3.8 Problems of Dimensionality
`3.8.1 An Unexpected Problem
`3.8.2 Estimating a Covariance Matrix
`3.8.3 The Capacity of a Separating Plane
`3.8.4 The Problem-Average Error Rate
`¯ 3.9 Estimating the Error Rate
`3.10 Bibliographical and Historical Remarks
`Problems
`
`4 NONPARAMETRIC TECHNIQUES
`
`Introduction
`4.1
`4.2 Density Estimation
`4.3 Parzen Windows
`4.3.1 General Discussion
`4.3.2 Convergence of the Mean
`4.3.3 Convergence of the Variance
`4.3.4 Two Examples
`4.4 k-Nearest Neighbor Estimation
`
`35
`36
`39
`
`44
`
`44
`45
`45
`47
`48
`49
`50
`51
`52
`52
`55
`55
`57
`59
`62
`66
`66
`67
`69
`70
`73
`76
`80
`
`85
`
`85
`85
`88
`88
`90
`91
`91
`95
`
`

`

`Case 1:14-cv-02396-PGG-MHD Document 153-14 Filed 06/28/19 Page 7 of 21
`
`CONTENTS
`
`4.5 Estimation of A Posteriori Probabilities
`4.6 The Nearest-Neighbor Rule
`4.6.1 General Considerations
`4.6.2 Convergence of the Nearest-Neighbor
`4.6.3 Error Rate for the Nearest-Neighbor Rule
`4.6.4 Error Bounds
`4.7 The k-Nearest-Neighbor Rule
`4.8 Approximations by Series Expansions
`4.9 Approximations for the Binary Case
`4.9.1 The Rademacher-Walsh Expansion
`4.9.2 The Bahadur-Lazarsfeld Expansion
`4.9.3 The Chow Expansion
`4.10 Fisher’s Linear Discriminant
`4.11 Multiple Discriminant Analysis
`4.12 Bibliographical and Historical Remarks
`Problems
`
`5
`
`LINEAR DISCRIMINANT FUNCTIONS
`
`5.1
`5.2
`
`Introduction
`Linear Discriminant Functions and Decision Surfaces
`5.2.1 The Two-Category Case
`5.2.2 The Multicategory Case
`5.3 Generalized Linear Discriminant Functions
`5.4 The Two-Category Linearly-Separable Case
`5.4.1 Geometry and Terminology
`5.4.2 Gradient Descent Procedures
`5.5 Minimizing the Perceptron Criterion Function
`5.5.1 The Perceptron Criterion Function
`5.5.2 Convergence Proof for Single-Sample Correction
`5.5.3 Some Direct Generalizations
`5.6 Relaxation Procedures
`5.6.1 The Descent Algorithm
`5.6.2 Convergence Proof
`5.7 Nonseparable Behavior
`5.8 Minimum Squared Error Procedures
`5.8.1 Minimum Squared Error and the Pseudoinverse
`5.8.2 Relation to Fisher’s Linear Discriminant
`5.8.3 Asymptotic Approximation to an Optimal Discriminant
`5.8.4 The Widrow-Hoff Procedure
`5.8.5 Stochastic Approximation Methods
`
`.°.
`XIII
`
`97
`98
`98
`99
`100
`101
`103
`105
`108
`108
`111
`113
`114
`118
`121
`126
`
`130
`
`130
`131
`131
`132
`134
`138
`138
`140
`141
`141
`142
`146
`147
`147
`148
`149
`151
`151
`152
`154
`155
`156
`
`

`

`Case 1:14-cv-02396-PGG-MHD Document 153-14 Filed 06/28/19 Page 8 of 21
`
`xiv
`
`CONTENTS
`
`5.9 The Ho-Kashyap Procedures
`5.9.1 The Descent Procedure
`5.9.2 Convergence Proof
`5.9.3 Nonseparable Behavior
`5.9.4 Some Related Procedures
`5.10 Linear Programming Procedures
`5.10.1 Linear Programming
`5.10.2 The Linearly Separable Case
`5.10.3 Minimizing the Perceptron Criterion Function
`5.10.4 Remarks
`5.11 The Method of Potential Functions
`5.12 Multicategory Generalizations
`5.12.1 Kesler’s Construction
`5.12.2 The Fixed-Increment Rule
`5.12.3 Generalization for MSE Procedures
`5.13 Bibliographical and Historical Remarks
`Problems
`
`6 UNSUPERVISED LEARNING AND CLUSTERING
`
`Introduction
`6.1
`6.2 Mixture Densities and Identifiability
`6.3 Maximum Likelihood Estimates
`6.4 Application to Normal Mixtures
`6.4.1 Case 1: Unknown Mean Vectors
`6.4.2 An Example
`6.4.3 Case 2: All Parameters Unknown
`6.4.4 A Simple Approximate Procedure
`6.5 Unsupervised Bayesian Learning
`6.5.1 The Bayes Classifier
`6.5.2 Learning the Parameter Vector
`6.5.3 An Example
`6.5.4 Decision-Directed Approximations
`6.6 Data Description and Clustering
`6.7 Similarity Measures
`6.8 Criterion Functions for Clustering
`6.8.1 The Sum-of-Squared-Error Criterion
`6.8.2 Related Minimum Variance Criteria
`6.8.3 Scattering Criteria
`6.8.3.1 The Scatter Matrices
`
`159
`159
`161
`163
`163
`166
`166
`167
`168
`169
`172
`174
`174
`176
`177
`179
`186
`
`189
`
`189
`190
`192
`193
`194
`195
`198
`201
`203
`203
`204
`207
`210
`211
`213
`217
`217
`219
`221
`221
`
`

`

`Case 1:14-cv-02396-PGG-MHD Document 153-14 Filed 06/28/19 Page 9 of 21
`
`CONTENTS
`
`xv
`
`6.8.3.2 The Trace Criterion
`6.8.3.3 The Determinant Criterion
`6.8.3.4
`Invariant Criteria
`Iterative Optimization
`6.9
`6.10 Hierarchical Clustering
`6.10.1 Definitions
`6.10.2 Agglomerative Hierarchical Clustering
`6.10.2.1 The Nearest-Neighbor Algorithm
`6.10.2.2 The Furthest-Neighbor Algorithm
`6.10.2.3 Compromises
`6.10.3 Stepwise-Optimal Hierarchical Clustering
`6.10.4 Hierarchical Clustering and Induced Metrics
`6.11 Graph Theoretic Methods
`6.12 The Problem of Validity
`6.13 Low-Dimensional Representations and Multidimensional Scaling
`6.14 Clustering and Dimensionality Reduction
`6.15 Bibliographical and Historical Remarks
`Problems
`
`Part II SCENE ANALYSIS
`
`7 REPRESENTATION AND INITIAL
`SIMPLIFICATIONS
`
`Introduction
`7.1
`7.2 Representations
`7.3 Spatial Differentiation
`7.4 Spatial Smoothing
`7.5 Template Matching
`7.5.1 Template Matching--Metric Interpretation
`7.5.2 Template Matching--Statistical Interpretation
`7.6 Region Analysis
`7.6.1 Basic Concepts
`7.6.2 Extensions
`7.7 Contour Following
`7.8 Bibliographical and Historical Remarks
`Problems
`
`222
`222
`223
`225
`228
`228
`230
`233
`233
`235
`235
`236
`237
`239
`243
`246
`248
`256
`
`263
`
`263
`264
`267
`272
`276
`276
`282
`284
`284
`288
`290
`293
`297
`
`

`

`Case 1:14-cv-02396-PGG-MHD Document 153-14 Filed 06/28/19 Page 10 of 21
`
`xvi
`
`CONTENTS
`
`8 THE SPATIAL FREQUENCY DOMAIN
`
`Introduction
`8.1
`8.2 The Sampling Theorem
`8.3 Template Matching and the Convolution Theorem
`8.4 Spatial Filtering
`8.5 Mean Square Estimation
`8.6 Bibliographical and Historical Remarks
`Problems
`
`9 DESCRIPTIONS OF LINE AND SHAPE
`
`Introduction
`9.1
`9.2 Line Description
`9.2.1 Minimum-Squared-Error Line Fitting
`9.2.2 Eigenvector Line Fitting
`9.2.3 Line Fitting by Clustering
`9.2.4 Line Segmentation
`9.2.5 Chain Encoding
`9.3 Shape Description
`9.3.1 Topological Properties
`9.3.2 Linear Properties
`9.3.3 Metric Properties
`9.3.4 Descriptions Based on Irregularities
`9.3.5 The Skeleton of a Figure
`9.3.6 Analytic Descriptions of Shape
`Integral Geometric Descriptions
`9.3.7
`9.4 Bibliographical and Historical Remarks
`Problems
`
`10 PERSPECTIVE TRANSFORMATIONS
`
`Introduction
`10.1
`10.2 Modelling Picture Taking
`10.3 The Perspective Transformation in Homogeneous Coordinates
`10.4 Perspective Transformations With Two Reference Frames
`Illustrative Applications
`10.5
`10.5.1 Camera Calibration
`10.5.2 Object Location
`10.5.3 Vertical Lines: Perspective Distortion
`10.5.4 Horizontal Lines and Vanishing Points
`
`298
`
`298
`302
`305
`308
`318
`322
`325
`
`327
`
`327
`328
`328
`332
`335
`337
`339
`341
`342
`345
`348
`352
`356
`362
`367
`372
`377
`
`379
`
`379
`380
`382
`386
`392
`392
`393
`394
`396
`
`

`

`Case 1:14-cv-02396-PGG-MHD Document 153-14 Filed 06/28/19 Page 11 of 21
`
`CONTENTS
`
`10.6 Stereoscopic Perception
`10.7 Bibliographical and Historical Remarks
`Problems
`
`11 PROJECTIVE INVARIANTS
`
`Introduction
`11.1
`11.2 The Cross Ratio
`11.3 Two-Dimensional Projective Coordinates
`11.4 The Inter-Lens Line
`11.5 An Orthogonal Projection Approximation
`11.6 Object Reconstruction
`11.7 Bibliographical and Historical Remarks
`Problems
`
`12 DESCRIPTIVE METHODS IN SCENE ANALYSIS
`
`Introduction
`12.1
`12.2 Descriptive Formalisms
`12.2.1 Syntactic Descriptions
`12.2.2 Relational Graphs
`12.3 Three-Dimensional Models
`12.4 The Analysis of Polyhedra
`12.4.1 Line Semantics
`12.4.2 Grouping Regions into Objects
`12.4.3 Monocular Determination of Three-Dimensional
`Structure
`12.5 Bibliographical and Historical Remarks
`Problems
`
`AUTHOR INDEX
`
`SUBJECT INDEX
`
`xvii
`
`398
`401
`404
`
`405
`
`405
`407
`411
`414
`418
`421
`422
`424
`
`425
`
`425
`426
`426
`434
`436
`441
`442
`449
`
`456
`462
`465
`
`467
`
`472
`
`

`

`Case 1:14-cv-02396-PGG-MHD Document 153-14 Filed 06/28/19 Page 12 of 21
`
`Part I
`
`PATTERN
`CLASSIFICATION
`
`

`

`Case 1:14-cv-02396-PGG-MHD Document 153-14 Filed 06/28/19 Page 13 of 21
`
`Chapter 1
`
`INTRODUCTION
`
`1.1 MACHINE PERCEPTION
`
`Since the advent of the digital computer there has been a constant effort to
`expand the domain of computer applications. Some of the motivation for
`this effort comes from important practical needs to find more efficient ways
`of doing things. Some of the motivation comes from the sheer challenge of
`building or programming a machine to do things that machines have never
`done before. Both of these motives are found in that area of artificial
`intelligence that we shall call machine perception.
`At present, the ability of machines to perceive their environment is very
`limited. A variety of transducers are available for converting light, sound,
`temperature, etc., to electrical signals. When the environment is carefully
`controlled and the signals have a simple interpretation, as is the case with
`the standard computer input devices, the perceptual problems become
`trivial. But as we move beyond having a computer read punched cards or
`magnetic tapes to having it read hand-printed characters or analyze bio-
`medical photographs, we move from problems of sensing the data to much
`more difficult problems of interpreting the data.
`The apparent ease with which vertebrates and even insects perform per-
`ceptual tasks is at once encouraging and frustrating. Psychological and
`physiological studies have given us a great many interesting facts about
`animal perception, but no understanding sufficient for us to duplicate their
`performance with a computer. The problem area has a certain unique
`fascination to it because perception is something everyone experiences but
`no one really understands. Introspection has not proved as helpful in dis-
`covering the nature of perception as one might hope, apparently because most
`everyday perceptual processes are carried out below the conscious level.
`Paradoxically, we are all expert at perception, but none of us knows much
`about it.
`The lack of a complete theory of perception has not prevented people
`from trying to solve more modest problems. Many of these involve pattern
`
`

`

`Case 1:14-cv-02396-PGG-MHD Document 153-14 Filed 06/28/19 Page 14 of 21
`
`2
`
`INTRODUCTION
`
`classification--the assignment of a physical object or event to one of several
`prespecified categories. Extensive study of classification problems has led
`to an abstract mathematical model that provides the theoretical basis for
`classifier design. Of course, in any specific application one ultimately must
`come to grips with the special characteristics of the problem at hand. Of the
`various problem areas, the domain of pictorial problems has received by
`far the most attention. The purpose of this book is to give a systematic
`account of those principles of pattern classification and those techniques of
`pictorial scene analysis that seem to have the widest applicability and interest.
`
`1.2 AN EXAMPLE
`
`To illustrate some of the types of problems we shall address, let us consider
`the following imaginary and somewhat fanciful example. Suppose that a
`lumber mill producing assorted hardwoods wants to automate the process of
`sorting finished lumber according to species of tree. As a pilot project, it is
`decided to try first to distinguish birch lumber from ash lumber using optical
`sensing. A system to perform this very specific task might well have the form
`shown in Figure 1.1. The camera takes a picture of the lumber and passes
`the picture on to a feature extractor, whose purpose is to reduce the data by
`measuring certain "features" or "properties" that distinguish pictures of
`birch lumber from pictures of ash lumber. These features (or, more precisely,
`the values of these features) are then passed to a classifier that evaluates
`the evidence presented and makes a final decision about the lumber type.
`Let us consider how the feature extractor and classifier might be designed.
`Suppose somebody at the lumber mill tells us that birch is often lighter
`colored than ash. Then brightness becomes an obvious feature, and we might
`attempt to classify the lumber merely by seeing whether or not the average
`brightness x exceeds some critical value x0. To choose x0, we could obtain
`some samples of the different types of wood, make brightness measurements,
`and inspect the results. Suppose that we do this, and obtain the histograms
`
`TRANSDUCER
`
`FEATURE L------J
`EXTRACTOR i
`r’~"~t CLASSIFIER ~ DECLSION
`
`FIGURE 1.1. A pattern classification system.
`
`

`

`Case 1:14-cv-02396-PGG-MHD Document 153-14 Filed 06/28/19 Page 15 of 21
`
`AN EXAMPLE ’ 3
`
`BIRCH
`
`Xo
`
`BRIGHTNESS --’ X
`
`FIGURE 1.2. Histograms for the brightness feature.
`
`shown in Figure 1.2. These histograms bear out the statement that birch is
`usually lighter than ash, but it is clear that this single criterion is not in-
`fallible. No matter how we choose x0, we can not reliably separate birch
`from ash by brightness alone.
`In our search for other features, we might try to capitalize on the observa-
`tion that ash typically has a more prominent grain pattern than birch. This
`feature is much more difficult to measure than average brightness, but it is
`reasonable to assume that we can obtain a measure of grain prominence from
`the magnitude and frequency of occurrence of light-to-dark transitions in
`the picture. Now we have two features for classifying lumber, the brightness
`x1 and the grain prominence x~. The feature extractor has thus reduced each
`picture to a point or feature vector x in a two-dimensional feature space,
`where
`
`X2
`
`Our problem now is to partition the feature space into two regions, where
`all the points in one region correspond to birch, and all points in the other
`correspond to ash. Suppose that we measure the feature vectors for our
`samples and obtain the scattering of points shown in Figure 1.3. This plot
`suggests the following rule for classifying the data: Classify the lumber as
`ash if its feature vector falls above the line AB, and as birch otherwise.
`While this rule appears to do a good job of separating our samples, we
`have no guarantee that it Will perform as well on new samples. It would
`certainly be prudent to obtain some more samples and see how many are
`
`

`

`Case 1:14-cv-02396-PGG-MHD Document 153-14 Filed 06/28/19 Page 16 of 21
`
`4
`
`INTRODUCTION
`
`¯ ASH
`
`O BIRCH
`
`o
`Oo o
`
`0000 0
`0
`
`mUm|
`
`o
`
`oB
`
`BRIGHTNESS -- Xi
`
`FIGURE 1.3. Scatter diagram for
`the feature vectors.
`
`correctly classified. This suggests that our problem has a statistical com-
`ponent, and that perhaps we should look for a Classification procedure that
`minimizes the probability of error.
`Without forgetting this idea, we should remember that we chose a simple
`problem for a pilot project. A more realistic problem might involve sorting
`many different classes of lumber. To separate oak from birch and ash, we
`might well require less obvious features, such as "straightness-of-grain."
`With more categories and more features, the graphical approach to designing
`the classifier will probably have to be abandoned. To proceed further, we
`shall need whatever theoretical help we can get.
`
`1.3 THE CLASSIFICATION MODEL
`
`The preceding example contains many of the elements of the most commonly
`used abstract model for pattern recognition, the classification model. This
`model contains three parts: a transducer, a feature extractor, and a classifier
`The transducer senses the input and converts it into a form suitable for
`machine processing. The feature extractor (also called the. receptor, property
`filter, attribute detector, or preprocessor) extracts presumably relevant
`information from the input data. The classifier uses this information to
`assign the input data to one of a finite number of categories.
`A discussion of transducer specification or design lies outside the province
`of this book. However, we shall be concerned with both feature extraction
`and classification. From a theoretical viewpoint, the line between these topics
`is arbitrary. An ideal feature extractor would make the job of the classifier
`trivial, and an omnipotent classifier would not need the help of a feature
`extractor. The distinction is forced upon us for practical, not theoretical
`reasons, but the distinction is important nevertheless.
`
`

`

`Case 1:14-cv-02396-PGG-MHD Document 153-14 Filed 06/28/19 Page 17 of 21
`
`THE DESCRIPTIVE APPROACH 5
`
`Generally speaking, the problem of feature extraction is much more
`problem dependent than the problem of classification. A good feature
`extractor for sorting lumber would probably be of little use for identifying
`fingerprints or classifying photomicrographs of blood cells. Nevertheless, a
`substantial body of techniques has been developed for extracting useful
`information from pictures. Part II of this book is devoted to an exposition
`of these techniques and their properties.
`The problem of classification is basically one of partitioning the feature
`space into regions, one region for each category. Ideally, one would like to
`arrange this partitioning so that none of the decisions is ever wrong. When
`this cannot be done, one would like to minimize the probability of error, or,
`if some errors are more costly than others, the average cost of errors. In
`this case, the problem of classification becomes a problem in statistical
`decision theory, a subject that has many applications to pattern classification.
`Part I of this book is concerned with these and related topics in classification
`theory.
`
`1.4 THE DESCRIPTIVE APPROACH
`
`There are many problems in pattern recognition and machine perception for
`which the classification model is clearly inappropriate. For example, in
`analyzing a picture of bubble-chamber particle tracks, one wants a descrip-
`tion, rather than just a classification, of the picture, Such a description should
`contain information about both the individual parts of the picture and about
`the relations among the parts. Ideally, it should directly reflect the structure
`present in the original scene.
`Consider, for example, the simple scene shown in Figure 1.4. It is possible
`that a simple classification such as "office scene" or "telephone present"
`would be adequate for some purposes. A more complete analysis would
`include an identification of all the major objects present--the telephone,
`note pad, cup, pencils, eraser, etc. An even more complete analysis would
`indicate the relations between these objects, and might result in a description
`such as "(Two pencils on top of a pad) in front of (a cup to the left of (an
`eraser in front of a telephone))."
`The problem of analyzing a visual scene and producing a structural de-
`scription has proved to be quite difficult. A generally accepted formalization
`of the problem, analogous to the classification model, has yet to emerge
`from the work that has been done to date. Attempts have been made to
`borrow concepts from the theory of formal languages and to produce a
`linguistic model for scene analysis. Here the scene is viewed as a statement
`in a language whose grammar defines the allowed structural relations. With
`
`

`

`Case 1:14-cv-02396-PGG-MHD Document 153-14 Filed 06/28/19 Page 18 of 21
`
`6
`
`INTRODUCTION
`
`FIGURE 1.4. A simple scene.
`
`this formulation, scene analysis is viewed as the process of using a picture
`grammar to parse the scene, producing a description of the scene as a com-
`position of related subscenes.
`The linguistic model does not exhaust the ways of using known structural
`relations among the elements of a picture to guide its analysis and produce
`a useful description. Some of the most interesting procedures that have been
`developed are ad hoc and heuristic; a unifying conceptual framework en-
`compassing all of these methods has not yet been developed. The procedures
`themselves can be described, however, and this general topic of descriptive
`approaches to scene analysis concludes Part II.
`
`1.5 SUMMARY OF THE BOOK BY CHAPTERS
`
`The orientation of Part I of this book is primarily statistical. Chapter 2 states
`the classification problem in decision-theoretic terms, and derives the general
`form for an optimal classifier. This solution is obtained under the assumption
`that all of the probability distributions involved are known. The remainder
`
`

`

`Case 1:14-cv-02396-PGG-MHD Document 153-14 Filed 06/28/19 Page 19 of 21
`
`BIBLIOGRAPHICAL REMARKS 7
`
`of Part I is concerned with ways of proceeding when the probabilistic structure
`is not known.
`In Chapter 3 we assume that everything is known except for some param-
`eters of the distributions. Both maximum likelihood and Bayesian pro-
`cedures for estimating parameters from samples are described. If none of the
`standard ways of parameterizing unknown distributions is suitable, one can
`resort to nonparametric techniques. These procedures, which are discussed
`in Chapter 4, exchange the need for knowledge of the forms of the distri-
`butions for the need for a large number of samples.
`In Chapter 5 we parameterize the classifier and study ways of using samples
`to determine the classifier directly. These techniques have their origin in
`Fisher’s linear discriminant, and include the well known perceptron and
`relaxation procedures, minimum-squared-error methods, stochastic approxi-
`mation, the method of potential functions, and linear programming tech-
`niques. Chapter 6 concludes Part I with a discussion of various techniques
`for unsupervised learning and clustering.
`Chapter 7 begins Part II with a discussion of procedures for representing
`pictures, and for performing such basic operations as sharpening, smoothing,
`template matching, and partitioning a picture into homogeneous regions.
`Chapter 8 develops the concept of spatial filtering, and interprets some of
`these operations in the frequency domain. Chapter 9 is concerned with a
`great variety of procedures for describing lines and shapes in pictures.
`Topological, linear, and metric properties of shape are described, together
`with a number of descriptive techniques based on these properties.
`Chapters 10 and 11 present important mathematical background relevant
`to pictures of three-dimensional objects. Chapter 10 develops the equations
`for perspective transformation, and shows how they can be usefully employed
`in scene analysis. Chapter 11 treats the subject of projective invariants,
`quantities that are the same in different pictures of the same object. Finally,
`Chapter 12 discusses some of the more important contemporary approaches
`to the difficult problem of completely analyzing visual scenes.
`
`1.6 BIBLIOGRAPHICAL REMARKS
`
`The published literature on pattern classification and scene analysis has grown
`to the point where even specialized bibliographies can contain hundreds of
`references. The references that we give at the end of each chapter hopefully
`will provide the reader with some historical perspective and a good starting
`point for further study. Additional guidance can be obtained from other
`texts and from a number of valuable survey articles.
`
`

`

`Case 1:14-cv-02396-PGG-MHD Document 153-14 Filed 06/28/19 Page 20 of 21
`
`8
`
`INTRODUCTION
`
`For a brief overview that treats pattern recognition as one topic in arti-
`ficial intelligence, the influential article by Minsky (1961) is highly recom-
`mended. Nagy (1968) provides an excellent survey of work done using the
`classification model. The text by Nilsson (1965) provides an exceptionally
`clear treatment of classification procedures. A lucid and more recent survey
`of these procedures is given by Ho and Agrawala (1968). Levine (1969) gives
`a comprehensive survey of the techniques that have been used to extract
`features from pictures. The general subject of automatic picture processing
`is well surveyed by Hawkins (1970), and is systematically treated in the
`scholarly text by Rosenfeld (1969).
`There are many interesting subject areas that are related to this book but
`beyond its scope. Readers interested in image enhancement and picture
`coding should be aware of the survey by Huang, Schreiber and Tretiak
`(1971). Those interested in the fascinating area of human and animal per-
`ception will find the surveys by Kolers (1968) and Gose (1969) most useful.
`Those interested in philosophical issues will find the books by Watanabe
`(1969) and Bongard (1970)thought provoking. Finally, those interested in
`the practical applications of all of this theory will find a wealth of references
`in the literature survey by Stevens (1970).
`
`REFERENCES
`
`1. Bongard, M., Pattern Recognition (Spartan Books, Washington, D.C.,
`1970).
`2. Gose, E. E., "Introduction to biological and mechanical pattern recognition,"
`in Methodologies of Pattern Recognition, pp. 203-252, S. Watanabe, ed. (Academic
`Press, New York, 1969).
`3. Hawkins, J. K., "Image processing principles and techniques," in Advances in
`Information Systems, Vol. 3, pp. 113-214, J. T. Tou, ed. (Plenum Press, New York
`and London, 1970).
`4. Ho, Y. C. and A. Agrawala, "On pattern classification algorithms: introduc-
`tion and survey," Proe. IEEE, 56, 2101-2114 (December 19"68).
`
`5. Huang, T. S., W. F. Schreiber, and O. J. Tretiak, "Image Processing," Proe.
`IEEE, 59, 1586-1609 (November 1971).
`6. Kolers, P. A., "Some psychological aspects of pattern recognition," in
`Recognizing Patterns, pp. 4-61, P. A. Kolers and M. Eden, eds. (MIT Press,
`Cambridge, Massachusetts, 1968).
`7. Levine, M.D., "Feature extraction: a survey," Proc. IEEE, 57, 1391-1407
`(August 1969).
`8. Minsky, M., "Steps toward artificial intelligence," Proe. IRE, 49, 8-30
`(January 1961); also in Computers and Thought, pp. 406--450, E. A. Feigenbaum
`and J. Feldman, eds. (McGraw-Hill, New York, 1963).
`
`

`

`Case 1:14-cv-02396-PGG-MHD Document 153-14 Filed 06/28/19 Page 21 of 21
`
`REFERENCES "9
`
`9. Nagy, G., "State of the art in pattern recognition," Proc. IEEE, 56, 836-862
`(May 1968).
`10. Nilsson, N. J., Learning Machines (McGraw-Hill, New York, 1965).
`11. Rosenfeld, A., Picture Processing by Computer (Academic Press, New
`York, 1969).
`12. Stevens, M. E.,"Research and development in the computer and information
`sciences. Volume 1. Information acquisition, sensing, and input--a selective
`literature review," National Bureau of Standards Monograph 113, Vol. 1 (March
`1970).
`13. Watanabe, M. S., Knowing and Guessing (John Wiley, New York, 1969).
`
`

This document is available on Docket Alarm but you must sign up to view it.


Or .

Accessing this document will incur an additional charge of $.

After purchase, you can access this document again without charge.

Accept $ Charge
throbber

Still Working On It

This document is taking longer than usual to download. This can happen if we need to contact the court directly to obtain the document and their servers are running slowly.

Give it another minute or two to complete, and then try the refresh button.

throbber

A few More Minutes ... Still Working

It can take up to 5 minutes for us to download a document if the court servers are running slowly.

Thank you for your continued patience.

This document could not be displayed.

We could not find this document within its docket. Please go back to the docket page and check the link. If that does not work, go back to the docket and refresh it to pull the newest information.

Your account does not support viewing this document.

You need a Paid Account to view this document. Click here to change your account type.

Your account does not support viewing this document.

Set your membership status to view this document.

With a Docket Alarm membership, you'll get a whole lot more, including:

  • Up-to-date information for this case.
  • Email alerts whenever there is an update.
  • Full text search for other cases.
  • Get email alerts whenever a new case matches your search.

Become a Member

One Moment Please

The filing “” is large (MB) and is being downloaded.

Please refresh this page in a few minutes to see if the filing has been downloaded. The filing will also be emailed to you when the download completes.

Your document is on its way!

If you do not receive the document in five minutes, contact support at support@docketalarm.com.

Sealed Document

We are unable to display this document, it may be under a court ordered seal.

If you have proper credentials to access the file, you may proceed directly to the court's system using your government issued username and password.


Access Government Site

We are redirecting you
to a mobile optimized page.





Document Unreadable or Corrupt

Refresh this Document
Go to the Docket

We are unable to display this document.

Refresh this Document
Go to the Docket