throbber
Word Sense Disambiguation for Free-text. Indexing
`Using a Massive Semantic Network
`
`Michael Sussna
`Department of Computer Science and Engineering
`University of California, San Diego
`La JoUa, California 92093-0114
`
`8ussna®c.s.ncsd.edu
`
`Abstract
`
`Setnantics-free, woid-based information retrieval is thwarted
`by two complementary problems. First, search for relevant
`documents returns irrelevant items when all meanings of a
`search term are used, rather than just the meaning intended.
`This causes low precision. Second, relevant items are missed
`when they are indexed not under the actual search terms,
`but rather under related terms. This causes low recall. With
`semantics-free approaches there is generally no way to im
`prove both precision and recall at the same time.
`Word sense disambiguation during document indexing
`should improve precision. We have investigated using the
`massive WordNct semantic network for disambiguation dur
`ing indexing. With the unconstrained text of the SMART
`retrieval environment, we have had to derive oiir own con
`tent description from the input text, given only part-of-
`speech tagging of the input.
`We employ the notion of semantic distance between net
`work nodes. Input text terms with multiple senses are dis-
`ambignated by finding the combination of stMsta from a set
`of contiguous terms which minimizes total pairwise distance
`between senses. Results so far have been encouraging. Im
`provement in disambiguation compared with :hance is clear
`and consistent.
`
`Keywords: Information retrieval, indexing, word sense
`disambiguation, semantic networks, free-text.
`
`1 Introduction
`
`Semantics-free, word-based information retrieval is thwarted
`by two complementary problems. First, search for relevant
`documents returns irrelevant items when all meanings of a
`search term are used, rather than just the meaning intended.
`This is the polysemy/false positives/low precision problem.
`Second, relevant items ate missed when they ate indexed
`not under the actual search terms, but rather under related
`terms. This is the synonymy/false negatives/low recall prob
`lem. With semantics-free approaclies there is generally no
`way to improve both precision and recall at the same lime.
`
`Piimitsfon to copy without f«t all or part of this matarlal ia
`grantad providad that lha eoplat ara not mado or ditlfibutad for
`diroel eommorciol advanltga, tha ACM eopyright notica and tha
`tilJo of tho puUioation and Ita data appaar, and notica ia givan
`that copying ia by parmiaaion ot lha Ataociaiion for Computing
`Machinery. To copy othorwiaa, or to ropubiiah, raquiros a faa
`and/or apecific paimistion.
`CIKM '93 • 1 t/93/O.C.. USA

` 1993 ACM 0-89791-626.3/93/0011
`
`50
`
`67
`
`Increasing one is done at the expense of the other [Salton
`and McGil], 1983; van Rijsbergen, 1983], For example, cast
`ing a wider net of search terms to improve recall of relevant
`items will also bring in an even greater proportion of irrele
`vant items, lowering precision.
`There is a many-to-many mapping between word forms
`and word meanings. A single word form can have multiple
`meanings, and a single meaning can be expressed by multi
`ple word forms. Both of these muHiplicities cause problems
`for any approach to content search based on word forms.
`We believe that in order to do near-human level retrieval we
`must go beyond words and get at meanings. Text disam
`biguation during indexing should improve precision by com
`bating polysemy (Krovetz and Croft, 1992]. We are looking
`into reducing the ambiguity of word forms during index
`ing by taking advantage of semantic networks. A number
`of these networks already exist and their implementation is
`fairly straightforward.
`As part of a larger research project exploring the ex
`ploitation of explicit semantics for overcoming both the pol
`ysemy and synonymy problems, we have performed prelim
`inary investigations of document indexing using a massive
`semantic network, WordNet. WordNet is a network of word
`meanings connected by a variety of lexical and semantic
`relations. Over 35,000 word senses are represented in the
`noun portion of WordNet alone. We have been working with
`WotdNci in the SMART information retrieval environment.
`In the unconstrained text of the SMART environment, no
`index terms have been assigned [Buckley, 198S]. We have
`had to derive our own content description from the input
`text, given only part-of-speech tagging of the input.
`Employing the notion of semantic distance between net
`work nodes, we have run a series of experiments. Input text
`terms with multiple senses have been disambiguated by find
`ing the combination of senses from a set of contiguous terms
`which minimizes total pairwise distant^ between senses. Re
`sults so far have been encouraging. Improvement in dis
`ambiguation compared with chance is clear and consistent,
`strongly suggesting that semantics-based indexing is worth
`pursuing further for transcending the polysemy problem. It
`U competitive with word-based approaches. A number of
`these have focused on only a few fixed terms whose senses
`were to be distinguished, rather than on unconstrained text
`[Lesk, 1986; Wilks et aL, 1969; Voorhees et al., 1993].
`In the following sections wc will discuss the research en
`vironment, network-based disambiguation, the experiments
`performed and results obtained.
`
`Page 1 of 8
`
`GOOGLE EXHIBIT 1033
`
`

`

`2 Research Environment
`
`Nouns in WordNet;
`
`The current project uses the SMART information retrieval
`environment. In SMART, documents do not have keyword
`descriptors. Instead, one must do one's own indexing of
`content. We are investigating using semantics for word sense
`disambiguation during document indexing.
`Semantics are supplied by the WordNet lexical/semantic
`database developed at the Cognitive Science Laboratory at
`Princeton University {Miller et at., 1930; Miller, 1990]. Of
`particular relevance and usefulness for our research is the
`noun portion of WordNet, which contains over 35,000 word
`meanings represented as network nodes called "synscts" (syn
`onym sets). Each sense of a word maps to a distinct synset.
`For example, one sense of the noun "strike" maps to (hit rap
`strike tap) which IS-A (impact bump thump blow); another
`maps to (strike work^toppage) which IS-A {direct.action).
`We work with the Time Magazine article collection, since
`it is the least specialized and technical, because WordNet is
`a general English lexicon.
`With SMART, the words in the documents are converted
`to lower case and parsed into strings. They can be stemmed
`down to base forms; e.g., "stemmed" and "stems" both be
`come "stem." Input words can also be labeled by part of
`speech, which is a feature that we took advantage of. Al
`though the part-of-speech tagger employed was not infalli
`ble, it was accurate enough to ^ve us a good working set of
`nouns to serve as input to semantic processing.
`One aspect of this input editing process which is a source
`for limiting the effectiveness of our efforts is the tillering
`out of terms. SMART uses a list of "stopwords," words
`to be ignored as "contentless." For example, prepositions,
`conjunctions, and articles are considered extraneous. After
`stopwords have been removed, and non-nouns removed from
`what remains, very little of the original article is left. So, we
`are working with a sparse sample of the original text by the
`time we get to decide which sense of each noun is intended.
`Nouns found in WordNet are the final distillation' that we
`begin to work with during disambiguation.
`The following example illustrates the fUtering process. It
`uses an excerpt &om Time document ], shown after succes
`sive filtering steps.
`
`After conversion to lowercase (part^of-speech tagging is
`omitted for readability; the first four words are actually the
`title):
`
`the allies after nassau in december 1960, the u.8 . first
`proposed to help nato develop its own nuclear strike force .
`but europe made no attempt to devise a plan . last week, as
`they studied the nassau accord between president kennedy
`and prime minister macmillan, europeans saw emerging the
`first outlines of the nuclear nato that the u.s . wants and
`will support . it all sprang from the anglo-u.s . ctisu over
`cancellation of the bug-ridden skybolt missile, and the u.s
`. offer to supply britain and france with the proved polaris
`(time, dec . 26).
`
`After stopword removal;
`
`allies . proposed nato develop nuclear strike force made
`attempt devise plan . week studi^ accord president kennedy
`prime minister macmillan emerging outlines nuclear nato .
`support sprang anglo crisis cancellation bug ridden skybolt
`missile offer supply britain france proved polaris time dec
`
`allies strike force attempt plan week accord president
`prime minister outlines support crisis cancellation bug mis
`sile fra;nce polaris time
`
`WordNet's noun portion has fairly rich connectivity as
`well as obvious comprehensiveness. The WordNet noun nodes
`are connected by nine relations. Bight of these form four
`pairs of complementary or inverse relations, while one is its
`own inverse. There is actually a tenth relation that is im
`plicit in the network structure, but does not label any net
`edges because it is intranode rather than internode. The
`relations are;
`
`aynonyny
`hjpemyay
`byponyay
`holonyay
`
`neronyny
`
`antonyay
`
`(haa sane neaning aa; intranode)
`(is a)
`(has instance)
`(is part of, is substance in,
`is nenber of; 3 relations)
`(has part, contains substance,
`has oenber; 3 relations)
`(is conplenent of; self-inverse)
`
`Hypernymy and hyponymy are the strictly hierarchical links.
`The holonymy/meronymy relations can also be considered
`"vertical" relations. Vertical relations are asymmetrical and
`order items. Synonymy and antonymy are "horizontal,"
`symmetrical, non-ordering relations (and of course are non-
`hierarchical).
`
`3 Net-based Disambiguation
`
`We have tried a variety of approaches to term disambigua
`tion, all based on minimizing an objective function utilizmg
`semantic distance between topics in WordNet. It is outside
`the scope of this paper to explain the distance determination
`logic. We will, however, describe the salient aspects of the
`network edge weighting scheme because this badcgronnd is
`necessary for discussion of the experiments where the net
`work weights were varied.
`
`3.1 Edge weighting
`
`Each edge consists of two inverse relations. Each relation
`type has a weight range between its own min and mac. The
`point in the range for a particular arc depends on the number
`of arcs of the same type leaving the node. This is the type-
`specific fanout (TSF) factor. TSF reflects dilution of the
`strength of connotation between a source and target node
`as a function of the number of like relations that the source
`node has.' The two inverse wdghts for an edge are averaged.
`The average is divided by the- depth of the edge within the'
`overall "tree." This process is called depth-relative seating
`and it is based on the observation that only-siblings deep in
`a tree are mote closely related than only-siblings higher in
`the tree.
`
`Dcllmtion. 1
`The edge between adjacent nodes A and B has distance
`or weight
`
`^ThU factor taker into account tho pocaibic aa/mmetry between
`two nodes, where the strength of connotation in one dinction dilTers
`from that in the other direction [Tvenky, 1977].
`
`68
`
`Page 2 of 8
`
`

`

`B) + w[B~*r' >1)
`w(A,B) =
`^
` ' u
`
`given w{X —Y) = maxr — mazr — mi'nr
`nr(.V)
`is its inverse, d is the
`U a relation of type r,
`where
`depth of the deeper of the two nodes, maxr and miiir are
`the maximum and minimum weights possible for a relation
`of type r respectively, and np(Jf) is the number of relations
`of type T leaving node X. □
`The synonym relation gets a weight of zero, while the
`nine intemode relation types have preliminary weight ranges
`as follows: hypcrnymy, hyponymy, holonymy, and meronymy
`all have wdghts ranging from 1 to 2. Antonymy arcs all get
`the value 2.5 (there is no range).
`
`3,2 Total distance minimization
`
`We utilize semantic distance between network nodes, cap
`tured by the weights on the edges along the shortest path
`connecting the nodes, as a measure of relatedness between
`the topics represented by the nodes. The shorter the dis
`tance, the greater the relatedness. For disambiguation the
`hypothesb is that, given a set of terms occurring near each
`other in the text, each of which might have multiple mean
`ings, by picking the senses that minimize distance we select
`the correct senses.
`Overall distance minimization works as follows. Por->a
`given set of terms T =
`l„}, each with possibly
`more than one candidate sense, each combination of n senses
`across the terms is tried, with one sense chosen at a time
`for each term. For example, given three terms ti.fj.ts, with
`2, 1, and 3 senses respectively, each of the 6 = 2 -'1 • 3
`combinations of senses is tried. For each combination of n
`senses, the pairwise distances between each pair of senses is
`found. The
`pairwise distances are summed to arrive
`at an overall vuue, H{T). The combination of senses which
`minimizes this sum is the "winning" combination.
`Definition 2
`(«}, let
`For a set of n^hboting terms T = {ti.fj
`5 be the set of all combinations of term senses, which has
`cardinality nLi l'«l> where |fj| is the number of senses of
`term t, and let o € ^ be a particular combination of senses
`(si,53,...,Sn}, where each Sj is a sense of tj.
`■ The winning combination is the S & S which produces
`the minimal "energy"
`
`ffiniii(T) = imn
`
`dtstance(z,y)
`
`Vi,y € S.D'
`
`We call this tedinique mutuaf conslruint among terms.
`There is a special case of mutual constraint where all terms
`except the one being disambiguated have had their senses
`determined and "frozen." Thus they have only one sense to
`work with now. When we are trying to disambiguate a term
`and work with previous frozen terms only, we speak of using
`a frozen pazt approach.
`
`<f«#lane«(y yjr)
`di»ianet{at,g) = 0.
`
`We have experimented with pure mutual constraint, pure
`frozen past, and a combination of the two. In ail cases there
`is a moptng window of terms currently in focus as wc move
`from the beginning of a document towards its end. In the
`pure cases there is only a moving window. In the case where
`mutual constraint and frozen past, a small set
`of initial text terms is processed with mutual constraint.
`This sets up a bias in semantic space for the processing of
`subsequent terms. The later terms are then processed with
`a moving frozen past window.
`Mutual constraint is more appealing conceptually than
`frozen past but is exponential in the number of combinations
`of term sen^s that need to be tried. Frozen past avoids this
`combinatoric explosion by reducing the problem to essen
`tially Unear-time processing, since there are only as many
`"combinations" to try as there are senses of the single term
`being disambiguated.
`Which tetm(s) gets its winning sense assigned varies de
`pending on the type of window used. When working with a
`frozen past window of size n, only the (n -f l)st term is as
`signed its sense. Each of the n window terms has already had
`its sense frozen. When working with a moving mutual con
`straint window, just the middle term is assigned its sense.
`Record is kept of the winning sense, but when that term
`plays a role other than "middle term," its senses are allowed
`to fully vary. This gives a middle term full benefit of both
`previous and subsequent context. All senses of surrounding
`terms are considered, not just their winning senses. For ini
`tial (as opposed to moving) mutual constraint windows, all
`of the terms in the window ate assigned their senses at the
`same time.
`
`4 ExpeilmeiitB
`We have performed a number of disambiguation experiments
`with the Time collection. One series of experiments varied
`window size and type, and a second series varied network
`weighting schemes. Before discussing our experimental re
`sults, we need to «»ver the subject of measuring performance
`during disambigaation.
`
`4.1 Performance evaluation
`How do we measure success in disambiguation? We need
`to know what the "right" answer is for each term being dis
`ambiguated. This knowledge is provided by manual analysis
`and disambiguation of the terms. Because this is tedious and
`problematic work, we originally only hand-disambiguated
`the first five Time documents.
`During that process it became evident that there are a
`number of situations that can arise when considering the
`input to the disambiguator. Seven situations can be distin
`guished:
`^
`1. There are multiple "good" senses — more than one sense
`of the input term is applicable in the context in which
`the term appears.
`2. There is exactly one good sense.
`3. There axe no applicable senses. This has five variations;
`3a. The item is not actually a noun here (e.g. "prime"
`in "prime minister")
`3b. The item is a noun, but not the one the program
`sees (e.g. "cent" from "per cent")
`
`69
`
`Page 3 of 8
`
`

`

`3c. The item was found as is, instead of after being
`stemmed ("acres* meaning "estate, demesne* in
`stead of the plural of "acre")
`3d. The item is really a proper noun ("time" as in
`Time Magazine)
`3e. The item is used in a sense not found in WordNct
`("time* as in "at that time")
`
`167 pose, trivial hits (good but no bad senses)
`319 poss. nontrivial hits (good and bad senses)
`
`749.9 naxiaun hit points
`
`As a baseline for comparison, senses were chosen ran
`domly. This "chance" performance yielded expected values
`as follows;
`
`We take these sitnations into account in deriving our
`measure of success or failure in disambiguation. Although
`the disambiguator in general works with every word that it
`is presented with, we focus only on those terms which have
`at least one good sense. In addition, we distinguish between
`"trivial" success and "nontrivial" success — words with at
`least one good sense but with no bad senses are trivial to
`disambiguate, since any choice is a success. Only when at
`least one sense is good and at least one is bad can we consider
`picking a correct sense a success worth rewarding. Thus we
`focus on nontrivial terms — those which are true tests of
`disambiguation prowess.
`One obvious way of evaluating success is to find the per
`centage of terms correctly disambiguated (out of the non-
`trivial terms). We will use this "hit-or-miss" measure as
`a secondary indicator. Since it does not reflect the diffi
`culty present for individual terms, we have chosen to focus
`on another measure that takes this difficulty into account.
`This is the "hit score" — the ratio of "actual hit points* to
`"maximum hit paints." Hit points are awarded as follows.
`Deflnltion 3
`
`For each term let a be the number of senses and let g be
`the number of good senses (in context). The hit points for
`a hit-are sfg — 1. Misses gel zero points. □
`The actual hit points for individual terms are summed,
`and this sum is divided by the sum of the maximum number
`of hit paints possible, derived by treating all nontrivial terms
`as having been disambiguated correctly and their hit points
`awarded accordingly. Formally, hit score over n terms equals
`hiipoints, where
`term,
`is a hit
`Ailpoiiifsi
`
`Hit scores range from 0 to 1.
`After manual disambiguation, the first five Time doc
`uments served as a standard against which to measure the
`performance of the semantic distance software. During man
`ual disambiguation, the several situations that can arise for
`a term whi^ were outlined above were taken into account
`when classifying the terms. The large majority of the terms
`had at least one good sense. Some basic quantities for the
`five documents are;
`
`1176 terae cenalning after stopeord renoval
`544 of thoBG are nouns and in HordNet
`123 type 1 terns (nultiple good senses)
`364 type 2 terns (one good sense)
`58 type 3 terns (no good senses) as lellovs:
`18 type 3a (not really a noun)
`4 type 3b (vxong noun)
`6 type 3c (unstenued, taken as is)
`7 type 3d (proper n'onn)
`23 type 3e (sense not in VordHet)
`
`486 possible hits (at least 1 good sense)
`
`nontrivial hits:
`X correct of nontrivial:
`hit points;
`hit score:
`
`124.6
`.391 (124.6 / 319)
`194.4
`.259
`
`(194.4 / 749.9)
`
`The standard deviation of the distribution of hit scores
`obtained from multiple runs of the "chance* software is ap
`proximately 0.04. In other words, taking a ± two standa^
`deviation range, the "chance* software will give a hit score
`in the range 0.259 d: 0.08 with high probability. Thus if an
`alternative method scores well above 0.259
`0.08 = 0.339,
`it is performing statistically significantly above the "chance*
`method.
`These chance values were derived analytically and then
`verified empirically. For 20 empirical random sense selection
`runs the average hit score was between .25 and .26.
`As "chance* provides a lower bound to compare our re
`sults against, hnman performance on the same tasks pro
`vides an upper bound. We had human subjects pick their
`estimate of the correct sense for each noun in WordNet for
`the first five Time documents. Two sets of printouts were
`distributed, each with the nouns in documents 1-5. Each
`noun's synset was given, along with its hypemym's synset
`and a gloss if available. "Hie subjects were thus given roughly
`the same sparse information that the software was getting.
`Although the humans could bring to bear their world knowl
`edge and linguistic knowledge, which should give them a
`large advantage, they were also handicapped by only re
`ceiving very local network data (node and parent only). In
`contrast', the software has the entire network at its disposal,
`albeit for its limited approach of looking at semantic dis
`tance. Also, the wording within synsets is quite terse and
`might not be highly suggestive of the actual sense intended.
`Thus, humans might find the information difficult to glean
`meaning from.
`Averaging over the two tests, the average percent correct
`was .782 and the average hit score .708. Of course this
`sample is too small for statistical robustness. Nevertheless,
`it succeeds in ^ving us an idea of how people do under these
`same conditions.
`For all of the experiments with the software, results ate
`given for hit score unless otherwise stated. Generally hit
`score is more informative than simple percent correct.
`
`4.2 Window vailation
`In the first series of experiments, window type and size were
`varied. First we tried frozen past windows of increasing size,
`from 1 to 100. These moving window results are given in
`Figures 1 and 2.
`As one can see, success climbs to a point and then ta
`pers off. This may be an effect of local discourse context
`size. As can be seen, the semantic distance approach pi^
`duccs results which are highly statistically significant. This
`is all the more significant, given the number of filters that
`the input text has gone through, and the amount of "noise"
`
`70
`
`Page 4 of 8
`
`

`

`' chance
`' software
`
`humm
`— chance
`—t— software
`
`30
`
`35
`frozen past window size
`
`Figure 3; Pinning down the optimal moving frozen past
`window; no mutn^ constraint window.
`
`frozen past window size
`
`Figure 1: Comparison of hit scores for chance, semantic
`distance software, and human subjects for Time documents
`1-5.
`
`' chance
`' software
`
`' chance
`' software
`
`frozen past window size
`
`Figure 2: The same data as in Figure 1 but with the vertical
`scale restricted.
`
`0
`
` 5
`
`10
`
`15
`
`initial mutual constraint window size
`
`Figure 4: Initial mutual constraint window with frozen past
`window s: 41.
`
`in the remaining 'Signal." Also, since the semantic net te-
`sources used arc relatively rudimentary compared to what
`they might be potentially, even greater success is possible.
`The next experiments attempted to pin down the peak
`performance seen near window sizes of 35 and 40. The best
`result was with a frozen past window size of 41, .437525. See
`Figure 3.
`Next, fixing the frozen past window size at 41, we tried
`augmenting this with an initial mutual constraint window.
`We were unable to proceed past an initial window size of 14
`because the runs were taking exponentially longer. The best
`results were with an initial mutual constraint window of size
`10, given the frozen past window of size 41 for all snbsequent
`terms (henceforth "(10,41)"). The hit score was .446771.
`All terms within the initial mutual constraint window had
`their sense selections fixed simultaneously once the objective
`function had determined the winning combination of senses.
`See Figure 4.
`We next tried a moving mutual constraint window. By
`the time we bad made the window size 9, the runs were
`taking about three hours, so we stopped there. The results
`were tantalizing, as the hit scores were just getting above .4
`at the point where we were forced to halt. See Figure 5.
`Note that these runs take longer per window size than
`the ones where only the initial terms are processed using
`mutual constraint. The moving mutual constraint window
`
`71
`
`Page 5 of 8
`
`

`

`chance
`
`sofCwaie
`
`chance
`—•— software
`original weights
`
`VW
`
`)2I
`
`0
`
` 2 4 6 8 10
`
`mutual constraint window size
`
`Figure 5: Moving mutual constraint window, no frozen past
`window.
`
`flrozen past window size
`
`Figure 7; Uniform weights, initial mutual constraint window
`= 10.
`
`chance
`—•— software
`— original weights
`
`a 8
`wk
`
`60
`
`a >
`
`■
`
`chance
`
`• no DRS
`ori^na] weights
`
`lirozen past window size
`
`Figure 6: No depth-relative scaling (DRS), initial mutual
`constraint window = 10.
`
`runs apply mutual constraint throughout an entire docu
`ment, not just the first set of terms. Thus, instead of a
`one-time cost incirrred for an exponential number of pair-
`wise comparisons, the cost here is incurred roughly n limes,
`where r is the number of terms in the document.
`
`4.3 Weight variation
`
`In the second series of experiments, we looked at the effects
`of varying the network weights in controlled ways. In each
`case we varied one parameter at a time.
`We varied the network edge weights to see the effect on
`disambiguation performance. First, we turned off depth-
`relative scaling. Interestingly, as shown in Figure 6, the
`hit scores over a range of frozen past window sizes with
`an initial mutual constraint window fixed at size 10 were
`low. This indicates that depth-relative scaling makes an
`important difference.
`Next, we tried making all weights equal. This gets rid of
`weight ranges and differences between relation types. The
`results indicate that this makes Uttle difference in the out
`come. Thus, the particular weights used may not make
`that much difference. Of course the original ranges were
`not that different from each other, nor that wide. The best
`results, still with an initial window of size 10 and a moving
`frozen past window size of 41, were lower than with the orig
`inal weighting scheme. So, perhaps weight distinctions and
`weight ranges help fine tune the performance. See Figure 7.
`
`0
`
`60
`40
` 20
`frozen past window size
`
`Figure 8: Privileged ajitonymy, initial mutual constraint
`10, frozen past.
`
`Next we gave antonyms privileged stains. Instead of
`having the highest weight of 2.5, we gave them the lowest
`weight, 0.5. This made negligible difference. See Figure 8.
`In the next experiment, we again saw a noticeable change
`in behavior. Here we made the network essentially hierarchi
`cal, deemphasizing the part/whole and antonymy relation
`ships and leaving the is-a relations dominant. The results
`are plotted in Figure 9.
`We see a fiat level of success across varying window
`sizes, and a mediocre one at that, the scores hovering in
`the range between .35 and .36. Thus, although using hier
`archical relationships gives us some power, we need to ex
`ploit the richness of additional relationships between topics
`to increase our ability to disambiguate. This Is evidence
`for the power of mized'link networks, containing both hi
`erarchic^ and nonhierarchical relations [Rada et ah, 1989;
`Kim and Kim, 1990].
`Removing type-specific fanout made little difference, but
`again the scores were slightly lower without it. See Fig
`ure 10.
`Finally, we tried the inverse of emphasizing the hierar
`chical relations. Here, we gave the hierarchical relations
`(hypemymy and hyponymy) large weights, ranging between
`5 and 10 instead of between 1 and 2. This also made little
`difference. See Figure 11.
`Other variations both in windowing and in weight varia
`tion are certainly conceivable. Nevertheless, we have pet-
`formed a number of preliminary experiments which have
`revealed useful insights. Specifically, it seems that depth-
`
`72
`
`Page 6 of 8
`
`

`

`—— chance
`—•— strictly hierarchical
`—
` original weights
`
`0
` 20
`frozen past window size
`
`Figure 9: Strictly hierarchical weighting, where part/whole
`and antonymy links are deemphasized; initial mutual con
`straint window = 10, frozen past.
`
`chance
`-•— software
`—— original weights
`
`60
`40
` 20
`0
`frozen past window size
`
`Figure 10: No type-spediic fanout, initial mutual constraint
`=: 10, frozen past.
`
`' chance
`' software
`' original weigltts
`
`0
` 20
`40 - 60
`frozen past window size
`
`Figure 11: Highly-weighted hierarchical relations, initial
`mutual constraint ~ 10, frozen past.
`
`relative scaling is important. In addition, it would appear
`that both hierarchical and nonhierarchical relations make
`contributions.
`We also looked at another set of five documents to see
`if the patterns would hold up. Only a few variations were
`tried, rather than the more comprehensive testing that was
`performed for the first five documents. Although the scores
`were lower, the overall trends of increase and plateauing
`were still recognizable. Although the number of terms in
`the second batch of documents was only slightly lower (272
`vs. 319), the expected hit score was much higher (c. .294).
`This was caused by a much higher proportion in the second
`batch of "high probability" terms, where there were very
`few senses for multi-sense terms. In the first batch there had
`been a larger number of difilcult terms, with many senses.
`It is not clear whether this difference made the software
`less effective for the second batch. Nevertheless, even with
`the results for the second batch included, the software has
`performed well.
`
`Documents 1-5
`.%
`hit
`correct
`
`score
`
`Documents 1-10
`%
`hit
`correct
`
`score
`
`chance
`(10.41)
`human
`
`.398
`.558
`.782
`
`.259
`.447
`.706
`
`.393
`.531
`
`.274
`.418
`
`—
`
`—
`
`Table 1. Summary of disambiguation results for nontrivial
`nouns in Time documents 1-S and 1-10. Results for human
`subjects are only available for documents 1-5.
`
`It is important to note that these figures are for the non-
`trivial terms only. Although such terms form a significant
`portion of the documents, focusing on them might give the
`erroneous impression that mote than half the terms in a doc
`ument would not be disambiguated properly. The truth is,
`taking the other document terms into account, most nouns
`in a document will be disambiguated correctly or will not
`need to be disambiguated in the first place. Therefore, docu
`ment content will actually be very well represented (at least
`at the individual term level).
`To substantiate tbb point, for documents 1 to 5 there
`are 544 nouns in WoidNet. Of these, only a small percent
`age are invalidated because there is no appropriate sense
`(type 3a-3e situations discussed earlier). 486 out of the 544
`terms are valid. Cut oi these 486 remaining terms, 319 are
`nontrivial and 167 arc trivial. Thus we get the 167 trivial
`terms correct for free. When we add that number to the
`178 nontrivial terms that (10,41) disambiguated correctly,
`we get 345 hits out of 486. This is 7l9S correct. Of course
`we are only looking here at the nouns. If we could bring the
`other parts of speech to bear, possibly without even invok
`ing sophisticated natural language processing techniques, we
`might strengthen our grasp of document content consider
`ably.
`
`5 Conclusion
`
`We have seen that applying a semantic network to mini
`mize semantic distance t^es us a long way towards the goal
`of removing extraneous search terms from free-text being
`indexed for retrieval. Yet the sophistication of natural lan
`guage processing required is kept minimal.
`The methods that we have employed trade off space for
`time — we use large data structures and keep them in main
`memory so that the runtime processing effort is kept to a
`
`73
`
`Page 7 of 8
`
`

`

`[Miller, 1990) George A. Miller. Nouns in WordNet: A lex
`ical inheritance system. IntemationalJaumal of Lexicog
`raphy, 3(4), 1990.
`[Rada' ct of,, 1989] Roy Rada, Hafcdh Mili, Ellen BIckneU,
`and Maria Blettner. Development and application of a
`metric on semantic nets. IBBB Tramactioni on Systems,
`Man, and Cybernetics, 19(l);17-30, Jaa./Feb. 1989.
`[Salton and McGiU, 1983] Gerard Salton and Michael J.
`McGill. /nfpoducfion to Afodern /n/orma(ion Retrieval.
`McGraw-Hill, 1983.
`[Tversky, 1977] A. Tversky. Features of similarity. Ptydio-
`logical Review, 84(4):327-3S2, 1977.
`[van Rysbcrgcn, 1983] C. J. van Rijsbergen. Information
`Retrieval Butter worths, second edition, 1983.
`[Voorhees ef of., 1992] Ellen M. Voorhees, Claudia Leacock,
`and Geoffrey Towell. Learning context to disambiguate
`word senses. In Proceeding

This document is available on Docket Alarm but you must sign up to view it.


Or .

Accessing this document will incur an additional charge of $.

After purchase, you can access this document again without charge.

Accept $ Charge
throbber

Still Working On It

This document is taking longer than usual to download. This can happen if we need to contact the court directly to obtain the document and their servers are running slowly.

Give it another minute or two to complete, and then try the refresh button.

throbber

A few More Minutes ... Still Working

It can take up to 5 minutes for us to download a document if the court servers are running slowly.

Thank you for your continued patience.

This document could not be displayed.

We could not find this document within its docket. Please go back to the docket page and check the link. If that does not work, go back to the docket and refresh it to pull the newest information.

Your account does not support viewing this document.

You need a Paid Account to view this document. Click here to change your account type.

Your account does not support viewing this document.

Set your membership status to view this document.

With a Docket Alarm membership, you'll get a whole lot more, including:

  • Up-to-date information for this case.
  • Email alerts whenever there is an update.
  • Full text search for other cases.
  • Get email alerts whenever a new case matches your search.

Become a Member

One Moment Please

The filing “” is large (MB) and is being downloaded.

Please refresh this page in a few minutes to see if the filing has been downloaded. The filing will also be emailed to you when the download completes.

Your document is on its way!

If you do not receive the document in five minutes, contact support at support@docketalarm.com.

Sealed Document

We are unable to display this document, it may be under a court ordered seal.

If you have proper credentials to access the file, you may proceed directly to the court's system using your government issued username and password.


Access Government Site

We are redirecting you
to a mobile optimized page.





Document Unreadable or Corrupt

Refresh this Document
Go to the Docket

We are unable to display this document.

Refresh this Document
Go to the Docket