throbber
Query by Image
`
`The QBIC System
`
`Myron Flickner, Harpreet Sawhney, Wayne Niblack, Jonathan Ashley, Qian Huang, Byron Dom,
`Monika Gorkani, Jim Hafher, Denis Lee, Dragutin Petkovie, David Steele, and Peter Yanker
`
`ZBMAlmaden Research Center
`
`m
`
`QBlC* lets users
`
`find pictorial information
`
`in large image and video
`
`databases based on color,
`
`shape, texture, and sketches.
`
`QBIC technology is part of
`
`several IBM products.
`
`‘To run an interacnve query, vult the QBIC Web sewer
`at http //imwqbic almaden ibm COW
`
`P
`
`icture yourself as a fashion designer needing images of fabrics
`with a particular mixture of colors, a museum cataloger looking
`for artifacts of a particular shape and textured pattern, or a movie
`producer needing a video clip of a red car-like object moving from right
`to left with the camera zooming. How do you find these images? Even
`though today’s technology enables us to acquire, manipulate, transmit,
`and store vast on-line image and video collections, the search method-
`ologies used to find pictorial information are still limited due to difficult
`research problems (see “Semantic versus nonsemantic” sidebar). Typ-
`ically, these methodologies depend on file IDS, keywords, or text associ-
`ated with the images. And, although powerful, they
`
`don’t allow queries based directly on the visual properties of the images,
`are dependent on the particular vocabulary used, and
`don’t provide queries for images similar to a given image.
`
`Research on ways to extend and improve query methods for image data-
`bases is widespread, and results have been presented in workshops, con-
`ferences,’.* and surveys.
`We have developed the QBIC (Query by Image Content) system to
`explore content-based retrieval methods. QBIC allows queries on large
`image and video databases based on
`
`example images,
`user-constructed sketches and drawings,
`selected color and texture patterns,
`
`Semantic versus nonsemantic information
`At first glance, content-based querying appears deceptively
`descriptions t o scenes through model matching-is an
`simple because we humans seem to be so good at it. If a pro-
`unsolved problem in image understanding. Humans are
`gram can be written to extract semantically relevant text
`much better than computers at extracting semantic descrip-
`phrases from images, the problem may be solved by using
`tions from pictures. Computers, however, are better than
`currently available text-search technology. Unfortunately, in
`humans at measuring properties and retaining these in
`an unconstrained environment, the task of writing this pro-
`long-term memory.
`gram is beyond the reach of current technology in image
`One of the guiding principles used by QBIC is to let com-
`understanding. At an artificial intelligence conference sev-
`puters do what they do best-quantifiable measurement-
`eral years ago, a challenge was issued to the audience to write
`and let humans do what they do best-attaching semantic
`a program that would identify all the dogs pictured in a chil-
`meaning. QBIC can find “fish-shaped objects,” since shape
`is a measurable property that can be extracted. However,
`dren’s book, a task most 3-year-olds can easily accomplish.
`Nobody in the audience accepted the challenge, and this
`since fish occur in many shapes, the only fish that will be
`found will have a shape close t o the drawn shape. This is not
`remains an open problem.
`the same as the much harder semantical query of finding
`Perceptual organization-the process of grouping image
`all the pictures of fish in a pictorial database.
`features into meaningful objects and attaching semantic
`
`0018-9162/95/$4 00
`
`1995 IEEE
`
`September 1995
`
`Page 1 of 10
`
`MINDGEEK EXHIBIT 1008
`
`

`

`Figure 1. QBlC query by drawn color. Drawn query specification on left; best 21 results sorted by similarity
`t o the query on right. The results were selected from a 12,968-picture database.
`
`~.
`
`,-\,
`S t i l l images
`
`)
`
`Scene
`
`\,
`
`/Motion objects
`\ Feature f
`
`‘ Shots
`
`Query interface
`
`Location
`Sketch
`Multiobject
`Shape
`ional Object Camera User Existlng
`exture motion motlon deflned
`
`1 *
`
`Match engine
`
`I
`
`Text
`
`I
`I
`I
`
`,
`
`-
`
`Color
`
`Texture
`
`Shape
`
`Multiobject
`
`Sketch
`
`Location
`
`Text
`
`Positional Object Camera User
`color/texture motion motion defined
`
`/’
`
`i
`
`User
`
`Computer
`
`Page 2 of 10
`
`MINDGEEK EXHIBIT 1008
`
`

`

`camera and object motion, and
`other graphical information.
`
`Two key properties of QBIC are (1) its
`use of image and video content-com-
`putable properties of color, texture, shape,
`and motion of images, videos, and their
`the queries, and (2) its graph-
`objects-in
`ical query language in which queries are
`posed by drawing, selecting, and other
`graphical means. Related systems, such as
`MIT’s Photobook3 and the Trademark and
`Art Museum applications from ETL,4 also
`address these common issues. This article
`describes the QBIC system and demon-
`strates its query capabilities.
`
`Figure 3. QBIC still image population interface. Entry for scene
`text at top. Tools in row are polygon outliner, rectangle outliner,
`ellipse outliner, paintbrush, eraser, line drawing, object
`translation, flood fill, and snake outliner.
`
`1
`
`I
`
`eJ
`
`$
`
`QBIC SYSTEM OVERVIEW
`Figure 1 illustrates a typical QBIC query.”
`The left side shows the query specification,
`where the user painted a large magenta cir-
`cular area on a green background using standard drawing
`tools. Query results are shown on the right: an ordered list of
`“hits” similar to the query. The order of the results is top to
`bottom, then left to right, to support horizontal scrolling. In
`general, all queries follow this model in that the query is spec-
`ified by using graphical means-drawing, selecting from a
`results
`color wheel, selecting a sample image, and so on-and
`are displayed as an ordered set of images.
`To achieve this functionality, QBIC has two main com-
`ponents: database population (the process of creating an
`image database) and database query. During the popula-
`tion, images and videos are processed to extract features
`describing their content-colors,
`textures, shapes, and
`camera and object motion-and
`the features are stored in
`a database. During the query, the user composes a query
`graphically. Features are generated from the graphical
`query and then input to a matching engine that finds
`images or videos from the database with similar features.
`Figure 2 shows the system architecture.
`
`Data model
`For both population and query, the QBIC data model has
`
`still images or scenes (full images) that contain objects
`(subsets of an image), and
`video shots that consist of sets of contiguous frames and
`contain motion objects.
`
`For still images, the QBIC data model distinguishes between
`“scenes” (or images) and “objects.” A scene is an image or
`single representative frame of video. An object is a part of
`a scene-for
`example, the fox in Figure 3-or
`a moving
`entity in a video. For still image database population, fea-
`tures are extracted from images and objects and stored in a
`database as shown in the top left part of Figure 2.
`Videos are broken into clips called shots. Representative
`
`frames, or r-frames, are generated for each extracted shot.
`R-frames are treated as still images, and features are
`extracted and stored in the database. Further processing
`of shots generates motion objects-for
`example, a car
`moving across the screen.
`Queries are allowed on objects (“Find images with a red,
`round object”), scenes (“Find images that have approxi-
`mately 30-percent red and 15-percent blue colors”), shots
`(“Find all shots panning from left to right”), or any com-
`bination (“Find images that have 30 percent red and con-
`tain a blue textured object”).
`In QBIC, similarity queries are done against the data-
`base of pre-extracted features using distance functions
`between the features. These functions are intended to
`mimic human perception to approximate a perceptual
`ordering of the database. Figure 2 shows the match
`engine, the collection of all distance functions. The match
`engine interacts with a filteringhndexing module (see
`“Fast searching and indexing” sidebar, next page) to sup-
`port fast searching methodologies such as indexing. Users
`interact with the query interface to generate a query spec-
`ification, resulting in the features that define the query.
`
`DATABASE POPULATION
`In still image database population, the images are
`reduced to a standard-sized icon called a thumbnail and
`annotated with any available text information. Object
`identification is an optional but key part of this step. It lets
`users manually, semiautomatically, or fully automatically
`identify interesting regions-which we call objects-in
`the images. Internally, each object is represented as a
`binary mask. There may be an arbitrary number of objects
`per image. Objects can overlap and can consist of multi-
`ple disconnected components like the set of dots on a
`polka-dot dress. Text, like “baby on beach,” can be associ-
`ated with an outlined object orwith the scene as a whole.
`
`’’ The scene image database used in thefigures consists of about 2 4 5 0
`imagesfrom the Mediasource Series of images and audiofrom Applred
`Optical Media Corp., 4,100 imagesfiom the PhotoDiscsampler CD, 950
`imagesfrom the Corel Professional Photo CD collection, and 450 images
`J?om an IBM collection.
`
`Object-outlining tools
`Ideally, object identification would be automatic, but
`this is generally difficult. The alternative-manual
`iden-
`tification-is
`tedious and can inhibit query-by-content
`
`September 1995
`
`Page 3 of 10
`
`MINDGEEK EXHIBIT 1008
`
`

`

`Fast searching and indexing
`Indexing tabular data for exact matching or range
`searches in traditional databases is a well-understood prob-
`lem, and structures like B-trees provide efficient access
`mechanisms. In this scenario, indexing assures sublinear
`search while maintaining completeness; that is, all records
`satisfying the query are returned without the need for
`examining each record in the database. However, in the con-
`text of similarity matching for visual content, traditional
`indexing methods may not be appropriate. For queries in
`which similarity is defined as a distance metric in high-
`dimensional feature spaces (for example, color histogram
`queries), indexing involves clustering and indexable repre-
`sentations of the clusters. In the case of queries that com-
`bine similarity matching with spatial constraints on objects,
`the problem is more involved. Data structures for fast access
`of high-dimensional features for spatial relationships must
`be invented.
`In a query, features from the database are compared to
`corresponding features from the query specification to
`determine which images are a good match. For a small data-
`base, sequentigl scanning of the features followed by
`straightforward similarity computations is adequate. But as
`the database grows, this combination can be too slow. To
`speed up the queries, we have investigated a variety of tech-
`niques. Two of the most promising follow.
`
`Filtering
`A computationally fast filter is applied to all data, and only
`items that passthrough the filter are operated on by the sec-
`ond stage, which computes the true similarity metric. For
`example, in QBlC we have shown that color histogram match-
`ing, which is based on a 256-dimensional color histogram and
`requires a 256 matrix-vector multiply, can be made efficient
`by filtering. The filtering step employs a much faster com-
`putation in a 3D space with no loss in accuracy. Thus, for a
`query on a database of 10,000 elements, the fast filter is
`applied to produce the best 1,000 color histogram matches.
`These filtered histograms are subsequently passed to the
`slower complete matching operation to obtain, say, the best
`200 matches t o displayto a user, with the guarantee that the
`global best 200 in the database have been found.
`
`Indexing
`For low-dimensional features such as average color and
`texture (each 3D), multidimensional indexing methods such
`as R*-trees can be used. For high-dimensional features-for
`example, our 20-dimensional moment-based shape feature
`vector-the dimensionality is reduced using the K-L, or prin-
`cipal component, transform. This produces a low-dimen-
`sional space, as low astwo or three dimensions, which could
`be indexed by using /?*-trees.
`
`applications. As a result, we have devoted considerable
`effort to developing tools to aid in this step. In recent
`work, we have successfully used fully automatic unsu-
`pervised segmentation methods along with a fore-
`ground/background model to identify objects in a re-
`stricted class of images. The images, typical of museums
`and retail catalogs, have a small number of foreground
`objects on a generally separable background. Figure 4
`shows example results. Even in this domain, robust algo-
`rithms are required because of the textured and varie-
`gated backgrounds.
`
`We also provide semiautomatic tools for identifying
`objects. One is an enhanced flood-fill technique. Flood-fill
`methods, found in most photo-editing programs, start
`from a single object pixel and repeatedly add adjacent pix-
`els whose values are within some given threshold of the
`original pixel. Selecting the chreshold, which must change
`from image to image and object to object, is tedious. We
`automatically calculate a dynamic threshold by having the
`user click on background as well as object points. For rea-
`sonably uniform objects that are distinct from the back-
`ground, this operation allows fast object identification
`
`Figure 4. Top row is the original image. Bottom row contains the automatically extracted objects using a
`foregroundhackground model. Heuristics encode the knowledge that objects tend to be in the center of
`the picture.
`
`Computer
`
`Page 4 of 10
`
`MINDGEEK EXHIBIT 1008
`
`

`

`without manually adjust-
`ing a threshold. The exam-
`ple in Figure 3 shows an
`object, a fox, identified by
`using only a few clicks.
`We designed another
`outlining tool to help users
`track object edges. This tool
`takes a user-drawn curve
`and automatically aligns it
`with nearby image edges.
`Based on the “snakes” con-
`cept developed in recent
`computer vision research,
`the tool finds the curve that
`maximizes the image gra-
`dient magnitude along the
`curve.
`The spline snake formu-
`lation we use allows for
`smooth solutions to the
`resulting nonlinear mini-
`mization problem. The
`computation is done at
`interactive speeds so that,
`as the user draws a curve, it
`is “rubber-banded’’ to lie
`along object boundaries.
`
`Video data
`For video data, database
`population has three major
`components:
`
`shot detection,
`representative frame cre-
`ation for each shot, and
`derivation of a layered representation of coherently
`moving structures/objects.
`
`Shots are short sequences of contiguous frames that we
`use for annotation and querying. For instance, a video clip
`may consist of a shot smoothly panning over the skyline
`of San Francisco, switching to a panning shot of the Bay
`meeting the ocean, and then to one that zooms to the
`Golden Gate Bridge. In general, a set of contiguous frames
`may be grouped into a shot because they
`
`depict the same scene,
`signify a single camera operation,
`contain a distinct event or an action like a significant
`presence and persistence of an object, or
`are chosen as a single indexable entity by the user.
`
`Our effort is to detect many shots automatically in a pre-
`processing step and provide an easy-to-use interface for
`the rest.
`
`SHOT DETECTION. Gross scene changes or scene cuts
`are the first indicators of shot boundaries. Methods for
`detecting scene cuts proposed in the literature essentially
`fall into two classes: (1) those based on global represen-
`
`Figure 5. Scene cuts automatically extracted from a 1,148-frame sales demo
`from Energy Productions.
`
`~
`
`U-M-I
`BEST COPY AVAILABLE
`-~
`tations like color/intensity histograms without any spa-
`tial information, and (2) those based on measuring dif-
`ferences between spatially registered features like
`intensity differences. The former are relatively insensi-
`tive to motion but can miss cuts when scenes look quite
`different but have similar distributions. The latter are
`sensitive to moving objects and camera. We have devel-
`oped a method that combines the strengths of the two
`classes of detection. We use a robust normalized corre-
`lation measure that allows for small motions and com-
`bines this with a histogram distance m e a s ~ r e . ~ Results
`on a few videos containing from 2,000 to 5,000 frames
`show no misses and only a few false cuts. Algorithms for
`signaling edit effects like fades and dissolves are under
`development. The results of cut detection on a video con-
`taining commercial advertisement clips are shown in
`Figure 5 .
`Shots may also be detected by finding changes in camera
`operation. Common camera transformations like zoom,
`pan, and illumination changes can be modeled as unknown
`affine 2 x 2 matrix transformations of the 2D image coor-
`dinate system and of the image intensities themselves. We
`have developed an algorithm6 that computes the dominant
`global view transformation while it remains insensitive to
`nonglobal changes resulting from independently moving
`
`September 1995
`
`Page 5 of 10
`
`MINDGEEK EXHIBIT 1008
`
`

`

`objects and local brightness changes. The
`affine transformations that result from this
`computation can be used for camera oper-
`ation detection, shot boundary detection
`based on the camera operation, and creat-
`ing a synthetic r-frame wherever appropri-
`ate.
`Shot boundaries can also be defined on
`the basis of events: appearance/disap-
`pearance of an object, distinct change in
`the motion of an object, or similar events.
`For instance, segmenting an object of inter-
`est based on its appearance and/or motion,
`and tracking it throughout its significant
`presence may be used for defining shots.
`
`REPRESENTATIVE FRAME GENERATION.
`Once the shot boundaries have been
`detected, each shot is represented using an
`r-frame. R-frames are used for several pur-
`poses. First, during database population,
`r-frames are treated as still images in which
`objects can be identified by using the pre-
`viously described methods. Secondly, dur-
`ing query, they are the basic units initially
`returned in a video query. For example, in
`a query for shots that are dominantly red,
`a set of r-frames will be displayed. To see
`the actual video shot, the user clicks on the
`displayed r-frame icon.
`The choice of an r-frame could be as sim-
`ple as a particular frame in the shot: the
`
`Figure 6. Top: Three frames from the charlie sequence and the
`resulting dynamic mosaics of the entire sequence. Below that is a
`mosaic from a video sequence of Yosemite National Park. Bottom:
`Original images and segmented motion layers for the flower gar-
`den sequence in which only the camera is moving. The flower
`bed, tree, and background have been separated into three layers
`shown in different shades of gray.
`
`U-M-I
`BEST COPY AVAILABLE
`
`Figure 7. Top: Query by histogram color. Histogram color query specification on left; best 21 results from a
`12,966-picture database on right. Bottom: A query for a red video r-frame. The color picker is on the left;
`the resulting r-frame thumbnails of the best matches are shown on the right. Each thumbnail is an active
`button that allows the user to play the shot.
`
`Computer
`
`Page 6 of 10
`
`MINDGEEK EXHIBIT 1008
`
`

`

`first, the last, or the middle. However, in situations such
`as a long panning shot, no single frame may be represen-
`tative of the entire shot. We use a synthesized r-frame7,8
`created by seamlessly mosaicking all the frames in a given
`shot using the computed motion transformation of the
`dominant background. This frame is an authentic depic-
`tion of all background captured in the whole shot. Any
`foreground object can be superimposed on the back-
`ground to create a single, static visual representation of
`the shot. The r-frame mosaicking is done by using warp-
`ing transforms that result from automatic dominant
`motion computation. Given a video sequence with domi-
`nant motion and moving object(s), the 2D motion esti-
`mation algorithm is applied between consecutive pairs of
`frames. Then, a reference frame is chosen, and all the
`frames are warped into the coordinate system of the ref-
`erence frame to create the mosaicked r-frame.
`Figure 6 illustrates mosaic-based r-frame creation on a
`video sequence of an IBM commercial. Three frames of
`this sequence plus the final mosaic are shown. Two dom-
`inant-component-only mosaics of the charlie sequence are
`shown in Figure 6. In one case, the moving object has been
`removed from the mosaic by using temporal median fil-
`tering on the frames in the shot. In the other case, the mov-
`ing object remains from the first frame in the sequence.
`We are also developing methods to visually represent the
`object motion in the r-frame.
`
`LAYERED REPRESENTATION. TO facilitate automatic
`segmentation of independently moving objects and sig-
`nificant structures, we take further advantage of the
`time-varying nature of video data to derive what is called
`a layered representation9 of video. The different layers
`are used to identify significant objects in the scene for fea-
`ture computations and querying. Our algorithm divides a
`shot into a number of layers, each with its own 2D affine
`motion parameters and region of support in each frame.l0
`The algorithm is first illustrated on a shot where the
`scene is static but the camera motion induces parallax
`
`motion onto the image plane due to the different depths
`in the scene. Therefore, surfaces and objects that may cor-
`respond to semantically useful entities can be segmented
`based on the coherence of their motion. Figure 6 (bottom
`row) shows the results for the layers from the flower gar-
`den sequence.
`
`SAMPLE QUERIES
`For each full-scene image, identified image object, r-
`frame, and identifiedvideo object resulting from the above ,
`processing, a set of features is computed to allow content-
`based queries. The features are computed and stored dur-
`ing database population. We present a brief description of
`the features and the associated queries. Mathematical
`details on the features and matching methods can be
`found in Ashley et al." and Niblack et a1.l2
`Average color queries let users find images or objects
`that are similar to a selected color, say from a color wheel,
`or to the color of an object. The feature used in the query
`is a 3D vector of Munsell color coordinates. Histogram
`color queries return items with matching color distribu-
`tions-say, a fabric pattern with approximately 40 per-
`cent red and 20 percent blue. For this case, the underlying
`feature is a 256-element histogram computed over a
`quantized version of the color space.
`Figure 7 shows a histogram query on still images and a
`color query on video r-frames. Note that in the query spec-
`ification for the histogram query of Figure 7, the user has
`selected percentages of two colors (blue and white) by
`adjusting sliders. Using such a query, an advertising agent
`could, for example, search for a picture of a beach scene,
`one predominantly blue (for sky and water) and white (for
`sand and clouds); or find images with similar color spreads
`for a uniform ad campaign. The average color query
`demonstrates a query against a video shot database where
`the user is searching for red r-frames. Again, the query
`specification is on the left and the best hits are on the right. 1
`Figure 8 shows an example texture query. In this case,
`the query is specified by selecting from a sampler-a
`set of
`
`~
`
`~
`
`September 1995
`
`Page 7 of 10
`
`MINDGEEK EXHIBIT 1008
`
`

`

`Figure 9. Top: Query by sketch. Sketched query specification on left; best 21 hits from a 12,965-image data-
`base on right. Bottom: A Multi-* object query. The query specification on the left describes a query for
`images with a red round object and a green textured object. Best 20 matches shown on right.
`
`prestored example images. The underlying texture fea-
`tures are mathematical representations of coarseness, con-
`trast, and directionality features. Coarseness measures the
`scale of a texture (pebbles versus boulders), contrast
`describes its vividness, and directionality describes
`whether it has a favored direction (like grass) or not (like
`a smooth object).
`An object shape query is shown in Figure 8. In this case,
`the query specification is the drawn shape on the left. Area,
`circularity, eccentricity, major-axis direction, features
`derived from the object moments, and a set of tangent
`angles around the object perimeter are the features used
`to characterize and match shapes.
`Figure 9 illustrates query by sketch. In this case, the
`query specification is a freehand drawing of the dominant
`lines and edges in the image. The sketch feature is an
`automatically extracted reduced-resolution “edge map.”
`Matching is done by using a template-matching tech-
`nique.
`A multiobject query asking for images that contain both
`a red round object and a green textured object is shown in
`the bottom of Figure 9. The features are standard color
`and texture. The matching is done by combining the color
`and texture distances. Combining distances is applied to
`arbitrary sets of objects and features to implement logical
`And semantics.
`
`WE HAVE DESCRIBED A PROTOTYPE SYSTEM that uses image
`and video content as the basis for retrievals. Technology
`from this prototype has already moved into a commercial
`stand-alone product, IBMs Ultimedia Manager, and is part
`
`of IBM’s Digital Library and DB2 series of products. Other
`companies are beginning to offer products with similar
`capabilities. Key challenges remain in making this tech-
`nology pervasive and useful.
`
`ANNOTATION AND DATABASE POPULATION TOOLS.
`Automatic methods (such as our Positional Color query)
`that don’t rely on object identification, methods that iden-
`tify objects automatically as in the museum image exam-
`ple, fast and easy-to-use semiautomatic outlining tools,
`and motion-based segmentation algorithms will enable
`additional application areas.
`
`FEATURE EXTRACTION AND MATCHING METHODS. New
`mathematical representations of video, image, and object
`attributes that capture “interesting” features for retrieval
`are needed. Features that describe new image properties
`such as alternate texture measures or that are based on
`fractals or wavelet representations, for example, may offer
`advantages of representation, indexability, and ease of
`similarity matching.
`
`INTEGRATION WITH TEXT AND PARAMETRIC ANNOTA-
`TION. Query by visual content complements and extends
`existing query methods. Systems must be able to integrate
`queries combining date, subject matter, price, and avail-
`ability with content properties such as color, texture, and
`shape.
`
`EXTENSIBILITY AND FLEXIBILITY. System architectures
`must support the addition of new features and new match-
`ing/similarity measures. Real applications often require
`
`Computer
`
`Page 8 of 10
`
`MINDGEEK EXHIBIT 1008
`
`

`

`new features, say a face-matching module, to add to their
`existing content-based retrieval capabilities.
`
`USER INTERFACE. The user interface must be designed
`to let users easily select content-based properties, allow
`these properties to be combined with each other and with
`text or parametric data, and let users reformulate queries
`and generally navigate the database.
`
`INDEXING AND PERFORMANCE. As image and video col-
`lections grow, system performance must not slow down
`proportionately. Indexing, clustering, and filtering meth-
`ods must be designed into the matching methods to main-
`tain performance.
`With these technologies, the QBIC paradigm of visual
`content querying, combined with traditional keyword and
`text querying, will lead to powerful search engines for mul-
`timedia archives. Applications will occur in areas such as
`decision support for retail marketing, on-line stock photo
`and video management, cataloging for library and
`museum collections, and multimedia-enabled applica-
`tions in art, fashion, advertising, medicine, and science. I
`
`References
`1. IFIP, Visual Database Systems I and 11, Elsevier Science Pub-
`lishers, North-Holland, 1989 and 1992.
`2. Proc. Storage and Retrieval forlmage and VideODatabases I, 11,
`andlll, Vols. 1,908; 2,185; and 2,420; W. Niblackand R. Jain,
`eds., SPIE, Bellingham, Wash., 1993,1994, and 1995.
`3. A. Pentland, R.W. Picard, and S. Sclaroff, “Photobook Tools
`for Content-Based Manipulation of Image Databases,” Proc.
`Storage and Retrieval for Image and Video Databases 11, Vol.
`2,185, SPIE, Bellingham, Wash., 1994, pp. 34-47.
`4. T. Kato, T. Kurita, and H. Shimogaki, “Intelligent Visual Inter-
`action with Image Database Systems-Toward
`the Multime-
`dia Personal Interface,” J. Information Processing (Japan),
`Vol. 14, No. 2,1991, pp. 134.143.
`5. A. Nagasaka and Y. Tanaka, “Automatic Video Indexing and
`Full-Video Search for Object Appearances,” Visual Database
`Systems, 11, IFIP Trans. A-7, Elsevier Science Publishers, North-
`Holland, 1992, pp. 113-127.
`6. H.S. Sawhney, S. Ayer, and M. Gorkani, “Model-Based 2D &
`3D Dominant Motion Estimation for Mosaicking and Video
`Representation,”hoc. Fifih Int’l Conf. Computer Vision, Order
`No. PR07042, IEEE CS Press, Los Alamitos, Calif., 1995, pp.
`583-590; http://www.almaden.ibm.com/pub/cs/reports/
`vision/dominant-moti0n.ps.Z.
`7. Y.T.A. Akutsu, K. Otsuji, and T. Sadakata, “VideoMAP and
`VideoSpaceIcon: Tools for Anatomizing Video Content,”ACM
`INTERCHI, 1993, pp. 131.136.
`8. L A . Teodosio and W. Bender, “Salient Video Stills: Content
`and Context Preserved,” ACM Int’l Conf. Multimedia, ACM,
`New York, 1993.
`9. J.Y.A. Wang and E.H. Adelson, “Layered Representation for
`Motion Analysis,” Proc. Computer Vision and Pattern Recog-
`nition Conf., IEEE CS Press, Los Alamitos, Calif., 1993, pp.
`361-366.
`10. S. Ayer and H.S. Sawhney, “Layered Representation of
`Motion Video Using Robust Maximum-Likelihood Estimation
`of Mixture Models and MDL Encoding,”Proc. Fqth Int’l Conf
`Computer Vision, Order No. PR07042, IEEE CS Press, Los
`
`MobiCom is ACM’s annual international conference, established
`to serve as the premier forum for addressing networks, systems,
`algorithms and applications that support the symbiosis of portable
`computers and wireless networks.
`
`MobiCom95 is the first conference to bring together computing
`researchers and professionals studying mobility issues from several
`different perspectives:
`0 Designing effective mobile computing environments that can
`deal with changing and sporadic connectivity
`Architecting end-to-end networks of wireless and fixed segments
`Building applications for providing robust ubiquitous service
`
`Two days of single track sessions, including:
`0 ATM & Wireless Networks
`0 The Impact of Mobility on
`0 Mobile Systems Performance
`Data Management
`0 Multimedia in a
`0 Mobile Network Protocols
`0 Location Management
`Mobile Environment
`0 TCPlIP over Wireless Networks
`Other Emerging Issues
`speciace-:
`Tutorials on Mobile Computing, Mobile IP and
`Wireless Mobile Networking, November 13
`0 Keynote speech on Nomadic Computing by
`Leonard Kleinrock, UCLA
`Invited talk: Packet Radio to MII, Barry Leiner, ARPA
`Panel: The future of mobile computing as shaped by
`government funding, by researchers, and by society
`Opportunities to demo; to “cross-pollinate” among people
`with overlapping interests but different emphases
`
`Page 9 of 10
`
`MINDGEEK EXHIBIT 1008
`
`

`

`Alamitos, Calif., 1995, pp. 777-784, http://www.almaden.
`ibm.com/pub/cs/reports/vision/layered_motion.ps.Z.
`11. J. Ashley et al., “Automatic and Semiautomatic Methods for
`Image Annotation and Retrieval in QBIC,” Proc. Storage and
`Retrievalfor Image and Video Databases 111, Vol. 2,420, SPIE,
`Bellingham, Wash., 1995, pp. 24-25.
`12. W. Niblack et al., “The QBIC Project: Querying Images by
`Content Using Color, Texture, and Shape,”Proc. Storage and
`Retrievalfor Image and Video Databases, Vol. 1,908, SPIE,
`Bellingham, Wash., 1993, pp. 173-187.
`
`Myron “Flick” Flickner architected parts of QBlC and
`implemented several shape-matching methods and the data-
`base population GUI. His current research interests include
`image and shape representations, content-based image
`retrieval, and vision-based man-machine interaction.
`
`Harpreet “Video” Sawhney has done much of the
`exploratory and implementation work to exten

This document is available on Docket Alarm but you must sign up to view it.


Or .

Accessing this document will incur an additional charge of $.

After purchase, you can access this document again without charge.

Accept $ Charge
throbber

Still Working On It

This document is taking longer than usual to download. This can happen if we need to contact the court directly to obtain the document and their servers are running slowly.

Give it another minute or two to complete, and then try the refresh button.

throbber

A few More Minutes ... Still Working

It can take up to 5 minutes for us to download a document if the court servers are running slowly.

Thank you for your continued patience.

This document could not be displayed.

We could not find this document within its docket. Please go back to the docket page and check the link. If that does not work, go back to the docket and refresh it to pull the newest information.

Your account does not support viewing this document.

You need a Paid Account to view this document. Click here to change your account type.

Your account does not support viewing this document.

Set your membership status to view this document.

With a Docket Alarm membership, you'll get a whole lot more, including:

  • Up-to-date information for this case.
  • Email alerts whenever there is an update.
  • Full text search for other cases.
  • Get email alerts whenever a new case matches your search.

Become a Member

One Moment Please

The filing “” is large (MB) and is being downloaded.

Please refresh this page in a few minutes to see if the filing has been downloaded. The filing will also be emailed to you when the download completes.

Your document is on its way!

If you do not receive the document in five minutes, contact support at support@docketalarm.com.

Sealed Document

We are unable to display this document, it may be under a court ordered seal.

If you have proper credentials to access the file, you may proceed directly to the court's system using your government issued username and password.


Access Government Site

We are redirecting you
to a mobile optimized page.





Document Unreadable or Corrupt

Refresh this Document
Go to the Docket

We are unable to display this document.

Refresh this Document
Go to the Docket