throbber
Analyzing Intrinsic Motion Textures
`Created from Naturalistic Video Captures
`
`Angus Graeme Forbes1, Christopher Jette1 and Andrew Predoehl2
`1School of Information: Science, Technology, and Arts, University of Arizona, Tucson, United States
`2Department of Computer Science, University of Arizona, Tucson, United States
`angus.forbes@sista.arizona.edu, cnj@email.arizona.edu, predoehl@email.arizona.edu
`
`Keywords:
`
`Intrinsic Motion Textures, Psychophysics, Perception, Metrics for Motion, Dynamic Textures.
`
`Abstract:
`
`This paper presents an initial exploration of the plausibility of incorporating subtle motions as a useful modal-
`ity for encoding (or augmenting the encoding of) data for information visualization tasks. Psychophysics
`research indicates that the human visual system is highly responsive to identifying and differentiating even the
`subtlest motions intrinsic to an object. We examine aspects of this intrinsic motion, whereby an object stays in
`one place while a texture applied to that object changes in subtle but perceptible ways. We hypothesize that the
`use of subtle intrinsic motions (as opposed to more obvious extrinsic motion) will avoid the clutter and visual
`fatigue that often discourages visualization designers from incorporating motion. Using transformed video
`captures of naturalistic motions gathered from the world, we conduct a preliminary user study that attempts
`ascertains the minimum amount of motion that is easily perceptible to a viewer. We introduce metrics which
`allow us to categorize these motions in terms of flicker (local amplitude and frequency), flutter (global ampli-
`tude and frequency), and average maximum contrast between a pixel and its immediate neighbors. Using these
`metrics (and a few others), we identify plausible ranges of motion that might be appropriate for visualization
`tasks, either on their own or in conjunction with other modalities (such as color or shape), without increasing
`visual fatigue. Based on an analysis of these initial preliminary results, we propose that the use of what we
`term “intrinsic motion textures” may be a promising modality appropriate for a range of visualization tasks.
`
`1
`
`INTRODUCTION
`
`Rustling leaves, flickering flames, sunlight sparkling
`on water – every day we are continually confronted
`with naturalistic motion as we navigate the world.
`Information visualization, as a field, examines how
`meaning is effectively conveyed through visually-
`encoded data. However, dynamic visualizations that
`encode data using motion have not been as widely ex-
`plored, as motion is generally considered to be too
`distracting a modality for representing information
`effectively. The human visual system is exception-
`ally adept at identifying differences in texture (Em-
`rith et al., 2010), and observing both the movement
`of objects or the movement within an object (Chalupa
`et al., 2004). While far from being fully understood,
`a growing body of research indicates the presence of
`multiple neurological mechanisms for processing ex-
`trinsic motion (the movement of an object) and intrin-
`sic motion (local movement within a single object)
`(Lu and Sperling, 2001; Nishida et al., 1997). Extrin-
`sic and intrinsic motion is also termed “first-order”
`
`or “second-order,” respectively: first-order motion re-
`ferring to the perception of a change in luminance
`across the visual field; second-order motion referring
`instead to the perception of changes in texture or con-
`trast. While both types of motion can indicate a global
`movement of a single entity, in this paper we examine
`the use of second-order motion to indicate motions
`within a stationary object. We introduce a method for
`transforming real-world motion into abstract motion
`textures with subtle motions, which we are calling in-
`trinsic motion textures. These transformed video cap-
`tures of real-world naturalistic motion allow us to ex-
`periment with non-distracting motion without includ-
`ing their representational aspects. Through the trans-
`formation of, say, the movement of water in a stream,
`we are able to capture the motion without directly re-
`minding users that they are looking at water. We also
`introduce an easy-to-calculate set of metrics to char-
`acterize these intrinsic motions. While the vast range
`of possible motions makes it rather daunting to at-
`tempt to encompass all types of movement via a sin-
`gle set of metrics, ours capture the main features of
`
`Lightricks Ltd.
`EX1018
`Page 1 of 7
`
`

`

`Figure 1: Stages of the video capture of naturalistic motion being transformed into an intrinsic motion texture. Here we show
`a single frame (A) as it is (B) cropped, (C) desaturated, (D) contrast-compressed with a low pixel range, and (E) pixelated. A
`further step uses temporal smoothing to mitigate extra flickering introduced from the pixelation.
`
`intrinsic motion. We conducted a user study asking
`participants to evaluate a set of intrinsic motion tex-
`tures. We were primarily interested in gathering in-
`sight into the following question – What is the least
`amount of motion needed in order to easily identify
`differences between highly similar motions? We dis-
`covered that even intrinsic motion textures with a low
`contrast range are easily distinguishable, as long as
`certain amounts of flicker, flutter, and average maxi-
`mum contrast (defined below) are present.
`In the influential text, “Semiology of Graphics,”
`Jacques Bertin introduces a typology of retinal vari-
`ables applicable to the communication of informa-
`tion. Although explicitly discussing static print me-
`dia, Bertin hints at the ability of texture, unlike other
`retinal variables, to produce a vibratory effect that is
`the “collusion” of physiological and psychological ef-
`fects. Although he does not explore this issue of vi-
`bratory effects further, he encourages designers “to
`make the most of this variation, to obtain the res-
`onance without provoking an uncomfortable sensa-
`tion, to flirt with ambiguity without succumbing to
`it.” Bertin speaks of texture in terms of the vary-
`ing thicknesses of lines, points, or shapes to indicate
`an overview similarity or differentiation between dif-
`ferent qualities, or an ordered-ness within the same
`quality. However, he does not believe that the use of
`texture is refined enough to allow viewers to perceive
`proportionality and therefore is not effective at allow-
`ing users to perform quantitative tasks using texture
`alone (Bertin, 2010).
`Nonetheless, investigations into the use of static
`textures in visualization contexts have found that they
`can be effective for representing quantitative data in
`certain tasks. A seminal paper (Ware and Knight,
`1995) uses Gabor patches to parametrically investi-
`gate aspects of texture that may be useful in informa-
`tion display, such as orientation, size, and contrast.
`Applied research introduced in (Kujala and Lukka,
`2003) explores the use of a “parameter hierarchy” in
`the procedural creation of perceptually distinct tex-
`tures for the effective display of information. More
`
`recently, (Byelas and Telea, 2009) uses overlapping
`textures to indicate multivariate data in software dia-
`grams, explicitly attempting to indicate an increased
`number of variables within the diagram. (House et al.,
`2006) describes a technique for the perceptual opti-
`mization of complex visualizations involving layered
`textures. And (Interrante and Kim, 2001) and (Kim
`et al., 2004) explore the efficacy of using various ori-
`entations and types of textures to facilitate shape per-
`ception. These examples are mostly explicitly con-
`cerned with the potential expanded range of variables
`that can be expressed, while at the same time aware
`that the cost of this expanded range might be percep-
`tual fatigue or cognitive overload, or worse, an in-
`ability to clearly distinguish differences. (Interrante,
`2000) examines the use of overlapping naturalistic
`textures to indicate multivariate data while mitigating
`against the “extraneous stress” that might occur with
`synthesized textures. In addition to being a potentially
`effective representation of a wide range of quantita-
`tive information, the use of certain aspects of natu-
`ralistic textures, such as variability, might be used to
`indicate extra information, such as uncertainty. Inter-
`estingly, recent perceptual experiments, such as (Em-
`rith et al., 2010), confirm that humans perceive even
`minute alterations in texture, but note that it is in fact
`somewhat easier to discern differences between syn-
`thetic textures than natural textures.
`Motion is often used to signal transitions between
`views and contexts, to signal interruptions, and to in-
`dicate temporal aspects of data. However, it is less
`frequently used as an encoding mechanism for quanti-
`tative or qualitative information. (Forbes et al., 2010)
`presents a data visualization framework than enables
`animation to be mapped to dynamic streams of data,
`and (Bostock et al., 2011) describes a framework
`that includes “transition” operators for animating data
`points. The use of motion in visualization elicits con-
`cern about visual clutter and perceptual fatigue, even
`while potentially providing an expanded toolset for
`representing information. At its most extreme, the in-
`judicious use of motion in information might cause
`
`Page 2 of 7
`
`

`

`significant visual stress (Ware, 2004). One group
`of security researchers (Conti et al., 2005) even de-
`scribes the potential for malicious hackers to takeover
`an information visualization system and alter its vi-
`sual output to induce epileptic seizures.
`Results from (Bartram and Ware, 2002) show that
`small, brief, and graphically simple extrinsic motions
`are perceptually efficient ways to distinguish objects
`in a crowded display. In particular, they note that a
`synchronization of elements is required in order for
`them to be effectively recognized as similar. That is,
`the timing of the motion is as important as the mo-
`tion itself. Research into extrinsic motion cues, or
`“moticons” (Bartram et al., 2003), finds that motion
`coding is independent from color and shape coding
`and that more subtle motions are less distracting to
`users yet easily perceived. A series of experiments
`that analyzed extrinsic aspects of motion– velocity,
`direction, and on-off blinking– finds that these prop-
`erties are all effective at encoding multiple data values
`in a prototype astrophysics simulation, provided they
`meet certain basic thresholds of perceptibility (Huber
`and Healey, 2005). A technique termed “motion high-
`lighting” explores the potential applicability of mo-
`tion to node-link diagrams (Ware and Bobrow, 2004).
`Results of motion highlighting experiments indicate
`that the translating or scaling of node is more useful
`for supporting rapid interactive queries on node-link
`diagrams than static highlighting methods.
`Dynamic textures are sequences of images that
`exhibit some form of temporal coherence, or more
`specifically, they are individual images that are “re-
`alizations of the output of a dynamical system driven
`by an independent and identically distributed pro-
`cess” (Doretto et al., 2003). Dynamic textures have
`been effectively used in scientific visualizations and
`have been more extensively investigated in computer
`graphics and computer vision research. For instance,
`(Van Wijk, 2002) uses an iterative series of texture
`distortions to represent fluid flows, and (Forbes and
`Odai, 2012; Forbes et al., 2013) applies this tech-
`nique to creative media arts projects. Work by (Lum
`et al., 2003) explores adding moving particles to a
`surface texture of a static object in which the parti-
`cles are placed along the principal curvature direction
`to better indicate the object’s shape and spatial rela-
`tionships. Within a more general computer graphics
`context, dynamic textures are used in a variety of ap-
`plications. For instance, dynamic textures have been
`used as a computationally efficient way to add realism
`to a scene. (Chuang et al., 2005) presents an interac-
`tive system that procedurally generates dynamic tex-
`tures from selected components of a single image that
`can then be added to a scene. Similarly, a variety of
`
`techniques have been introduced to automatically cre-
`ate “temporal textures” from a single image in order
`to mimic natural phenomenon such as clouds, water,
`and fire (Lai and Wu, 2007; Ruiters et al., 2010; Ok-
`abe et al., 2011). In addition to having the potential
`to be used as an effective modality for representing
`quantitative information, recent research has explored
`the use of dynamic textures as a medium for provid-
`ing semantically contextualized information. (Lock-
`yer et al., 2011) explores the “expressive scope” of
`ambient motion textures for “emphasis and more sub-
`tle ambient visualization.” In particular, this research
`focused on the effective communication of particu-
`lar emotions through the use of intrinsic motion cues
`within a dynamic texture.
`For the most part, research on motion in informa-
`tion visualization is concerned with extrinsic motion,
`or at least does not differentiate between extrinsic and
`intrinsic motion. For instance, (Ware and Bobrow,
`2004), also cited above, discusses a motion highlight-
`ing technique whereby the animation of a station-
`ary link generate a “crawling” motion. Although it
`is not presented specifically as a dynamic texture, it
`is clear that this “crawling” motion is of a different
`nature than the translation patterns used to highlight
`nodes. A recent evaluation found that animated rep-
`resentations were more effective than almost all static
`representations of link representations (Holten et al.,
`2011).
`It seems reasonable that other visualization
`systems could utilize a conflation of textures and mo-
`tion; rather than attempting to procedurally generate
`dynamic textures, we could gather them directly. This
`would have the immediate advantage that they were,
`at least in some degree, inherently non-distracting for
`the simple reason that they occur continually in the
`real-world. An earlier (unpublished) study by one of
`the authors found that the use of moving sinusoidal
`gratings introduced visual fatigue precisely because
`of the qualities that made it unrealistic: its fixed ro-
`tation, its predictable frequency and amplitude, and
`its repetitive sequence of pixel values.
`instead, Dy-
`namic textures using fluctuating, intrinsic, real-world
`motion are cohesive without being repetitive; differ-
`entiable without being distracting.
`In order to expedite the creation of intrinsic mo-
`tion textures in order to analyze their potential effec-
`tiveness in visualization systems, we gathered real-
`world video of natural phenomena containing intrin-
`sic motion: fire, water, clouds, etc. Although we be-
`lieve that these textures have a representational com-
`ponent that might be useful in some visualization cir-
`cumstances, for this study we isolated the phenomena
`using these steps: (A) record the natural phenomenon;
`(B) crop the resulting video; (C) desaturate the video;
`
`Page 3 of 7
`
`

`

`The variability metrics reflect the spatial variation
`in a single video frame. Let Q′ (cid:26) Q be the set of
`pixel positions in the interior of the grid. Each pixel
`position q 2 Q′ therefore has eight spatially adjacent
`neighbors, the set of which we denote N(q). The
`roughness R is defined as the average intensity dif-
`ference of the highest-contrast neighbors:
`Definition 2 (Roughness).
`T(cid:229)
`(cid:1) 1
`(cid:229)
`T
`q2Q′
`Edginess Eq is the average number of large-
`contrast juxtapositions, per pixel, per frame. The con-
`trast is regarded as large if the intensity difference is
`at least q:
`Definition 3 (Edginess).
`T(cid:229)
`(cid:1) 1
`1
`(cid:229)
`W H
`T
`q2Q′
`t=1
`In the current study, we set the threshold value q equal
`to one-eighth the maximum contrast range of the rele-
`vant collection (G20, G40, G60, or G80). For example,
`when comparing two videos from group G40, we used
`the E40=8 = E5 edginess metric.
`Our flicker metrics depend on the local maxima
`and minima (peaks and valleys) of pixel intensity in
`the time domain. To specify them formally, we intro-
`duce definitions of peak and valley as follows.
`Definition 4 (Peak intensity value). Iq;t is a peak in-
`tensity value of width j, provided
`(cid:15) j > 0,
`(cid:15) 1 < t (cid:20) T (cid:0) j,
`(cid:15) Iq;t(cid:0)1 < Iq;t = Iq;t+1 = (cid:1)(cid:1)(cid:1) = Iq;t+ j(cid:0)1, and
`(cid:15) Iq;t+ j(cid:0)1 > Iq;t+ j.
`this definition is consistent
`Though cumbersome,
`with an intuitive notion of local maximum. We also
`define valley intensity value analogously for local
`minima. Let nq denote the number of peaks of width
`1 or more at position q. Let p1;q; p2;q; : : : ; pnq;q be the
`peak intensity values at position q, such that pi;q is the
`intensity value of the ith peak, in chronological order.
`Similarly, let v1;q;v2;q; : : : ;vnq;q denote the valley in-
`tensity values (assuming they are equally numerous
`as the peaks). Now we can precisely state our met-
`rics for flicker. First, the average number of peaks per
`pixel, per frame, is called local frequency, FL:
`Definition 5 (Local frequency).
`(cid:1) 1
`1
`W H
`T
`
`nq:
`
`(cid:229) q
`
`2Q
`
`FL =
`
`R =
`
`1
`W H
`
`jIq;t (cid:0) Is;tj:
`
`max
`s2N(q)
`
`t=1
`
`Eq =
`
`1ft : q(cid:20)maxs2N(q)
`
`jIq;t(cid:0)Is;tjg(t):
`
`(D) constrain the pixel range of the video; (E) pixelate
`the video; (F) apply temporal smoothing to the video.
`The video is thus transformed into a low-contrast, de-
`saturated, pixelated, and mostly unrecognizable ver-
`sion of itself that nonetheless retains important qual-
`ities of the original natural motion. Figure 1 shows
`a frame from a naturalistic video as it is processed
`through this pipeline. These intrinsic motion textures
`can then be defined metrically, in terms of particular
`features, and included in studies where we can asso-
`ciate these features with empirical observations, such
`as discernibility and differentiability.
`
`2 Metric definitions
`
`Much research has been done to develop ways to ef-
`fectively and efficiently characterize motion. A com-
`monly used method, optical flow, assumes that there
`is a unique velocity vector at each pixel. Other meth-
`ods relax that assumption. For instance, (Langer and
`Mann, 2003) introduces “optical snow,” which is able
`to characterize motions that have prevalent disconti-
`nuities between frames. The intrinsic motions that
`we have gathered likewise include large amounts of
`flickering that are not captured with optical flow type
`analyses. Since we are, for now, interested primar-
`ily in the single task of determining the ease of dis-
`crimination betweens motions, we constructed a way
`to create simpler metrics that define a video via a set
`of eight features that sufficiently characterize intrinsic
`motions. They are grouped into the following cate-
`gories: flicker, flutter, and variability.
`We now introduce notation that defines our met-
`rics precisely. Let Iq;t denote the integer intensity
`value of a pixel in the video, at spatial position q
`and at time t. Assuming the video pixels lie in a
`rectangular grid of width W , height H, and that the
`video has duration of T discrete video frames, q 2
`Q = f1; : : : ;Wg(cid:2)f1; : : : ;Hg and t 2 f1; : : : ;Tg. The
`first characteristic of interest we define is the contrast
`{
`{
`}
`range K of a video:
`Definition 1 (Contrast range).
`K = max
`Iq;t
`q2Q
`q2Q
`1(cid:20)t(cid:20)T
`1(cid:20)t(cid:20)T
`In the present study, instead of using K as a video
`feature for comparison, we partitioned our test videos
`into collections with of similar contrast range, be-
`cause two videos with widely differing contrasts are
`very obviously distinct. Specifically, we gathered
`videos with contrast ranges of 20 or less, 21 to 40,
`41 to 60, and 61 to 80, into groups called G20, G40,
`G60, and G80, respectively.
`
`Iq;t
`
`:
`
`}(cid:0) min
`
`Page 4 of 7
`
`

`

`Again, in our study we set the this metric’s threshold
`just as those of Defs. 3 and 7. So, when comparing
`videos from group G40, we compute metric CG;5.
`These metrics are easy to calculate and capture
`both the global and the local, pixel level aspects of
`the videos. They may not however capture larger-
`scale movement within the motion textures. However,
`since our express aim is to use motion textures that do
`not contain large-scale motion, we believe that these
`metrics are appropriate as a first attempt to character-
`ize intrinsic motions.
`
`3 User Study
`
`The main goal of our preliminary user study was
`to determine the minimum amount of movement re-
`quired in order for a participant to quickly differen-
`tiate between similar motions. Since motion can be
`highly distracting and since humans are exception-
`ally good at noticing differences in motion, by finding
`lower bounds on various parameters that make motion
`distinguishable we can identify the minimum values
`of easily-discernible features. Future work will use a
`more rigorously defined empirical study using tech-
`niques to measure just-noticeable difference, as well
`as explore user response to visualization tasks incor-
`porating motion textures. For this preliminary study
`we wanted to obtain an initial sense of what attributes
`were most easily noticeable at low-contrast ranges,
`and which of these attributes were thought to be the
`least distracting.
`To find this minimum feature set, we created a
`study that presented the participant with a pair of
`videos. The user was then asked to indicate whether
`he or she agreed or disagreed with a series of state-
`ments about the videos. We gathered 32 unique
`videos of naturalistic motion and processed them as
`described in section 3.1. We created 4 “bins” and
`made versions of each of these videos with different
`levels of contrast. Bin 1 contained videos with a con-
`trast range of +/- 10; Bin 2, +/- 20; Bin 3, +/- 30; Bin
`4, +/- 40. For Bin 1, it was very difficult to tell most of
`the videos apart, especially when looking at a single
`(unmoving) frame from the video. In other words, the
`contrast was so low that without movement it would
`be almost impossible to tell them apart. For Bin 2,
`it seemed that about half of the time it was easy to
`tell the videos apart and the other half of the time it
`was difficult. For Bin 3, it became much easier to tell
`any of the videos apart from any of the others. And
`finally, for Bin 4 it was easy to tell all of the videos
`apart. However, we thought that if the movements be-
`came more chaotic (higher absolute flicker amplitude
`
`We define the average peak-to-valley difference,
`per pixel, as local amplitude, AL:
`Definition 6 (Local amplitude).
`nq(cid:229)
`(pi;q (cid:0) vi;q);
`
`i=1
`
`1 n
`(cid:229) q
`
`q
`
`1
`W H
`
`AL =
`
`2Q
`if nq > 0, otherwise AL = 0.
`We define local choppiness, CL;q, as the average
`number of large intensity jumps per pixel, per frame.
`A jump at position q is large if equals or exceeds
`threshold value q.
`Definition 7 (Local choppiness).
`T(cid:229)
`(cid:1) 1
`(cid:229)
`T
`q2Q
`Similar to our use of Def. 3, in our study we set
`the threshold value here to one-eighth of the contrast
`range of whichever videos we are comparing. So
`when comparing videos from group G40, we use lo-
`cal choppiness metric CL;5.
`Our flutter metrics are similar to the flicker met-
`rics, but they depend on the average intensity of the
`entire video frame at a given moment. Let Jt denote
`the average pixel intensity of the frame at time t, i.e.,
`W H (cid:229)q2Q Iq;t. This sequence also has peaks and
`Jt = 1
`valleys, which we assume are m in number. At the
`risk of confusion, we will denote its peak values and
`valley values (regardless of width) as p1; p2; : : : ; pm
`and v1;v2; : : : ;vm respectively, in chronological order.
`(Note that these values have only one subscript.) The
`average number of these peaks, per frame, is called
`global frequency, FG:
`Definition 8 (Global frequency).
`
`CL;q =
`
`1
`W H
`
`1ft : jIq;t(cid:0)Iq;t(cid:0)1j(cid:21)qg(t)
`
`t=2
`
`FG = m=T:
`
`The average of these peak-to-valley differences is
`the global amplitude, AG:
`Definition 9 (Global amplitude).
`m(cid:229)
`(pi (cid:0) vi);
`
`AG =
`
`1 m
`
`i=1
`if m > 0, otherwise AG = 0.
`We define global choppiness CG;q as the average
`number of large increases in average intensity, with
`threshold q:
`Definition 10 (Global choppiness).
`T(cid:229)
`
`1ft : jJt(cid:0)Jt(cid:0)1j(cid:21)qg(t):
`
`t=2
`
`1 T
`
`CG;q =
`
`Page 5 of 7
`
`

`

`and frequency) then in those cases the videos in Bin
`3 and Bin 4 would be hard to tell apart. We did not
`test any of the videos against a video with a different
`range of pixel values as it is easy to discern the dif-
`ferences in videos when one had a higher maximum
`and lower minimum pixel value than the other. Each
`of the videos was analyzed with custom software that
`output the features described by our metrics system.
`We further calculated the absolute difference between
`the feature vectors of each video.
`We included a series of four Likert items per test
`designed to elicit the participant’s opinion about the
`discernibility of flicker and frequency. We ran vari-
`ous “batches” of our test over the course of a week
`and a half on Amazon Mechanical Turk. We received
`a total of 144 completed studies. For most of these
`batches, we randomly chose one of the 4 bins for each
`test. The majority of our “workers,” 107, indicated
`that they were from India; 24 were from the United
`States; the rest came from the United Kingdom, Mex-
`ico, Sri Lanka, Canada, Pakistan, and Nigeria. 2 par-
`ticipants chose “Other” as their nationality. There
`were an equal number of male and female participants
`(72 each). The minimum and maximum age was 19
`and 63, respectively, with a median age of 31. Fol-
`lowing the suggestions in (Heer and Bostock, 2010),
`which describes some of the advantages and disad-
`vantages of conducting studies via Mechanical Turk,
`we made a substantial effort to encourage reliable par-
`ticipation and mitigate inaccurate or random answers,
`ultimately obtaining 476 samples from the 144 partic-
`ipants.
`All features, except for the frequency of the flutter
`(the global frequency of a direction change in aver-
`age pixel value for a frame) were positively correlated
`with easy differentiability. We built a statistical model
`to characterize the relationship between the video mo-
`tion metrics and the participants’ responses. We fo-
`cused on contrast ranges 20 and 40 and modeled the
`data as a two-category classification problem: in each
`video comparison, the videos are either difficult to
`distinguish (category CD) or not (category CE ), gen-
`erated by the user’s Likert responses. For our binary
`classifier, any value greater than 2 was given a placed
`in category CE, and any value less than or equal to 2
`was placed in category CD. Ideally, the model would
`effectively predict the category for a video compar-
`ison, based only on our video features. We modeled
`these two categories by assuming a multivariate Gaus-
`sian distribution of the feature vectors (which are the
`absolute differences of each of the eight metrics for
`the pair of videos being compared). In other words,
`we computed the maximum-likelihood mean vector
`µC;r and covariance matrix SC;r of all feature vec-
`
`tors for category C 2 fCD;CEg, and contrast range
`r 2 f20;40g. For the purpose of classification, we also
`make use of the empirical frequency of CD and CE
`classes, denoted p(CD) and p(CE ). Given a new data
`vector x for contrast range r, we would classify it in
`category CD provided it satisfies p(CDjx) > p(CEjx),
`(
`)
`where by Bayes’ theorem, for C 2 fCD;CEg,
`(
`)
`x;µC;r;SC;r
`N
`p(C)
`p(Cjx) =
`x;µB;r;SB;r
`(cid:229)B2fCD;CEg N
`p(B)
`In the above, N (x;µ;S) denotes the probability den-
`sity function for the multivariate Gaussian with mean
`µ and covariance S. One advantage of a Gaussian
`characterization is that we can easily marginalize any
`subset of features. Thus we can see the average in-
`teraction between any two features and can list the
`thresholds for the classifier with all other features
`marginalized (Table 1). In particular, even small dif-
`ferences in flickering (especially in the frequency and
`choppiness) at the individual pixel level were the main
`predictors of whether or not a video pair was likely to
`be easily differentiable.
`
`:
`
`Table 1: Threshold values between features.
`l chop
`0.0011901
`l amp
`0.12419
`l freq
`0.00081854
`g chop
`0.048523
`g amp
`0.037790
`g freq
`0.0056176
`rough
`0.57185
`edge
`0.0033416
`
`4 Conclusion
`
`This paper presents an initial foray into exploring
`the potential usefulness of intrinsic motion textures.
`We provide a method for generating cohesive, non-
`repetitive, intrinsic motions textures from real-world
`video captures; a method for characterizing the fea-
`tures of intrinsic motions; a preliminary user study
`that indicates minimal differences necessary for dif-
`ferentiation between motions; and an analysis of this
`study that identifies thresholds on these features. This
`initial exploration of motion textures created from
`video captures of naturalistic movement seems to in-
`dicate that this may be a promising area for future in-
`vestigations. Future work will involve the design and
`analysis of more rigorous empirical studies to deter-
`mine the validity of our claims regarding the notice-
`ability and distraction of these types of textures.
`
`Page 6 of 7
`
`

`

`REFERENCES
`
`Bartram, L. and Ware, C. (2002). Filtering and brushing
`with motion. Information Visualization, 1(1):66–79.
`Bartram, L., Ware, C., and Calvert, T. (2003). Moticons:
`detection, distraction and task. International Journal
`of Human-Computer Studies, 58(5):515–545.
`Bertin, J. (2010). Semiology of graphics: diagrams, net-
`works, maps, pages 79–81. Esri Press.
`Bostock, M., Ogievetsky, V., and Heer, J. (2011). D3: Data-
`IEEE Transactions on Visualiza-
`driven documents.
`tion and Computer Graphics (TVCG), 17(12):2301–
`2309.
`Byelas, H. and Telea, A. (2009). Visualizing multivariate
`attributes on software diagrams. In Software Mainte-
`nance and Reengineering, 2009. CSMR ’09. 13th Eu-
`ropean Conference on, pages 335 –338.
`Chalupa, L., Werner, J., and of Technology, M. I. (2004).
`The visual neurosciences. MIT Press.
`Chuang, Y., Goldman, D., Zheng, K., Curless, B., Salesin,
`D., and Szeliski, R. (2005). Animating pictures with
`stochastic motion textures. ACM Transactions on
`Graphics (TOG), 24(3):853–860.
`Conti, G., Ahamad, M., and Stasko, J. (2005). Attacking in-
`formation visualization system usability overloading
`and deceiving the human. In Proceedings of the 2005
`symposium on Usable privacy and security, pages 89–
`100. ACM.
`Doretto, G., Chiuso, A., Wu, Y. N., and Soatto, S. (2003).
`Dynamic textures. International Journal of Computer
`Vision, 51:91–109. 10.1023/A:1021669406132.
`Emrith, K., Chantler, M., Green, P., Maloney, L., and
`Clarke, A. (2010). Measuring perceived differences in
`surface texture due to changes in higher order statis-
`tics. JOSA A, 27(5):1232–1244.
`Forbes, A. G., H¨ollerer, T., and Legrady, G. (2010). Be-
`haviorism: A framework for dynamic data visualiza-
`IEEE Transactions on Visualization and Com-
`tion.
`puter Graphics (TVCG), 16(6):1164–1171.
`Forbes, A. G., H¨ollerer, T., and Legrady, G. (2013). Genera-
`tive fluid profiles for interactive media arts projects. In
`Proceedings of the International Symposium on Com-
`putational Aesthetics in Graphics, Visualization, and
`Imaging (CAe), pages 123–129, Anaheim, California.
`Forbes, A. G. and Odai, K. (2012). Iterative synaesthetic
`In Proceed-
`composing with multimedia signals.
`ings of the International Computer Music Conference
`(ICMC), pages 573–578, Ljubjiana, Slovenia.
`Heer, J. and Bostock, M. (2010). Crowdsourcing graphical
`perception: using mechanical turk to assess visualiza-
`In ACM Human Factors in Computing
`tion design.
`Systems (CHI), pages 203–212. ACM.
`Holten, D., Isenberg, P., van Wijk, J., and Fekete, J. (2011).
`An extended evaluation of the readability of tapered,
`animated, and textured directed-edge representations
`in node-link graphs. In Pacific Visualization Sympo-
`sium (PacificVis), 2011 IEEE, pages 195 –202.
`House, D., Bair, A., and Ware, C. (2006). An approach
`to the perceptual optimization of complex visualiza-
`
`tions. IEEE Transactions on Visualization and Com-
`puter Graphics (TVCG), 12(4):509–521.
`Huber, D. and Healey, C. (2005). Visualizing data with
`motion. In Proceedings of IEEE Visualization (VIS),
`pages 527–534.
`Interrante, V. (2000). Harnessing natural textures for multi-
`variate visualization. Computer Graphics and Appli-
`cations, IEEE, 20(6):6 –11.
`Interrante, V. and Kim, S. (2001). Investigating the effect
`of texture orientation on the perception of 3d shape.
`In Human Vision and Electronic Imaging VI, volume
`4299, pages 330–339.
`Kim, S., Hagh-Shenas, H., and Interrante, V. (2004). Con-
`veying shape with texture: Experimental investiga-
`tions of texture’s effects on shape categorization judg-
`ments. IEEE Transactions on Visualization and Com-
`puter Graphics (TVCG), 10(4):471–483.
`Kujala, J. and Lukka, T. (2003). Rendering recognizably
`unique textures. In In

This document is available on Docket Alarm but you must sign up to view it.


Or .

Accessing this document will incur an additional charge of $.

After purchase, you can access this document again without charge.

Accept $ Charge
throbber

Still Working On It

This document is taking longer than usual to download. This can happen if we need to contact the court directly to obtain the document and their servers are running slowly.

Give it another minute or two to complete, and then try the refresh button.

throbber

A few More Minutes ... Still Working

It can take up to 5 minutes for us to download a document if the court servers are running slowly.

Thank you for your continued patience.

This document could not be displayed.

We could not find this document within its docket. Please go back to the docket page and check the link. If that does not work, go back to the docket and refresh it to pull the newest information.

Your account does not support viewing this document.

You need a Paid Account to view this document. Click here to change your account type.

Your account does not support viewing this document.

Set your membership status to view this document.

With a Docket Alarm membership, you'll get a whole lot more, including:

  • Up-to-date information for this case.
  • Email alerts whenever there is an update.
  • Full text search for other cases.
  • Get email alerts whenever a new case matches your search.

Become a Member

One Moment Please

The filing “” is large (MB) and is being downloaded.

Please refresh this page in a few minutes to see if the filing has been downloaded. The filing will also be emailed to you when the download completes.

Your document is on its way!

If you do not receive the document in five minutes, contact support at support@docketalarm.com.

Sealed Document

We are unable to display this document, it may be under a court ordered seal.

If you have proper credentials to access the file, you may proceed directly to the court's system using your government issued username and password.


Access Government Site

We are redirecting you
to a mobile optimized page.





Document Unreadable or Corrupt

Refresh this Document
Go to the Docket

We are unable to display this document.

Refresh this Document
Go to the Docket