SlideShare a Scribd company logo
1 of 8
Download to read offline
Understanding the Benefits of Gaze Enhanced Visual Search
                                        Pernilla Qvarfordt, Jacob T. Biehl, Gene Golovchinsky, Tony Dunningan
                                                              FX Palo Alto Laboratory, Inc.
                                                      3400 Hillview Bldg. 4, Palo Alto, CA 94304
                                                       {pernilla, biehl, gene, tonyd}@fxpal.com




Abstract                                                                                           For instance, a study investigating the utility of triple examination
                                                                                                   showed error rates still at 11% [Markus et al. 1990]. Error rates go
                                                                                                   even higher when non-skilled examiners (e.g., MDs other than
In certain applications such as radiology and imagery analysis, it is
                                                                                                   Radiologists) are involved [Shaw et al. 1990]. A central recom-
important to minimize errors. In this paper we evaluate a struc-
                                                                                                   mendation from many of these studies is to employ a systematic
tured inspection method that uses eye tracking information as a
                                                                                                   methodology in how images are examined [Goddard et al. 2001].
feedback mechanism to the image inspector. Our two-phase
                                                                                                   That is, there is a need for a more formal, structured analysis meth-
method starts with a free viewing phase during which gaze data is
                                                                                                   odology that ensures full examination of all anatomical compo-
collected. During the next phase, we either segment the image,
                                                                                                   nents [Vock 1987].
mask previously seen areas of the image, or combine the two tech-
niques, and repeat the search. We compare the different methods
proposed for the second search phase by evaluating the inspection                                  Similar recommendations have been made for other critical task
method using true positive and false negative rates, and subjective                                domains, such as transportation security. Many of the recommen-
workload. Results show that gaze-blocked configurations reduced                                    dations call for changing the current methodology from only
the subjective workload, and that gaze-blocking without segmenta-                                  requiring secondary screening on potential true positives to requir-
tion showed the largest increase in true positive identifications and                              ing mandatory multi-stage screening [Butler and Poole 2002,
the largest decrease in false negative identifications of previously                               Fenga et al. 2008]. While false negative statistics are not openly
unseen objects.                                                                                    published for the domain, such recommendations would not be
                                                                                                   made if there was not a meaningful need to reduce false negative
                                                                                                   error rates.
CR Categories: H.5.2 [Information Interfaces and Presentation]:
User Interfaces
                                                                                                   In this paper, we present a novel methodology for performing
                                                                                                   structured image analysis that is both systematic and adaptive to an
Keywords: gaze-enhanced visual search, two-phase search, multi-
                                                                                                   examiner’s search behavior. The goal of the methodology is to pro-
ple targets.
                                                                                                   vide a more structured image examination process. Specifically,
                                                                                                   we propose a two-phase process that consists of an initial phase of
1     Introduction                                                                                 free search, followed by a second phase designed to help the exam-
                                                                                                   iner to cover the whole image. This second phase can consist either
Human-based image analysis is a commonly performed task in                                         of the image being divided into smaller sub-images (segments), of
many contemporary professions. For instance, radiologists and                                      having previously viewed (fixated) segments blocked out, or of a
other medical professionals frequently examine medical images to                                   combination of these techniques. This methodology was designed
diagnose and treat patients, airport security agents scan x-rays of                                to increase the total coverage of examination while reducing
luggage for prohibited items, and factory workers perform visual                                   redundant examination with the goal of reducing the overall num-
inspection of goods to assure quality. In these tasks, the examiner                                ber of false negatives, particularly false negatives that the exam-
must combine knowledge of the domain and a high degree of men-                                     iner did not view or judge. Our proposed methodology is not
tal concentration within a short amount of time to classify or inter-                              domain dependent; we believe it can be applied to many domains,
pret the images.                                                                                   including security images, satellite images, maps, astronomical
                                                                                                   images, and scientific visualizations.
While there have been many advances both in the technology used
to create images and in the training of human examiners, many                                      We evaluated the inspections techniques proposed for the struc-
image analysis tasks are still prone to significant error. For exam-                               tured viewing in a controlled laboratory study. The results shows
ple, Goddard et al. [2001] showed that radiological image exami-                                   that masking previously well inspected areas reduced the subjec-
nation still has rates nearing 20% for clinically significant or major                             tive workload. When the previous gaze is masked over the whole
errors. Even with multiple examination (multiple radiologists                                      image, the increase of true positive identifications and reduction of
investigating an image or set of images), the error rate is still high.                            false negative identifications are larger than in the other combina-
                                                                                                   tions. We believe these results can provide insights and design les-
                                                                                                   sons for the construction of new visual search systems and tech-
                                                                                                   niques that improve search performance.
.
Copyright © 2010 by the Association for Computing Machinery, Inc.
Permission to make digital or hard copies of part or all of this work for personal or
classroom use is granted without fee provided that copies are not made or distributed              2    Related work
for commercial advantage and that copies bear this notice and the full citation on the
first page. Copyrights for components of this work owned by others than ACM must be
honored. Abstracting with credit is permitted. To copy otherwise, to republish, to post on         Research on visual search have been pursued in two directions,
servers, or to redistribute to lists, requires prior specific permission and/or a fee.             either to investigate the perceptual and cognitive processes under-
Request permissions from Permissions Dept, ACM Inc., fax +1 (212) 869-0481 or e-mail
permissions@acm.org.
                                                                                                   lying visual search, or to design procedures or tools for improving
ETRA 2010, Austin, TX, March 22 – 24, 2010.
© 2010 ACM 978-1-60558-994-7/10/0003 $10.00

                                                                                             283
the performance of professional image analysts. Our research falls             looking for suspected tumors in an X-ray, or for camflaged weap-
in the latter category, but findings from perceptual and cognitive             ons in a satellite image.
studies provide additional context.
                                                                               2.2    Improving Visual Inspection
2.1    Visual Search
                                                                               Visual inspection is important in many domains such as finding
Characteristics of visual search have been investigated extensively            tumors in X-rays and identifying threats in airport luggage scans.
in cognitive psychology and perception. This line of research is               Yet, visual inspection is error prone. Most attempts to lower the
focused on building models of perception and of higher-level cog-              error has focused on training, see for example [Kollera, Drury, and
nitive processes. Although visual search studies are often labora-             Schwaninger, 2009]. Nickles, Melloy and Gramopadhye [2003]
tory studies removed from real world image search tasks, there are             used dynamic tool for training inspectors to follow a specific scan-
findings that carry over to the work of professional image analysts.           ning pattern during the inspection, but they could not show that the
                                                                               dynamic training was better than verbal instructions. Another
When searching for an item in an image, the eye scans the image to             study [Sadasivan et al. 2005] used gaze information from experts
find the item. The character of these scanpaths has caused a long              to create a visualization of a preferred scanpath for cargo bay
debate within psychology. Both early observations, e.g. [Norton                inspection. This visualization was used to train novice inspectors
and Stark 1971] and more recent studies, e.g. [Scinto, Pillalamarri,           and proved to be successful in improving performance.
and Karsh 1986], have characterized the scanpaths as random.
Although people tend to look at similar things in the image, people
                                                                               Rather than focus solely on training, others have focused on sup-
rarely follow the same scanpath, i.e., they look at different parts of
                                                                               porting the inspection task. For instance, Haimson et al. [2004]
the image in different order.
                                                                               added partitioning to a radar screen and thereby improved perfor-
                                                                               mance in finding targets. Forlines and Balakrishnan [2009] used
However, scanpaths do not always appear random. Findlay and
                                                                               image segmentation to separate objects in the image and to recom-
Brown [2006] showed that people can employ different strategies
                                                                               pose the objects. Techniques for recomposing the image showed
when searching through images with randomly placed targets. One
such strategy, for example, is to follow the global external counter           promise in reduced error rate, but longer search time was also
that the objects create. When the displays are arranged in a more              found for two of the recomposition techniques. Image segmenta-
structured ways, e.g. in a grid pattern, people tend to read the grid          tion as an underlying technique may work well on some kinds of
as they would read text [Gilchrist and Harvey 2006]. The practical             targets. Forlines and Balakrishnan focused on blood cell inspec-
implications from scanpath theories is that expert inspectors                  tion; other type of images decomposition may not be as easy to
appear to develop structured scanpaths [Sadasivan et al. 2005].                decompose as cell slides used in their experiment.
There is, however, evidence from a luggage-screening experiment
that performance gains appear to be related to changes in ability to           As mentioned above, Nodine and Kundel [1987] found that the
recognize targets rather than to changes in scanpaths [Mccarley et             fixation times for false positive responses were lower than those of
al. 2004].                                                                     true positive when inspecting chest X-rays. They used this knowl-
                                                                               ege in a system which collected gaze information on where radiol-
There is also evidence that the characteristics of fixation change             ogists looked during a free form inspection. In a second phase,
depending on judgments viewers make of the items in view. Pom-                 they asked the radiologists to re-inspect areas they had dwelled on
plun, Reingold and Shen [2001] found that people who were asked                for an extended time, but in which they did not find any incidence
to match two sets of images fixated for a longer time on matching              of tumors. In that work, participants were not asked to view areas
objects compared to mismatched objects. Manning, Ethell and                    they had not looked at previously, but only to re-evaluate already
Donovan [2004] showed that when radiologists inspected chest x-                looked at areas.
rays, their gaze duration was half as long on lesions they did not
report as a detected tumor as on lesions they reported. Another                3     Inspection Methodologies
study [Nodine and Kundel 1987] reported longer gaze durations
when giving a true positive judgment about a lesion compared to a
false negative judgment. These results suggest that it might be pos-           In this work we compare four different inspection methodologies
sible to use gaze durations to provide feedback to the user of which           for visual search. These methodologies follow a two-phase inspec-
parts of an image have been well inspected and which parts have                tion approach. In the first phase, searchers are allowed to view the
not.                                                                           image unconstrained, enabling them to gain an understanding of
                                                                               the image and its gross structure. In the second phase, the view of
One interesting finding is that people have a higher error rate when           the image is restricted to create a more structured search. Below
targets are rare than when targets occure frequently [Wolfe et al.             we describe each methodology studied and the rationale for its
2007]. A practical example of this is that when a radiologist                  selection.
screens mammograms for cancer, very few mammograms will
show any signs of tumors. Research in low prevalence has sug-                  Structured Segmentation. Inspired by previous work that found
gested several ways to counter the effect, such as rewarding image             performance benefits in systematic search of images [Sadasivan et
searchers appropriately for finding targets [Navalpakkam, Kock                 al. 2005] and [Nickles et al 2003], this technique subdivides the
and Perona 2009], and allowing people to go back and correct mis-              image into a tiled grid and then displays the grid tiles serially
takes [Fleck and Mitroff 2007].                                                (Figure 1). By reducing the viewable search area at any one time,
                                                                               this technique forces the user to focus attention on the currently-
In most of perception studies that explore low prevalence of tar-              visible area. It also ensures that all parts of the display get equal
gets, participants are asked to find only one target. While the over-          attention and reduces distractions from other parts of the display.
all true positive rate maybe low, targets often do not exist in isola-         In our study, we divided each image into a three by three grid,
tion. That is, there are likely to be many true positives in a single          keeping the area small enough to be able to quickly scan and mark
image, all of which need to be identified to correctly perform the             during the time limit set for the experiment, and large enough to
search task. Examples of such recall-oriented activity include                 contain information when gaze-blocking was also used.



                                                                         284
Phase I                                Phase II                               Phase I                               Phase II

                Figure 1: Structured segmentation                                    Figure 3: Structured segmentation with gaze-blocking




           Phase I                                                                        Phase I                                Phase II
                                                  Phase II


                      Figure 2: Gaze blocking                                                       Figure 4: Full-image search

Gaze-Blocking. This methodology uses a searcher’s gaze behavior
from the first free-viewing search phase to determine where the                4.1    Method
searcher has previously inspected the image and blocks out those               We used a 2x2 within-subjects design, where the independent vari-
areas in the second phase to prevent unnecessary re-examination                ables were type of image display (image display), and visibility of
(Figure 2). The blocked-out areas are constructed from clusters of             fixation clusters (gaze display). Each of the two independent vari-
identified fixations from the first phase using a dispersion-based             able had two levels for image display: full image or segmented
fixation detection algorithm [Salvucci and Goldberg 2004]. Fixa-               image, and for gaze display: with gaze-blocking or without gaze-
tions are clustered using a minimum-spanning tree algorithm                    blocking. This design gave us four conditions for displaying the
based on their distance to other fixations. Only fixations clusters            image in phase 2: full image with previous gaze blocking (FB),
with a total gaze duration of more than 270 ms were blocked out.               segmented image with gaze-blocking (SB), full image without
We based this threshold on 275 ms fixations for image search                   gaze-blocking (F), and segmented image without gaze-blocking
reported in [Rayner 1998], and based on our own experimentation.               (S). These conditions directly match the proposed inspection meth-
                                                                               odologies discussed in the previous section.
This threshold ensured that fixation clusters reflected some higher
level of cognitive processes. Each fixation in the clusters is repre-
                                                                               4.2    Participants
sented as a 40 pixel circle in our system, centered at the gaze loca-
tion. This size was chosen as it represents the 1 visual degree of the         Seven men and one woman participated in the study. Participants
fovea at the distance of 60 cm from the computer screen using 96               were recruited using a broadcast email solicitation within our orga-
pixels per inch. Each blocked-out region was replaced with a white             nization. Participants’ age ranged between 30 and 60 years. All
patch over the image.                                                          were screened before the study to make sure they had normal or
                                                                               corrected vision by contact lenses (4 participants wore contact
Structured Segmentation and Gaze-Blocking. This technique com-                 lenses) and that the eye tracker could properly track their eyes.
bines the previous two methodologies. The image is re-presented
in segments, each segment containing gaze blocks of previously                 4.3    Stimuli and Tasks
inspected areas (Figure 3). This retains the advantages of struc-              Careful consideration was given to the design of the task stimuli.
tured search while eliminating redundant examination.                          Motivated by the fact that most inspection tasks are not only per-
                                                                               ceptual but also involve higher-level cognitive processes, we
Full Image. In addition to these three methodologies, we also                  wanted to simulate a task that involved not only perception of tar-
included a baseline condition in which the full image was re-pre-              gets, but also associated judgment. Furthermore, we did not want
sented to the searcher (Figure 4).                                             to require specific domain knowledge to avoid introducing
                                                                               domain-specific and participant-specific effects.
4    Experiment
                                                                               We designed two types of targets and distractors: symmetric and
                                                                               asymmetric (see Figure 5 shapes in black outline). For symmetric
We conducted an experiment to evaluate the inspection methodol-                targets a mirror image of the target is indistinguishable from the
ogies. We designed a controlled experiment to explore the effects              target. Asymmetric targets on the other hand, have a mirror image
of using both segmentation and occlusion of well-inspected areas.              that when rotated do not match the target (e.g. in Figure 5 first gray



                                                                         285
shape to left of the assymetric target). Symmetric targets differed
from distractors by the size of the insections (notches shown in
Figure 5) in the basic shape. To identify symmetric targets, partici-
pants needed to judge details of the target. In Figure 5 for example,
the participants need to judge if the circular insection was of the
same size as the example target.To identify asymmetric targets,
participants needed to perform mental rotation of potential targets.

Each target had one to three similarly looking distractors (close
distractors, targets with gray outline in Figure 5). Targets were ran-
domly assigned to locations where they would not overlap other
targets or distractors. Targets and close distractors were not placed
on segmentation boundries. Besides close distractors, stimulus
images also include a random collection of distractors (random
distractors). Targets and distractors were randomly rotated in steps
of 30 degrees, starting a 0 degrees..


                                                                                     Figure 6: Example of stimuli image used in the study.
           symmetric                     asymmetric                            set to 100 ms. In addition, we only displayed fixation clusters in
                                                                               the gaze-block conditions where gaze duration was greater than
  Figure 5: Examples of symmetric and asymmetric targets.                      270 ms.
  The black shape is targets and the gray close distractors.
                                                                               4.5    Procedure
In addition, the task required target objects to match a particular            Each participant’s session took about one hour to complete. A
color. To ensure color matching was non-trivial, we placed the tar-            researcher first demonstrated to the participant the basics of the
gets and distractors against uniquely generated, non-uniform col-              experimental task by explaining and showing a sample search task.
ored background and had distractors (both close and random) col-               The participant was then asked to perform the experimental image
ored in the same color. For the targets, we used five different basic
                                                                               search tasks. The experimental session was divided into four parts,
colors (blue, green, orange, purple, and red), and three shades of
                                                                               one for each condition (baseline-full image, segmented, gaze-
each color. Figure 6 shows an example of stimuli used with a mid-
                                                                               blocked, and combined segmented and gaze-blocked). Ordering of
shade orange color as the target color.
                                                                               the conditions was counterbalanced using a 4x4 Latin Square.
Figure 6 further illustrates the combination of requirements and
constraints of the study tasks. A legend at the top of the image               In each condition, participants started the task by calibrating the
describes the target shape and the color separately. In all, 24                eye tracker. Next, participants were shown a screen that described
images were generated. Four images included no targets matching                the current experimental condition. Pressing the space bar loaded
the description in the legend. Twenty images, ten with symmetric               the first stimulus image. The image would be displayed for 67.5
targets and ten with asymmetric targets, were generated with a ran-            seconds (Phase 1). Next, the screen was blanked for three seconds.
dom number of targets ranging between 5-20. In addition each                   The image was then displayed according to one of the four condi-
image had between 10-40 target shapes with wrong color, 35-40                  tions (Phase 2). The participant was given another 67.5 seconds to
close distractors with the target correct color, 10-40 close distrac-          inspect the image and find targets. In conditions including segmen-
tors with wrong color. Each stimulus image had 260-300 shapes in               tation of the image, each segment was shown for 7.5 seconds.
total. The difference between the total number of shapes and the               Another three second break was given between successive tasks,
targets plus close distractors were filled with random distractors in          for a total of 24 tasks per participant.
random colors. All shapes had approximately the same size,
around 24 pixels high and wide.                                                A total of six images were presented in each condition. The first
                                                                               image was a training image to familiarize participants with the pre-
4.4    Experimental Setup                                                      sentation method and timing in the second phase. One image con-
                                                                               tained no targets. The no-target image and the training image were
A ViewSonic 18” CRT monitor set at 1024 x 768 pixels resolution
                                                                               not included in the analysis. The order of images after the training
was used to present task stimuli to the participants. A CRT was
chosen over a LCD panel as it showed consistent colors indepen-                image was randomized within each condition. Tasks were bal-
dent of viewing angle. This was important because the task                     anced over conditions.
required participants to carefully judge colors. A Tobii X120 eye
tracker was positioned in front of the monitor. Participants were              Before repeating the search task for the remaining conditions, par-
seated approximately 60 cm from the monitor.                                   ticipants filled out a NASA TLX questionnaire [Hart and Staveland
                                                                               1988] for assessing their subjective workload. After participants
The Tobii X120 was configured to collect gaze data at a rate of 60             had gone through all four conditions, they ranked the rating scales,
Hz. We found that this rate meet the input requirements of our sys-            as specified in the final step in the NASA TLX procedure. Finally,
tem and provided more robust results compared to higher collec-                the participants were asked a few debriefing questions about their
tion rates. The dispersion threshold for the fixation detection algo-          experience of using the four conditions. They were also asked to
rithm was set to 40 pixels. The minimum duration of a fixation was             rank the four conditions according to their preference.



                                                                         286
4.6                                    Data Analysis                                                                                     Condition       TP    TPadj       FN            FNadj
In all analysis, if nothing else is mentioned, we used a 2x2 full fac-                    Overall                                                        86%    88%         14%            11%
torial repeated measurements ANOVA. We expected that partici-
                                                                                          Block                                                          83%    88%         17%            11%
pants’ performance would not differ during the first phase, the free
viewing phase. This assumption was confirmed on all dependent                             No Block                                                       88%    88%         12%            12%
measures reported in the results section. Hence, we focused our                           Segmentation                                                   85%    87%         15%            13%
analysis on the performance during the second phase. We looked at                         Full Image                                                     86%    90%         14%            10%
the True Positive (TP) rate, the False Positive (FP) rate and the
                                                                                          Block + seg                                                    83%    86%         14%            10%
False Negative (FN) rate. The false negative rate requires some
additional elaboration because it consists of items viewed (gazed                         Block + full image                                             83%    90%         14%             7%
at) and not judged as matching, and also of items not viewed at all.                      No Block + seg                                                 87%    87%         11%            11%
Thus we calculated FNV and FNNV rates (where FNV + FNNV =                                 No Block + full image                                          90%    90%          8%             8%
FN). We calculated the overall performance in phase 2 and an
adjusted performance based on visible targets not selected for TP                                             Table 1: TP, TPadj, FN, FNadj rates at end of phase 2.
and FN. The visible targets were the total number of targets present
in the image minus the targets blocked or partially blocked by the                      False positives classifications (FP) were rare in our experiment.
gaze-blocking. The adjusted TP and FN hence reflect the number of                       Across all conditions, only 6% of the total number of selections
targets in the second phase that the participants could see and                         were FPs (SD=7.6). No significant difference was found between
judge in full.                                                                          conditions.

We analyzed patterns of gaze on targets, and found that partici-                        The false negative rate (FN), incorrectly identifying a true target as
pants often fixated multiple times on a target. For this reason we                      false, was 15% (SD=10.9) across all conditions. During the second
analyzed total gaze duration on the target rather than the fixation                     phase, participants reduced FNs by, on average, 59% (SD=21.7).
durations. This method of analysis is also more consistent with                         We did not find any main effects for gaze display and image dis-
how fixation clusters were calculated.                                                  play variables.

5                       Results                                                         For many image inspection tasks, reducing the FN rate is impor-
                                                                                        tant. In this work, we are particularly interested in reducing the
                                                                                        FNNV rate. We compared the number of FNNV outcomes in phase 1
5.1                                    Performance and Error
                                                                                        with the three different outcomes after phase 2: FNNV (still not
In any inspection task, it is important to have as many true posi-                      viewed), FNV (viewed but not selected) and TP (viewed and
tives (TPs), correctly identifying a true target, as possible. In our                   selected). Table 1 shows ratios for these three categories for all
study the participants found on average 86% (SD=10.7) of all tar-                       conditions. As the table illustrates, we again did not find any main
gets after both phases across all conditions. In phase 2, they                          effects, but we found a significant interaction for the FNNV in
improved their performance on average 31% (SD=20.2) across all                          phase 1 transitioning to TPs durring phase 2 (F(1, 7)= 6.321,
conditions. See Table 1 for a summary of TP and TP adjusted for                         p<0.05). In line with the previously found interaction, the combi-
visible targets (TPadj) for all conditions. We did not find any main                    nation of gaze-blocked and no segmentation (FB) gave more TP
effect on TP or TPabjfor the independent variables image display                        from previously not seen targets, an average of 82% compared to
and gaze display, or any effect of gaze blocking or image segmen-                       the overall mean of 67%.
tation. However, when testing the improvement of TP from
phase 1 to phase 2 adjusted for visible targets, we found a border-                     5.2 Revisitations of Targets and Image Search
line significant interaction between gaze blocking and segmenta-                        Strategy
tion (F(1,7)=4.589, p=0.069). As Figure 7 shows, the combination
of gaze-blocking and no segments gave a higher improvement in                           Previous literature has shown that people revisit targets and that
adjusted TP rate (49%) compared to all other combination of con-                        revisitation may change the person’s judgement or classification
ditions.
       100
                                                                                        [Nodine and Kundle 1987]. In our study, the gaze-blocking condi-
                                                                No segmentation
                                                                Segmentation                                                           100
                                                                                                                                                                       No segmentation
                                                                                                                                                                       Segmentation
                                       80

                                                                                                                                       80
                                                                                             Percent change from FN not viewed to TP
      Percentage increase of TP rate




                                       60
                                                                                                                                       60



                                       40
                                                                                                                                       40



                                       20
                                                                                                                                       20




                                       0
                                                                                                                                        0
                                               BLOCK           NO BLOCK
                                                                                                                                                 BLOCK             NO BLOCK
       Figure 7: Average improvement of TP rate adjusted for                              Figure 8: Average transition of FN not viewed in phase 1 to
        visible targets. Error bars indicate ±1 standard error                            TP during phase 2. Error bars indicate ± 1 standard error.




                                                                                  287
Condition             FNV     FNNV         TP                                    1.0                                                         FN
                                                                                                                                                                                                        TP
  Overall                                                                          19%     13%         68%
  Block                                                                            19%     14%         66%                                  0.8

  No Block                                                                         19%     11%         69%
  Segmentation                                                                     26%     15%         59%




                                                                                                                    Survival rate of gaze
                                                                                                                                            0.6
  Full Image                                                                       12%     11%         77%
  Block + seg                                                                      29%     20%         51%
                                                                                                                                            0.4
  Block + full image                                                               10%      8%         82%
  No Block + seg                                                                   24%      9%         67%
  No Block + full image                                                            15%     13%         71%                                  0.2


Table 2: Ratio of FNNV from phase 1 to FNV, FNNV, or TP at end
                          of phase 2.                                                                                                       0.0


                                                                                                                                                  0     500   1000         1500           2000   2500   3000

                                                                                                                                                                     Gaze duration (ms)
                                                        1200
                                                               No segmentation
                                                               Segmentation
                                                                                                                                            Figure 10: Survival rate of gaze duration on TP and FN
                                                                                                                                                           targets during phase 1.
    Average gaze duration (ms) for FNs during phase I




                                                        1000

                                                                                                                   analysis indicates that participants likely used different strategies
                                                        800
                                                                                                                   for searching for targets in different inspection conditions. During
                                                                                                                   the first phase of the gaze-blocking conditions, participants likely
                                                                                                                   searched for easily identifiable targets and used the second phase
                                                        600                                                        to identify targets that required a higher degree of examination to
                                                                                                                   distinguish targets from distractors. For example, one participant
                                                        400
                                                                                                                   indicated in the debriefing interview that he would examine less
                                                                                                                   dense clusters of objects first, and then move on to examine more
                                                                                                                   dense object clusters. Another participant stated he would try to
                                                        200                                                        identify objects “that jumped out at [him]” first.

                                                          0
                                                                                                                   5.3                            Subjective Workload and Preference
                                                                     BLOCK                 NO BLOCK
                                                                                                                   Although the differences between the different conditions for the
Figure 9: Average duration on FN during phase 1. Error bars                                                        task load index (TLX) were small in magnitude, we found a signif-
                indicate ± 1 Standard Error                                                                        icant main effect for the gaze display variable (F(1, 7)=6.905,
                                                                                                                   p<0.05). When gaze-blocking was present, participants experi-
tions make areas previously visited non-viewable if they are well                                                  enced a significantly lower workload (mean=63.6) compared to it
inspected. The goal is to guide examiners to focus on finding new                                                  was not present (mean=67.5). This correspond to 6% reduction in
targets. If people frequently revisit targets and change their mind,                                               task load. We did not find any effect due to segmentation.
this might negatively effect the performance of the gaze-blocked
level. For this reason, we investigated how frequently participants                                                Participants were asked to rank their preferred methodology from
revisited visible targets and how often they changed judgments                                                     1 (most preferred) to 4 (least preferred). Differences among meth-
from FN to TP. We found that 51% (SD=23.1) of all new TP in                                                        odologies were small and no one methodology was overwhelming
phase 2 had been viewed in phase 1. This difference was not sig-                                                   preferred across all participants. The highest overall rated method-
nificant, F(1, 7)=.222, ns. There was no main effects for the two                                                  ology was no segmentation and no gaze-blocking (F: summed
independent variables.                                                                                             score: 17), followed by no segmentation and gaze-blocking (FB:
                                                                                                                   summed score: 18). However, when we correlated the preference
This result shows that adding a second inspection phase is benefi-                                                 scores with the measured workload, we found a significant correla-
cial since people improve performance by re-examining potential                                                    tion (r=0.45, t(30)=2.785, p<0.01). That is, participants tended to
targets. Further, it appears that the gaze-blocking did not signifi-                                               prefer conditions where they reported lowest subjective workload.
cantly affect the number of targets being re-examined. The number
of FNV blocked in phase 2 was small: only three of eight partici-                                                  5.4                            Fixation Patterns on Targets
pants experienced any blocking of FNVs, and of these, two partici-
pants had FNVs blocked in both gaze-blocking conditions, where                                                     For a gaze-blocking to be feasible, gaze durations on targets need
on average 17% of FNV were blocked.                                                                                to be distinguishable. When designing gaze-blocking, we relied on
                                                                                                                   observations that fixations on TPs were longer than those on FNs
When examining the gaze duration in phase 1 of viewed FN targets                                                   [Nodine and Kundle 1987] in the domain of radiology. We exam-
which the participants changed to TP after re-examination, we saw                                                  ined gaze durations on targets, and found that in our domain too
a difference in gaze duration between the two levels in the gaze                                                   TPs were gazed at longer than FNs. The average duration of FNs
display variable (Figure 9). On average the gaze duration in the                                                   were 661ms (SD=628.5), and for TPs 1386 ms (SD=1032.5). This
gaze-blocking conditions was 443 ms (SD=251.2), while in the                                                       difference was significant (F(1,7)=16.533, p<0.01). This result
non-blocking conditions it was 746 ms (SD=493.7). This differ-                                                     shows that the duration of gaze on targets differed depending the
ence was very close to significant (F(1, 7)=5.574, p=.0503). This                                                  participant’s interpretation of the target.



                                                                                                             288
As illustrated in Figure 10, there is a definite pattern in the data           Our results showed that slightly less than half of the performance
which indicates TP, on average, are dwelled on longer by examin-               gain when previous gaze was blocked in phase 2 came from previ-
ers compared to FN. For instance, at 550 ms, the participants had              ously non-viewed objects. This is a promising result, in particular
stopped looking at 50% of FNs, but only stopped looking at 32% of              for situations where image analysis is done under time pressure.
the TPs. The 50% gaze duration survival rate for TPs was at 1032               When time is unlimited, more targets may be viewed during the
ms. On the other hand, at gaze duration of 1049 ms, 70% of the FN              first phase. An interesting research issue to look at in the future is
were shorter. Although both TPs and FNs has a large spread, these              how people can optimize their performance using a two phase
numbers reaffirm that it should be possible to use gaze duration as            inspection method and what trade-offs the image analyst faces
an implicit relevance feedback mechanism to support image analy-               when previous gaze is blocked during a second phase.
sis.
                                                                               Revisitation of previously looked at targets is another important
We also found a significant difference in fixation duration between            factor for the performance increase during phase 2. Although gaze
phase 1 and phase 2 (F(1, 7)=71.597, p<0.001). During phase 2,                 blocking masked previously well-inspected areas, we could not
participants’ fixation duration was on average 813 ms (SD=749)                 find a significant impact of gaze-blocking during the second phase
and in phase 2 the duration was on average 1235 ms (SD=1037).                  in terms of revisitations. In our study, a low percentage of the pre-
                                                                               viously viewed and un-selected targets were blocked out, and in
There was no difference between the conditions. This result is fur-
                                                                               particular, in gaze-blocking and full-image conditions, participants
ther evidence that the participants shifted strategy between the two
                                                                               made up for that performance loss by finding previously-unseen
phases.
                                                                               targets. This is a promising result, since one of our goals was to
                                                                               lower the number FNs during the second phase.
6    Discussion
                                                                               Another interesting finding in our study is that participants
Intuitively, masking off where you have looked during an image                 adjusted their behavior during the gaze-blocking condition. During
search task is both a bad idea and a compelling idea. When viewed              the first phase, they adopted the strategy of scanning the image for
parts of the image are blocked out, the searcher loses context and it          the easiest to find targets. In the next phase, they would focus on
is not possible to correct previous decisions. On the other hand, in           targets that were harder to find and harder to interpret. It is possi-
a complex image it can be hard to identify all occurrences, and                ble that this strategy contributed to lowered workload in the gaze-
                                                                               blocking condition. This result is intriguing since it suggests more
even to remember what you looked at already. In addition, when
                                                                               research on how to design image search techniques that can pro-
the task involves not just perceptual processes but higher level
                                                                               mote a more efficient use of time and mental resources.
cognitive processes, fatigue sets in quickly due to the demanding
nature of the task, resulting decreased performance. We described
                                                                               Gaze patterns collected during study confirm that it is possible to
four different inspection methods using combinations of gaze-                  give indications to image searchers of which parts of an image
blocking and image segmentation techniques. In the study, we took              they have not inspected in enough detail. As also observed in other
an extreme approach to understand limitations of the methods. The              studies, we found that on average, participants spent significantly
performance results from our study show two things: masking an                 longer looking at targets they ended up selecting (TPs) than those
image based on where searchers looked previously did not nega-                 they discarded as distractors (FNs). Because of the large individual
tively impact performance, and masking gazed-at areas has poten-               difference in gaze duration on TP targets and FN targets, thresholds
tial.                                                                          for gaze-blocking algorithms may need to be set dynamically
                                                                               based on each user’s behavior and preference, or by using machine
The results that gaze blocking reduced participants’ workload is an            learning approaches. In systems using manual pointing to for
important finding. Image search can be a demanding task and if the             marking selections, users’ selections can also be used to get gaze
workload can be reduced, performance can be improved. We                       durations for TPs. In this study, we only looked at the sum of all
believe it is important to find methods supporting image analysts              fixations on the target. Other gaze characteristics, such as number
that also reduce their workload.                                               of fixations on target, duration between fixations, etc. can poten-
                                                                               tially be informative. The image search task may also influence
In terms of performance, not surprisingly the gaze-blocking condi-             what method the system uses to determine which parts of the
tion did on average somewhat worse in true positive rate, in partic-           image would be helpful for the user to look at in more detail.
ular true positive rate not adjusted for visible targets. This differ-
ence was not significant, and hence could just as well as have been            Our study does not describe a realistic system; it merely tests the
caused by other factors than those manipulated. However, when                  limits and the feasibility of using eye tracking to support image
adjusted for targets blocked by the gaze mask, the gaze-block in               search by masking viewed areas. A realistic system would need
combination with whole image showed a larger increase in the TP                ways to make correction, which in the study was only possible on
rate and a larger reduction in FN during phase 2 than the other con-           non-masked areas. A realistic system would also need to allow
ditions. This result indicates that allowing participants to focus on          image analysts to control the pace, segmentation, and masking of
areas not previously viewed helped them find and select more of                viewed areas. However, results from this study are encouraging,
the remaining targets compared to the other three inspection meth-             indicating that eye tracking can provide useful support during
ods. This result is promising for future work.                                 image search.

Segmenting the image did not seem to affect the performance,                   7    Conclusion
expect in combination with gaze block. The SB condition had the
overall lowest performance. Interestingly, when the transition rate            In this paper, we proposed a two-phased approach to image search
from FNNV to TP were highest in gaze block on full image (FB), it              and investigated the performance, workload, and user preference
was also lowest in SB. It appears that restricing the viewable area            trade-offs of four different second phase image re-presentation
too much has negative affects on performance. Future research is               methodologies. Results from a controlled laboratory study showed
needed to explain this phenonemon.                                             that in a two-phase approach the gaze-blocking treatment, which



                                                                         289
masks well examined regions of the image in the second phase of            gage screening due to training. Ergonomics, 52(6), 644–656.
inspection, significantly reduced overall subjective workload.
Gaze block on a full image increased the transition of previously          MANNING, D.J., ETHELL, S.C. AND DONOVAN, T. 2004.
unseen false negatives to true positives, and was nearly as pre-           Detection or decision errors? Missed lung cancer from posterante-
ferred as free-form search. Further, our results also suggest that         rior chest radiograph. British Journal of Radiology, 77, 231-235.
segmentation and gaze-blocking likely encourage searchers to               MCCARLEY, J.S, KRAMER, A.F., WICKENS, C.D., VIDONI,
adopt an more structured search strategy in both phases of exami-          E.D. AND BOOT, W. R. 2004. Visual Skills in Airport-Security
nation. Results from our study can have broad impact on the                Screening. Psychological Science, 15(5), 302-306.
design of future techniques and systems that aid in visual search.
                                                                           NAVALPAKKAM, V., KOCH, C., AND PERONA, P. 2009.
                                                                           Homo economicus in visual search. Journal of Vision, 9(1):31, 1-
8   Acknowledgments                                                        16.
                                                                           NODINE, C. F. AND KUNDEL, H. L. 1987. Using eye move-
We thank all our participants for their time and effort, and Lynn          ments to study visual search and to improve tumor detection.
Wilcox and Larry Rowe for supporting this research.                        RadioGraphics 7(6) 1241-1250.
                                                                           NORTON, D., AND STARK, L. 1971. Eye movements and visual
9   References
                                                                           perception. Scientific American, 224, 34-43.

BUTLER, V. AND POOLE, R.W. 2002. Rethinking Checked-Bag-                   MARKUS, J.B., SOMERS, S., O’MALLEY, B.P., AND STE-
gage Screening. Reason Public Policy Institute; July. 25p. Policy          VENSON, G.W. 1990. Double-contrast barium enema studies:
Study No. 297.                                                             effect of multiple reading on perception error. Radiology 175, 155-
                                                                           156.
FENGA, Q., HANDE, S., AND KAPURB, K.C. 2008. Designing
                                                                           POMPLUN, M., REINGOLD, E.M. AND SHEN, J. 2001. Investi-
Airport Checked-Baggage-Screening Strategies Considering Sys-
                                                                           gating the visual span in comparative search: the effect of task dif-
tem Capability and Reliability. Reliability Engineering & System
                                                                           ficulty and divided attention. Cognition, 8, B57-B67.
Safety. 94(2) 618-627.
                                                                           RAYNER, K. 1998. Eye movements in reading and information
FLECK., M.S. AND MITROFF, S. 2007. Rare targets are rarely
                                                                           processing: 20 years of research. Psychological Bulletin, 124(3),
missed in correctable search. Psychological Science, 18(11), 943-
                                                                           372-422.
947.
                                                                           SADASIVAN, S., GREENSTEIN, J. S., GRAMOPADHYE, A.K.,
FORLINES, C., BALAKRISHNAN, R. 2009. Improving visual                      DUCHOWSKI. 2005. Use of eye movement as feedforward train-
search with image segmentation. In Proceedings of CHI’2009, Pp.            ing for a synthetic aircraft inspection task. In Proceedings of
1093-1102.                                                                 CHI’2005, 141-149.
GILCHRIST, I. D. AND HARVEY, M. 2006. Evidence for a sys-                  SANTELLA, A. AND DECARLO, D. 2004. Robust clustering of
tematic component within scan paths in visual search. Visual Cog-          eye movement recordings for quantification of visual interest. In
nition, 14, 704-71                                                         Proceedings of ETRA’04. Pp27-34
GODDARD, P., LESLIE, A., JONES, A., WAKELEY, C., AND                       SCINTO, L. F., PILLALAMARRI, R., KARSH, R. 1986. Cogni-
KABALA, J. 2001. Error in radiology. British Journal Radiology             tive strategies for visual search. Acta Psychologica. 62(3), 263-
74, 949-951.                                                               292.
HAIMSON, C, BOTHELL, D., DOUGLASS, S.A., ANDERSON,                         SHAW, N., HENDRY, M., AND EDEN, O. 1990. Inter-observer
J.R. 2004. Partitioning Visual Displays Aids Task-Directed Visual          variation in interpretation of chest X-rays. Scott Med J 35, 140-
Search. Human Factors, 46(3), 551-566                                      141.
HART S.G., STAVELAND L.E. 1988. Development of a NASA-                     VOCK, P. 1987. Frequent errors in the interpretation of chest x-ray
TLX (Task Load Index): Results of Empirical and Theoretical                films. Schweiz Med Wochenschr 117, 928-932.
Research.” In Hancock PS, Meshkati N, (eds). Human Mental
Workload. Amsterdam: North-Holland: pp. 139–83.                            WOLFE, J.M., HOROWITZ, T.S., VAN WERT, M.J., KENNER,
                                                                           N.M., PLACE, S.S., KIBBI, N. 2007. Low target prevalence is a
KOLLERA, S.M., DRURY, C.G., AND SCHWANINGER, A.                            stubborn source of errors in visual search tasks. Journal of Experi-
2009. Change of search time and non-search time in X-ray bag-              mental Pshychology: General. 136(6), 623-638.




                                                                     290

More Related Content

What's hot

Statistics: Final test study guide (word 2003)
Statistics: Final test study guide (word 2003)Statistics: Final test study guide (word 2003)
Statistics: Final test study guide (word 2003)Mark default
 
Image restoration model with wavelet based fusion
Image restoration model with wavelet based fusionImage restoration model with wavelet based fusion
Image restoration model with wavelet based fusionAlexander Decker
 
FACE VERIFICATION ACROSS AGE PROGRESSION USING ENHANCED CONVOLUTION NEURAL NE...
FACE VERIFICATION ACROSS AGE PROGRESSION USING ENHANCED CONVOLUTION NEURAL NE...FACE VERIFICATION ACROSS AGE PROGRESSION USING ENHANCED CONVOLUTION NEURAL NE...
FACE VERIFICATION ACROSS AGE PROGRESSION USING ENHANCED CONVOLUTION NEURAL NE...sipij
 
Automated Diagnosis of Glaucoma using Haralick Texture Features
Automated Diagnosis of Glaucoma using Haralick Texture FeaturesAutomated Diagnosis of Glaucoma using Haralick Texture Features
Automated Diagnosis of Glaucoma using Haralick Texture FeaturesIOSR Journals
 
MultiModal Identification System in Monozygotic Twins
MultiModal Identification System in Monozygotic TwinsMultiModal Identification System in Monozygotic Twins
MultiModal Identification System in Monozygotic TwinsCSCJournals
 
(2006) Impact of Image Quality on Performance: Comparison of Young and Elderl...
(2006) Impact of Image Quality on Performance: Comparison of Young and Elderl...(2006) Impact of Image Quality on Performance: Comparison of Young and Elderl...
(2006) Impact of Image Quality on Performance: Comparison of Young and Elderl...International Center for Biometric Research
 
ANALYSIS OF WATERMARKING TECHNIQUES FOR MEDICAL IMAGES PRESERVING ROI
ANALYSIS OF WATERMARKING TECHNIQUES FOR MEDICAL IMAGES PRESERVING ROI ANALYSIS OF WATERMARKING TECHNIQUES FOR MEDICAL IMAGES PRESERVING ROI
ANALYSIS OF WATERMARKING TECHNIQUES FOR MEDICAL IMAGES PRESERVING ROI cscpconf
 
Comparison of Machine Learning Techniques for Identification of Disease
Comparison of Machine Learning Techniques for Identification of DiseaseComparison of Machine Learning Techniques for Identification of Disease
Comparison of Machine Learning Techniques for Identification of DiseaseIJCSIS Research Publications
 
Klingner Fixation Aligned Pupillary Response Averaging
Klingner Fixation Aligned Pupillary Response AveragingKlingner Fixation Aligned Pupillary Response Averaging
Klingner Fixation Aligned Pupillary Response AveragingKalle
 
MEDICAL IMAGE TEXTURE SEGMENTATION USINGRANGE FILTER
MEDICAL IMAGE TEXTURE SEGMENTATION USINGRANGE FILTERMEDICAL IMAGE TEXTURE SEGMENTATION USINGRANGE FILTER
MEDICAL IMAGE TEXTURE SEGMENTATION USINGRANGE FILTERcscpconf
 

What's hot (11)

Statistics: Final test study guide (word 2003)
Statistics: Final test study guide (word 2003)Statistics: Final test study guide (word 2003)
Statistics: Final test study guide (word 2003)
 
Image restoration model with wavelet based fusion
Image restoration model with wavelet based fusionImage restoration model with wavelet based fusion
Image restoration model with wavelet based fusion
 
FACE VERIFICATION ACROSS AGE PROGRESSION USING ENHANCED CONVOLUTION NEURAL NE...
FACE VERIFICATION ACROSS AGE PROGRESSION USING ENHANCED CONVOLUTION NEURAL NE...FACE VERIFICATION ACROSS AGE PROGRESSION USING ENHANCED CONVOLUTION NEURAL NE...
FACE VERIFICATION ACROSS AGE PROGRESSION USING ENHANCED CONVOLUTION NEURAL NE...
 
Automated Diagnosis of Glaucoma using Haralick Texture Features
Automated Diagnosis of Glaucoma using Haralick Texture FeaturesAutomated Diagnosis of Glaucoma using Haralick Texture Features
Automated Diagnosis of Glaucoma using Haralick Texture Features
 
MultiModal Identification System in Monozygotic Twins
MultiModal Identification System in Monozygotic TwinsMultiModal Identification System in Monozygotic Twins
MultiModal Identification System in Monozygotic Twins
 
(2006) Impact of Image Quality on Performance: Comparison of Young and Elderl...
(2006) Impact of Image Quality on Performance: Comparison of Young and Elderl...(2006) Impact of Image Quality on Performance: Comparison of Young and Elderl...
(2006) Impact of Image Quality on Performance: Comparison of Young and Elderl...
 
ANALYSIS OF WATERMARKING TECHNIQUES FOR MEDICAL IMAGES PRESERVING ROI
ANALYSIS OF WATERMARKING TECHNIQUES FOR MEDICAL IMAGES PRESERVING ROI ANALYSIS OF WATERMARKING TECHNIQUES FOR MEDICAL IMAGES PRESERVING ROI
ANALYSIS OF WATERMARKING TECHNIQUES FOR MEDICAL IMAGES PRESERVING ROI
 
Comparison of Machine Learning Techniques for Identification of Disease
Comparison of Machine Learning Techniques for Identification of DiseaseComparison of Machine Learning Techniques for Identification of Disease
Comparison of Machine Learning Techniques for Identification of Disease
 
01
0101
01
 
Klingner Fixation Aligned Pupillary Response Averaging
Klingner Fixation Aligned Pupillary Response AveragingKlingner Fixation Aligned Pupillary Response Averaging
Klingner Fixation Aligned Pupillary Response Averaging
 
MEDICAL IMAGE TEXTURE SEGMENTATION USINGRANGE FILTER
MEDICAL IMAGE TEXTURE SEGMENTATION USINGRANGE FILTERMEDICAL IMAGE TEXTURE SEGMENTATION USINGRANGE FILTER
MEDICAL IMAGE TEXTURE SEGMENTATION USINGRANGE FILTER
 

Viewers also liked

Tien Measuring Situation Awareness Of Surgeons In Laparoscopic Training
Tien Measuring Situation Awareness Of Surgeons In Laparoscopic TrainingTien Measuring Situation Awareness Of Surgeons In Laparoscopic Training
Tien Measuring Situation Awareness Of Surgeons In Laparoscopic TrainingKalle
 
nextNY Online Marketing School - SEM Presentation
nextNY Online Marketing School - SEM PresentationnextNY Online Marketing School - SEM Presentation
nextNY Online Marketing School - SEM PresentationnextNY
 
Lone Star College Tomball Community Library
Lone Star College Tomball Community LibraryLone Star College Tomball Community Library
Lone Star College Tomball Community Libraryguest54f40230
 
Jarodzka A Vector Based, Multidimensional Scanpath Similarity Measure
Jarodzka A Vector Based, Multidimensional Scanpath Similarity MeasureJarodzka A Vector Based, Multidimensional Scanpath Similarity Measure
Jarodzka A Vector Based, Multidimensional Scanpath Similarity MeasureKalle
 
Prats Interpretation Of Geometric Shapes An Eye Movement Study
Prats Interpretation Of Geometric Shapes An Eye Movement StudyPrats Interpretation Of Geometric Shapes An Eye Movement Study
Prats Interpretation Of Geometric Shapes An Eye Movement StudyKalle
 
Ebtee Presentation
Ebtee PresentationEbtee Presentation
Ebtee Presentationsdemey
 
Mollenbach Single Gaze Gestures
Mollenbach Single Gaze GesturesMollenbach Single Gaze Gestures
Mollenbach Single Gaze GesturesKalle
 
Intercentros carrefour
Intercentros carrefourIntercentros carrefour
Intercentros carrefouroscargaliza
 
ZFConf 2010: Zend Framework & MVC, Model Implementation (Part 2, Dependency I...
ZFConf 2010: Zend Framework & MVC, Model Implementation (Part 2, Dependency I...ZFConf 2010: Zend Framework & MVC, Model Implementation (Part 2, Dependency I...
ZFConf 2010: Zend Framework & MVC, Model Implementation (Part 2, Dependency I...ZFConf Conference
 
Uunified Communications For General Business
Uunified Communications For General BusinessUunified Communications For General Business
Uunified Communications For General BusinessLarryDougan
 
Moshnyaga The Use Of Eye Tracking For Pc Energy Management
Moshnyaga The Use Of Eye Tracking For Pc Energy ManagementMoshnyaga The Use Of Eye Tracking For Pc Energy Management
Moshnyaga The Use Of Eye Tracking For Pc Energy ManagementKalle
 
Manifesto ao presidente da xunta de galicia maio 2012
Manifesto ao presidente da xunta de galicia maio 2012Manifesto ao presidente da xunta de galicia maio 2012
Manifesto ao presidente da xunta de galicia maio 2012oscargaliza
 
Skmbt 50112081315521 escrito conselleria
Skmbt 50112081315521 escrito conselleriaSkmbt 50112081315521 escrito conselleria
Skmbt 50112081315521 escrito conselleriaoscargaliza
 

Viewers also liked (20)

Tien Measuring Situation Awareness Of Surgeons In Laparoscopic Training
Tien Measuring Situation Awareness Of Surgeons In Laparoscopic TrainingTien Measuring Situation Awareness Of Surgeons In Laparoscopic Training
Tien Measuring Situation Awareness Of Surgeons In Laparoscopic Training
 
nextNY Online Marketing School - SEM Presentation
nextNY Online Marketing School - SEM PresentationnextNY Online Marketing School - SEM Presentation
nextNY Online Marketing School - SEM Presentation
 
Lone Star College Tomball Community Library
Lone Star College Tomball Community LibraryLone Star College Tomball Community Library
Lone Star College Tomball Community Library
 
Greeting1
Greeting1Greeting1
Greeting1
 
Jarodzka A Vector Based, Multidimensional Scanpath Similarity Measure
Jarodzka A Vector Based, Multidimensional Scanpath Similarity MeasureJarodzka A Vector Based, Multidimensional Scanpath Similarity Measure
Jarodzka A Vector Based, Multidimensional Scanpath Similarity Measure
 
Prats Interpretation Of Geometric Shapes An Eye Movement Study
Prats Interpretation Of Geometric Shapes An Eye Movement StudyPrats Interpretation Of Geometric Shapes An Eye Movement Study
Prats Interpretation Of Geometric Shapes An Eye Movement Study
 
Ebtee Presentation
Ebtee PresentationEbtee Presentation
Ebtee Presentation
 
Test
TestTest
Test
 
Sport Equipment Ukr
Sport Equipment UkrSport Equipment Ukr
Sport Equipment Ukr
 
Mollenbach Single Gaze Gestures
Mollenbach Single Gaze GesturesMollenbach Single Gaze Gestures
Mollenbach Single Gaze Gestures
 
Intercentros carrefour
Intercentros carrefourIntercentros carrefour
Intercentros carrefour
 
TEMA 1A GRAMMAR INFINITIVES
TEMA 1A GRAMMAR INFINITIVESTEMA 1A GRAMMAR INFINITIVES
TEMA 1A GRAMMAR INFINITIVES
 
ZFConf 2010: Zend Framework & MVC, Model Implementation (Part 2, Dependency I...
ZFConf 2010: Zend Framework & MVC, Model Implementation (Part 2, Dependency I...ZFConf 2010: Zend Framework & MVC, Model Implementation (Part 2, Dependency I...
ZFConf 2010: Zend Framework & MVC, Model Implementation (Part 2, Dependency I...
 
Uunified Communications For General Business
Uunified Communications For General BusinessUunified Communications For General Business
Uunified Communications For General Business
 
Full MSSQL Injection PWNage
Full MSSQL Injection PWNageFull MSSQL Injection PWNage
Full MSSQL Injection PWNage
 
Moshnyaga The Use Of Eye Tracking For Pc Energy Management
Moshnyaga The Use Of Eye Tracking For Pc Energy ManagementMoshnyaga The Use Of Eye Tracking For Pc Energy Management
Moshnyaga The Use Of Eye Tracking For Pc Energy Management
 
Manifesto ao presidente da xunta de galicia maio 2012
Manifesto ao presidente da xunta de galicia maio 2012Manifesto ao presidente da xunta de galicia maio 2012
Manifesto ao presidente da xunta de galicia maio 2012
 
Skmbt 50112081315521 escrito conselleria
Skmbt 50112081315521 escrito conselleriaSkmbt 50112081315521 escrito conselleria
Skmbt 50112081315521 escrito conselleria
 
Stagnation
StagnationStagnation
Stagnation
 
jvum2013a niftycloud
jvum2013a niftycloudjvum2013a niftycloud
jvum2013a niftycloud
 

Similar to Understanding the Benefits of Gaze Enhanced Visual Search

A Hybrid Approach to Face Detection And Feature Extraction
A Hybrid Approach to Face Detection And Feature ExtractionA Hybrid Approach to Face Detection And Feature Extraction
A Hybrid Approach to Face Detection And Feature Extractioniosrjce
 
A Survey on Person Detection for Social Distancing and Safety Violation Alert...
A Survey on Person Detection for Social Distancing and Safety Violation Alert...A Survey on Person Detection for Social Distancing and Safety Violation Alert...
A Survey on Person Detection for Social Distancing and Safety Violation Alert...IRJET Journal
 
DEEP FACIAL DIAGNOSIS: DEEP TRANSFER LEARNING FROM FACE RECOGNITION TO FACIAL...
DEEP FACIAL DIAGNOSIS: DEEP TRANSFER LEARNING FROM FACE RECOGNITION TO FACIAL...DEEP FACIAL DIAGNOSIS: DEEP TRANSFER LEARNING FROM FACE RECOGNITION TO FACIAL...
DEEP FACIAL DIAGNOSIS: DEEP TRANSFER LEARNING FROM FACE RECOGNITION TO FACIAL...IRJET Journal
 
APPLICATION OF CNN MODEL ON MEDICAL IMAGE
APPLICATION OF CNN MODEL ON MEDICAL IMAGEAPPLICATION OF CNN MODEL ON MEDICAL IMAGE
APPLICATION OF CNN MODEL ON MEDICAL IMAGEIRJET Journal
 
Application of generative adversarial networks (GAN) for ophthalmology image ...
Application of generative adversarial networks (GAN) for ophthalmology image ...Application of generative adversarial networks (GAN) for ophthalmology image ...
Application of generative adversarial networks (GAN) for ophthalmology image ...Daniel983829
 
Pneumonia Detection Using Deep Learning and Transfer Learning
Pneumonia Detection Using Deep Learning and Transfer LearningPneumonia Detection Using Deep Learning and Transfer Learning
Pneumonia Detection Using Deep Learning and Transfer LearningIRJET Journal
 
LEARNING BASES OF ACTICITY
LEARNING BASES OF ACTICITYLEARNING BASES OF ACTICITY
LEARNING BASES OF ACTICITYPadma Kannan
 
Image Forgery Detection Using Deep Neural Network
Image Forgery Detection Using Deep Neural NetworkImage Forgery Detection Using Deep Neural Network
Image Forgery Detection Using Deep Neural NetworkIRJET Journal
 
harsh final ppt (2).pptx
harsh final ppt (2).pptxharsh final ppt (2).pptx
harsh final ppt (2).pptxAkbarali206563
 
Face Verification Across Age Progression using Enhanced Convolution Neural Ne...
Face Verification Across Age Progression using Enhanced Convolution Neural Ne...Face Verification Across Age Progression using Enhanced Convolution Neural Ne...
Face Verification Across Age Progression using Enhanced Convolution Neural Ne...sipij
 
Ultrasound image segmentation through deep learning based improvised U-Net
Ultrasound image segmentation through deep learning based  improvised U-NetUltrasound image segmentation through deep learning based  improvised U-Net
Ultrasound image segmentation through deep learning based improvised U-Netnooriasukmaningtyas
 
Care expert assistant for Medicare system using Machine learning
Care expert assistant for Medicare system using Machine learningCare expert assistant for Medicare system using Machine learning
Care expert assistant for Medicare system using Machine learningIRJET Journal
 
Top Cited Articles in Signal & Image Processing 2021-2022
Top Cited Articles in Signal & Image Processing 2021-2022Top Cited Articles in Signal & Image Processing 2021-2022
Top Cited Articles in Signal & Image Processing 2021-2022sipij
 
Image annotation - Segmentation & Annotation
Image annotation - Segmentation & AnnotationImage annotation - Segmentation & Annotation
Image annotation - Segmentation & AnnotationTaposh Roy
 

Similar to Understanding the Benefits of Gaze Enhanced Visual Search (20)

K017247882
K017247882K017247882
K017247882
 
A Hybrid Approach to Face Detection And Feature Extraction
A Hybrid Approach to Face Detection And Feature ExtractionA Hybrid Approach to Face Detection And Feature Extraction
A Hybrid Approach to Face Detection And Feature Extraction
 
A Survey on Person Detection for Social Distancing and Safety Violation Alert...
A Survey on Person Detection for Social Distancing and Safety Violation Alert...A Survey on Person Detection for Social Distancing and Safety Violation Alert...
A Survey on Person Detection for Social Distancing and Safety Violation Alert...
 
DEEP FACIAL DIAGNOSIS: DEEP TRANSFER LEARNING FROM FACE RECOGNITION TO FACIAL...
DEEP FACIAL DIAGNOSIS: DEEP TRANSFER LEARNING FROM FACE RECOGNITION TO FACIAL...DEEP FACIAL DIAGNOSIS: DEEP TRANSFER LEARNING FROM FACE RECOGNITION TO FACIAL...
DEEP FACIAL DIAGNOSIS: DEEP TRANSFER LEARNING FROM FACE RECOGNITION TO FACIAL...
 
APPLICATION OF CNN MODEL ON MEDICAL IMAGE
APPLICATION OF CNN MODEL ON MEDICAL IMAGEAPPLICATION OF CNN MODEL ON MEDICAL IMAGE
APPLICATION OF CNN MODEL ON MEDICAL IMAGE
 
Application of generative adversarial networks (GAN) for ophthalmology image ...
Application of generative adversarial networks (GAN) for ophthalmology image ...Application of generative adversarial networks (GAN) for ophthalmology image ...
Application of generative adversarial networks (GAN) for ophthalmology image ...
 
Pneumonia Detection Using Deep Learning and Transfer Learning
Pneumonia Detection Using Deep Learning and Transfer LearningPneumonia Detection Using Deep Learning and Transfer Learning
Pneumonia Detection Using Deep Learning and Transfer Learning
 
Af044212215
Af044212215Af044212215
Af044212215
 
LEARNING BASES OF ACTICITY
LEARNING BASES OF ACTICITYLEARNING BASES OF ACTICITY
LEARNING BASES OF ACTICITY
 
P045058186
P045058186P045058186
P045058186
 
Image Forgery Detection Using Deep Neural Network
Image Forgery Detection Using Deep Neural NetworkImage Forgery Detection Using Deep Neural Network
Image Forgery Detection Using Deep Neural Network
 
harsh final ppt (2).pptx
harsh final ppt (2).pptxharsh final ppt (2).pptx
harsh final ppt (2).pptx
 
Face Verification Across Age Progression using Enhanced Convolution Neural Ne...
Face Verification Across Age Progression using Enhanced Convolution Neural Ne...Face Verification Across Age Progression using Enhanced Convolution Neural Ne...
Face Verification Across Age Progression using Enhanced Convolution Neural Ne...
 
(2011) Comparison of Face Image Quality Metrics
(2011) Comparison of Face Image Quality Metrics(2011) Comparison of Face Image Quality Metrics
(2011) Comparison of Face Image Quality Metrics
 
Ultrasound image segmentation through deep learning based improvised U-Net
Ultrasound image segmentation through deep learning based  improvised U-NetUltrasound image segmentation through deep learning based  improvised U-Net
Ultrasound image segmentation through deep learning based improvised U-Net
 
Care expert assistant for Medicare system using Machine learning
Care expert assistant for Medicare system using Machine learningCare expert assistant for Medicare system using Machine learning
Care expert assistant for Medicare system using Machine learning
 
Top Cited Articles in Signal & Image Processing 2021-2022
Top Cited Articles in Signal & Image Processing 2021-2022Top Cited Articles in Signal & Image Processing 2021-2022
Top Cited Articles in Signal & Image Processing 2021-2022
 
TBerger_FinalReport
TBerger_FinalReportTBerger_FinalReport
TBerger_FinalReport
 
Image annotation - Segmentation & Annotation
Image annotation - Segmentation & AnnotationImage annotation - Segmentation & Annotation
Image annotation - Segmentation & Annotation
 
S4502115119
S4502115119S4502115119
S4502115119
 

More from Kalle

Blignaut Visual Span And Other Parameters For The Generation Of Heatmaps
Blignaut Visual Span And Other Parameters For The Generation Of HeatmapsBlignaut Visual Span And Other Parameters For The Generation Of Heatmaps
Blignaut Visual Span And Other Parameters For The Generation Of HeatmapsKalle
 
Zhang Eye Movement As An Interaction Mechanism For Relevance Feedback In A Co...
Zhang Eye Movement As An Interaction Mechanism For Relevance Feedback In A Co...Zhang Eye Movement As An Interaction Mechanism For Relevance Feedback In A Co...
Zhang Eye Movement As An Interaction Mechanism For Relevance Feedback In A Co...Kalle
 
Yamamoto Development Of Eye Tracking Pen Display Based On Stereo Bright Pupil...
Yamamoto Development Of Eye Tracking Pen Display Based On Stereo Bright Pupil...Yamamoto Development Of Eye Tracking Pen Display Based On Stereo Bright Pupil...
Yamamoto Development Of Eye Tracking Pen Display Based On Stereo Bright Pupil...Kalle
 
Wastlund What You See Is Where You Go Testing A Gaze Driven Power Wheelchair ...
Wastlund What You See Is Where You Go Testing A Gaze Driven Power Wheelchair ...Wastlund What You See Is Where You Go Testing A Gaze Driven Power Wheelchair ...
Wastlund What You See Is Where You Go Testing A Gaze Driven Power Wheelchair ...Kalle
 
Vinnikov Contingency Evaluation Of Gaze Contingent Displays For Real Time Vis...
Vinnikov Contingency Evaluation Of Gaze Contingent Displays For Real Time Vis...Vinnikov Contingency Evaluation Of Gaze Contingent Displays For Real Time Vis...
Vinnikov Contingency Evaluation Of Gaze Contingent Displays For Real Time Vis...Kalle
 
Urbina Pies With Ey Es The Limits Of Hierarchical Pie Menus In Gaze Control
Urbina Pies With Ey Es The Limits Of Hierarchical Pie Menus In Gaze ControlUrbina Pies With Ey Es The Limits Of Hierarchical Pie Menus In Gaze Control
Urbina Pies With Ey Es The Limits Of Hierarchical Pie Menus In Gaze ControlKalle
 
Urbina Alternatives To Single Character Entry And Dwell Time Selection On Eye...
Urbina Alternatives To Single Character Entry And Dwell Time Selection On Eye...Urbina Alternatives To Single Character Entry And Dwell Time Selection On Eye...
Urbina Alternatives To Single Character Entry And Dwell Time Selection On Eye...Kalle
 
Takemura Estimating 3 D Point Of Regard And Visualizing Gaze Trajectories Und...
Takemura Estimating 3 D Point Of Regard And Visualizing Gaze Trajectories Und...Takemura Estimating 3 D Point Of Regard And Visualizing Gaze Trajectories Und...
Takemura Estimating 3 D Point Of Regard And Visualizing Gaze Trajectories Und...Kalle
 
Stevenson Eye Tracking With The Adaptive Optics Scanning Laser Ophthalmoscope
Stevenson Eye Tracking With The Adaptive Optics Scanning Laser OphthalmoscopeStevenson Eye Tracking With The Adaptive Optics Scanning Laser Ophthalmoscope
Stevenson Eye Tracking With The Adaptive Optics Scanning Laser OphthalmoscopeKalle
 
Stellmach Advanced Gaze Visualizations For Three Dimensional Virtual Environm...
Stellmach Advanced Gaze Visualizations For Three Dimensional Virtual Environm...Stellmach Advanced Gaze Visualizations For Three Dimensional Virtual Environm...
Stellmach Advanced Gaze Visualizations For Three Dimensional Virtual Environm...Kalle
 
Skovsgaard Small Target Selection With Gaze Alone
Skovsgaard Small Target Selection With Gaze AloneSkovsgaard Small Target Selection With Gaze Alone
Skovsgaard Small Target Selection With Gaze AloneKalle
 
San Agustin Evaluation Of A Low Cost Open Source Gaze Tracker
San Agustin Evaluation Of A Low Cost Open Source Gaze TrackerSan Agustin Evaluation Of A Low Cost Open Source Gaze Tracker
San Agustin Evaluation Of A Low Cost Open Source Gaze TrackerKalle
 
Ryan Match Moving For Area Based Analysis Of Eye Movements In Natural Tasks
Ryan Match Moving For Area Based Analysis Of Eye Movements In Natural TasksRyan Match Moving For Area Based Analysis Of Eye Movements In Natural Tasks
Ryan Match Moving For Area Based Analysis Of Eye Movements In Natural TasksKalle
 
Rosengrant Gaze Scribing In Physics Problem Solving
Rosengrant Gaze Scribing In Physics Problem SolvingRosengrant Gaze Scribing In Physics Problem Solving
Rosengrant Gaze Scribing In Physics Problem SolvingKalle
 
Porta Ce Cursor A Contextual Eye Cursor For General Pointing In Windows Envir...
Porta Ce Cursor A Contextual Eye Cursor For General Pointing In Windows Envir...Porta Ce Cursor A Contextual Eye Cursor For General Pointing In Windows Envir...
Porta Ce Cursor A Contextual Eye Cursor For General Pointing In Windows Envir...Kalle
 
Pontillo Semanti Code Using Content Similarity And Database Driven Matching T...
Pontillo Semanti Code Using Content Similarity And Database Driven Matching T...Pontillo Semanti Code Using Content Similarity And Database Driven Matching T...
Pontillo Semanti Code Using Content Similarity And Database Driven Matching T...Kalle
 
Park Quantification Of Aesthetic Viewing Using Eye Tracking Technology The In...
Park Quantification Of Aesthetic Viewing Using Eye Tracking Technology The In...Park Quantification Of Aesthetic Viewing Using Eye Tracking Technology The In...
Park Quantification Of Aesthetic Viewing Using Eye Tracking Technology The In...Kalle
 
Palinko Estimating Cognitive Load Using Remote Eye Tracking In A Driving Simu...
Palinko Estimating Cognitive Load Using Remote Eye Tracking In A Driving Simu...Palinko Estimating Cognitive Load Using Remote Eye Tracking In A Driving Simu...
Palinko Estimating Cognitive Load Using Remote Eye Tracking In A Driving Simu...Kalle
 
Nakayama Estimation Of Viewers Response For Contextual Understanding Of Tasks...
Nakayama Estimation Of Viewers Response For Contextual Understanding Of Tasks...Nakayama Estimation Of Viewers Response For Contextual Understanding Of Tasks...
Nakayama Estimation Of Viewers Response For Contextual Understanding Of Tasks...Kalle
 
Nagamatsu User Calibration Free Gaze Tracking With Estimation Of The Horizont...
Nagamatsu User Calibration Free Gaze Tracking With Estimation Of The Horizont...Nagamatsu User Calibration Free Gaze Tracking With Estimation Of The Horizont...
Nagamatsu User Calibration Free Gaze Tracking With Estimation Of The Horizont...Kalle
 

More from Kalle (20)

Blignaut Visual Span And Other Parameters For The Generation Of Heatmaps
Blignaut Visual Span And Other Parameters For The Generation Of HeatmapsBlignaut Visual Span And Other Parameters For The Generation Of Heatmaps
Blignaut Visual Span And Other Parameters For The Generation Of Heatmaps
 
Zhang Eye Movement As An Interaction Mechanism For Relevance Feedback In A Co...
Zhang Eye Movement As An Interaction Mechanism For Relevance Feedback In A Co...Zhang Eye Movement As An Interaction Mechanism For Relevance Feedback In A Co...
Zhang Eye Movement As An Interaction Mechanism For Relevance Feedback In A Co...
 
Yamamoto Development Of Eye Tracking Pen Display Based On Stereo Bright Pupil...
Yamamoto Development Of Eye Tracking Pen Display Based On Stereo Bright Pupil...Yamamoto Development Of Eye Tracking Pen Display Based On Stereo Bright Pupil...
Yamamoto Development Of Eye Tracking Pen Display Based On Stereo Bright Pupil...
 
Wastlund What You See Is Where You Go Testing A Gaze Driven Power Wheelchair ...
Wastlund What You See Is Where You Go Testing A Gaze Driven Power Wheelchair ...Wastlund What You See Is Where You Go Testing A Gaze Driven Power Wheelchair ...
Wastlund What You See Is Where You Go Testing A Gaze Driven Power Wheelchair ...
 
Vinnikov Contingency Evaluation Of Gaze Contingent Displays For Real Time Vis...
Vinnikov Contingency Evaluation Of Gaze Contingent Displays For Real Time Vis...Vinnikov Contingency Evaluation Of Gaze Contingent Displays For Real Time Vis...
Vinnikov Contingency Evaluation Of Gaze Contingent Displays For Real Time Vis...
 
Urbina Pies With Ey Es The Limits Of Hierarchical Pie Menus In Gaze Control
Urbina Pies With Ey Es The Limits Of Hierarchical Pie Menus In Gaze ControlUrbina Pies With Ey Es The Limits Of Hierarchical Pie Menus In Gaze Control
Urbina Pies With Ey Es The Limits Of Hierarchical Pie Menus In Gaze Control
 
Urbina Alternatives To Single Character Entry And Dwell Time Selection On Eye...
Urbina Alternatives To Single Character Entry And Dwell Time Selection On Eye...Urbina Alternatives To Single Character Entry And Dwell Time Selection On Eye...
Urbina Alternatives To Single Character Entry And Dwell Time Selection On Eye...
 
Takemura Estimating 3 D Point Of Regard And Visualizing Gaze Trajectories Und...
Takemura Estimating 3 D Point Of Regard And Visualizing Gaze Trajectories Und...Takemura Estimating 3 D Point Of Regard And Visualizing Gaze Trajectories Und...
Takemura Estimating 3 D Point Of Regard And Visualizing Gaze Trajectories Und...
 
Stevenson Eye Tracking With The Adaptive Optics Scanning Laser Ophthalmoscope
Stevenson Eye Tracking With The Adaptive Optics Scanning Laser OphthalmoscopeStevenson Eye Tracking With The Adaptive Optics Scanning Laser Ophthalmoscope
Stevenson Eye Tracking With The Adaptive Optics Scanning Laser Ophthalmoscope
 
Stellmach Advanced Gaze Visualizations For Three Dimensional Virtual Environm...
Stellmach Advanced Gaze Visualizations For Three Dimensional Virtual Environm...Stellmach Advanced Gaze Visualizations For Three Dimensional Virtual Environm...
Stellmach Advanced Gaze Visualizations For Three Dimensional Virtual Environm...
 
Skovsgaard Small Target Selection With Gaze Alone
Skovsgaard Small Target Selection With Gaze AloneSkovsgaard Small Target Selection With Gaze Alone
Skovsgaard Small Target Selection With Gaze Alone
 
San Agustin Evaluation Of A Low Cost Open Source Gaze Tracker
San Agustin Evaluation Of A Low Cost Open Source Gaze TrackerSan Agustin Evaluation Of A Low Cost Open Source Gaze Tracker
San Agustin Evaluation Of A Low Cost Open Source Gaze Tracker
 
Ryan Match Moving For Area Based Analysis Of Eye Movements In Natural Tasks
Ryan Match Moving For Area Based Analysis Of Eye Movements In Natural TasksRyan Match Moving For Area Based Analysis Of Eye Movements In Natural Tasks
Ryan Match Moving For Area Based Analysis Of Eye Movements In Natural Tasks
 
Rosengrant Gaze Scribing In Physics Problem Solving
Rosengrant Gaze Scribing In Physics Problem SolvingRosengrant Gaze Scribing In Physics Problem Solving
Rosengrant Gaze Scribing In Physics Problem Solving
 
Porta Ce Cursor A Contextual Eye Cursor For General Pointing In Windows Envir...
Porta Ce Cursor A Contextual Eye Cursor For General Pointing In Windows Envir...Porta Ce Cursor A Contextual Eye Cursor For General Pointing In Windows Envir...
Porta Ce Cursor A Contextual Eye Cursor For General Pointing In Windows Envir...
 
Pontillo Semanti Code Using Content Similarity And Database Driven Matching T...
Pontillo Semanti Code Using Content Similarity And Database Driven Matching T...Pontillo Semanti Code Using Content Similarity And Database Driven Matching T...
Pontillo Semanti Code Using Content Similarity And Database Driven Matching T...
 
Park Quantification Of Aesthetic Viewing Using Eye Tracking Technology The In...
Park Quantification Of Aesthetic Viewing Using Eye Tracking Technology The In...Park Quantification Of Aesthetic Viewing Using Eye Tracking Technology The In...
Park Quantification Of Aesthetic Viewing Using Eye Tracking Technology The In...
 
Palinko Estimating Cognitive Load Using Remote Eye Tracking In A Driving Simu...
Palinko Estimating Cognitive Load Using Remote Eye Tracking In A Driving Simu...Palinko Estimating Cognitive Load Using Remote Eye Tracking In A Driving Simu...
Palinko Estimating Cognitive Load Using Remote Eye Tracking In A Driving Simu...
 
Nakayama Estimation Of Viewers Response For Contextual Understanding Of Tasks...
Nakayama Estimation Of Viewers Response For Contextual Understanding Of Tasks...Nakayama Estimation Of Viewers Response For Contextual Understanding Of Tasks...
Nakayama Estimation Of Viewers Response For Contextual Understanding Of Tasks...
 
Nagamatsu User Calibration Free Gaze Tracking With Estimation Of The Horizont...
Nagamatsu User Calibration Free Gaze Tracking With Estimation Of The Horizont...Nagamatsu User Calibration Free Gaze Tracking With Estimation Of The Horizont...
Nagamatsu User Calibration Free Gaze Tracking With Estimation Of The Horizont...
 

Understanding the Benefits of Gaze Enhanced Visual Search

  • 1. Understanding the Benefits of Gaze Enhanced Visual Search Pernilla Qvarfordt, Jacob T. Biehl, Gene Golovchinsky, Tony Dunningan FX Palo Alto Laboratory, Inc. 3400 Hillview Bldg. 4, Palo Alto, CA 94304 {pernilla, biehl, gene, tonyd}@fxpal.com Abstract For instance, a study investigating the utility of triple examination showed error rates still at 11% [Markus et al. 1990]. Error rates go even higher when non-skilled examiners (e.g., MDs other than In certain applications such as radiology and imagery analysis, it is Radiologists) are involved [Shaw et al. 1990]. A central recom- important to minimize errors. In this paper we evaluate a struc- mendation from many of these studies is to employ a systematic tured inspection method that uses eye tracking information as a methodology in how images are examined [Goddard et al. 2001]. feedback mechanism to the image inspector. Our two-phase That is, there is a need for a more formal, structured analysis meth- method starts with a free viewing phase during which gaze data is odology that ensures full examination of all anatomical compo- collected. During the next phase, we either segment the image, nents [Vock 1987]. mask previously seen areas of the image, or combine the two tech- niques, and repeat the search. We compare the different methods proposed for the second search phase by evaluating the inspection Similar recommendations have been made for other critical task method using true positive and false negative rates, and subjective domains, such as transportation security. Many of the recommen- workload. Results show that gaze-blocked configurations reduced dations call for changing the current methodology from only the subjective workload, and that gaze-blocking without segmenta- requiring secondary screening on potential true positives to requir- tion showed the largest increase in true positive identifications and ing mandatory multi-stage screening [Butler and Poole 2002, the largest decrease in false negative identifications of previously Fenga et al. 2008]. While false negative statistics are not openly unseen objects. published for the domain, such recommendations would not be made if there was not a meaningful need to reduce false negative error rates. CR Categories: H.5.2 [Information Interfaces and Presentation]: User Interfaces In this paper, we present a novel methodology for performing structured image analysis that is both systematic and adaptive to an Keywords: gaze-enhanced visual search, two-phase search, multi- examiner’s search behavior. The goal of the methodology is to pro- ple targets. vide a more structured image examination process. Specifically, we propose a two-phase process that consists of an initial phase of 1 Introduction free search, followed by a second phase designed to help the exam- iner to cover the whole image. This second phase can consist either Human-based image analysis is a commonly performed task in of the image being divided into smaller sub-images (segments), of many contemporary professions. For instance, radiologists and having previously viewed (fixated) segments blocked out, or of a other medical professionals frequently examine medical images to combination of these techniques. This methodology was designed diagnose and treat patients, airport security agents scan x-rays of to increase the total coverage of examination while reducing luggage for prohibited items, and factory workers perform visual redundant examination with the goal of reducing the overall num- inspection of goods to assure quality. In these tasks, the examiner ber of false negatives, particularly false negatives that the exam- must combine knowledge of the domain and a high degree of men- iner did not view or judge. Our proposed methodology is not tal concentration within a short amount of time to classify or inter- domain dependent; we believe it can be applied to many domains, pret the images. including security images, satellite images, maps, astronomical images, and scientific visualizations. While there have been many advances both in the technology used to create images and in the training of human examiners, many We evaluated the inspections techniques proposed for the struc- image analysis tasks are still prone to significant error. For exam- tured viewing in a controlled laboratory study. The results shows ple, Goddard et al. [2001] showed that radiological image exami- that masking previously well inspected areas reduced the subjec- nation still has rates nearing 20% for clinically significant or major tive workload. When the previous gaze is masked over the whole errors. Even with multiple examination (multiple radiologists image, the increase of true positive identifications and reduction of investigating an image or set of images), the error rate is still high. false negative identifications are larger than in the other combina- tions. We believe these results can provide insights and design les- sons for the construction of new visual search systems and tech- niques that improve search performance. . Copyright © 2010 by the Association for Computing Machinery, Inc. Permission to make digital or hard copies of part or all of this work for personal or classroom use is granted without fee provided that copies are not made or distributed 2 Related work for commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, to republish, to post on Research on visual search have been pursued in two directions, servers, or to redistribute to lists, requires prior specific permission and/or a fee. either to investigate the perceptual and cognitive processes under- Request permissions from Permissions Dept, ACM Inc., fax +1 (212) 869-0481 or e-mail permissions@acm.org. lying visual search, or to design procedures or tools for improving ETRA 2010, Austin, TX, March 22 – 24, 2010. © 2010 ACM 978-1-60558-994-7/10/0003 $10.00 283
  • 2. the performance of professional image analysts. Our research falls looking for suspected tumors in an X-ray, or for camflaged weap- in the latter category, but findings from perceptual and cognitive ons in a satellite image. studies provide additional context. 2.2 Improving Visual Inspection 2.1 Visual Search Visual inspection is important in many domains such as finding Characteristics of visual search have been investigated extensively tumors in X-rays and identifying threats in airport luggage scans. in cognitive psychology and perception. This line of research is Yet, visual inspection is error prone. Most attempts to lower the focused on building models of perception and of higher-level cog- error has focused on training, see for example [Kollera, Drury, and nitive processes. Although visual search studies are often labora- Schwaninger, 2009]. Nickles, Melloy and Gramopadhye [2003] tory studies removed from real world image search tasks, there are used dynamic tool for training inspectors to follow a specific scan- findings that carry over to the work of professional image analysts. ning pattern during the inspection, but they could not show that the dynamic training was better than verbal instructions. Another When searching for an item in an image, the eye scans the image to study [Sadasivan et al. 2005] used gaze information from experts find the item. The character of these scanpaths has caused a long to create a visualization of a preferred scanpath for cargo bay debate within psychology. Both early observations, e.g. [Norton inspection. This visualization was used to train novice inspectors and Stark 1971] and more recent studies, e.g. [Scinto, Pillalamarri, and proved to be successful in improving performance. and Karsh 1986], have characterized the scanpaths as random. Although people tend to look at similar things in the image, people Rather than focus solely on training, others have focused on sup- rarely follow the same scanpath, i.e., they look at different parts of porting the inspection task. For instance, Haimson et al. [2004] the image in different order. added partitioning to a radar screen and thereby improved perfor- mance in finding targets. Forlines and Balakrishnan [2009] used However, scanpaths do not always appear random. Findlay and image segmentation to separate objects in the image and to recom- Brown [2006] showed that people can employ different strategies pose the objects. Techniques for recomposing the image showed when searching through images with randomly placed targets. One such strategy, for example, is to follow the global external counter promise in reduced error rate, but longer search time was also that the objects create. When the displays are arranged in a more found for two of the recomposition techniques. Image segmenta- structured ways, e.g. in a grid pattern, people tend to read the grid tion as an underlying technique may work well on some kinds of as they would read text [Gilchrist and Harvey 2006]. The practical targets. Forlines and Balakrishnan focused on blood cell inspec- implications from scanpath theories is that expert inspectors tion; other type of images decomposition may not be as easy to appear to develop structured scanpaths [Sadasivan et al. 2005]. decompose as cell slides used in their experiment. There is, however, evidence from a luggage-screening experiment that performance gains appear to be related to changes in ability to As mentioned above, Nodine and Kundel [1987] found that the recognize targets rather than to changes in scanpaths [Mccarley et fixation times for false positive responses were lower than those of al. 2004]. true positive when inspecting chest X-rays. They used this knowl- ege in a system which collected gaze information on where radiol- There is also evidence that the characteristics of fixation change ogists looked during a free form inspection. In a second phase, depending on judgments viewers make of the items in view. Pom- they asked the radiologists to re-inspect areas they had dwelled on plun, Reingold and Shen [2001] found that people who were asked for an extended time, but in which they did not find any incidence to match two sets of images fixated for a longer time on matching of tumors. In that work, participants were not asked to view areas objects compared to mismatched objects. Manning, Ethell and they had not looked at previously, but only to re-evaluate already Donovan [2004] showed that when radiologists inspected chest x- looked at areas. rays, their gaze duration was half as long on lesions they did not report as a detected tumor as on lesions they reported. Another 3 Inspection Methodologies study [Nodine and Kundel 1987] reported longer gaze durations when giving a true positive judgment about a lesion compared to a false negative judgment. These results suggest that it might be pos- In this work we compare four different inspection methodologies sible to use gaze durations to provide feedback to the user of which for visual search. These methodologies follow a two-phase inspec- parts of an image have been well inspected and which parts have tion approach. In the first phase, searchers are allowed to view the not. image unconstrained, enabling them to gain an understanding of the image and its gross structure. In the second phase, the view of One interesting finding is that people have a higher error rate when the image is restricted to create a more structured search. Below targets are rare than when targets occure frequently [Wolfe et al. we describe each methodology studied and the rationale for its 2007]. A practical example of this is that when a radiologist selection. screens mammograms for cancer, very few mammograms will show any signs of tumors. Research in low prevalence has sug- Structured Segmentation. Inspired by previous work that found gested several ways to counter the effect, such as rewarding image performance benefits in systematic search of images [Sadasivan et searchers appropriately for finding targets [Navalpakkam, Kock al. 2005] and [Nickles et al 2003], this technique subdivides the and Perona 2009], and allowing people to go back and correct mis- image into a tiled grid and then displays the grid tiles serially takes [Fleck and Mitroff 2007]. (Figure 1). By reducing the viewable search area at any one time, this technique forces the user to focus attention on the currently- In most of perception studies that explore low prevalence of tar- visible area. It also ensures that all parts of the display get equal gets, participants are asked to find only one target. While the over- attention and reduces distractions from other parts of the display. all true positive rate maybe low, targets often do not exist in isola- In our study, we divided each image into a three by three grid, tion. That is, there are likely to be many true positives in a single keeping the area small enough to be able to quickly scan and mark image, all of which need to be identified to correctly perform the during the time limit set for the experiment, and large enough to search task. Examples of such recall-oriented activity include contain information when gaze-blocking was also used. 284
  • 3. Phase I Phase II Phase I Phase II Figure 1: Structured segmentation Figure 3: Structured segmentation with gaze-blocking Phase I Phase I Phase II Phase II Figure 2: Gaze blocking Figure 4: Full-image search Gaze-Blocking. This methodology uses a searcher’s gaze behavior from the first free-viewing search phase to determine where the 4.1 Method searcher has previously inspected the image and blocks out those We used a 2x2 within-subjects design, where the independent vari- areas in the second phase to prevent unnecessary re-examination ables were type of image display (image display), and visibility of (Figure 2). The blocked-out areas are constructed from clusters of fixation clusters (gaze display). Each of the two independent vari- identified fixations from the first phase using a dispersion-based able had two levels for image display: full image or segmented fixation detection algorithm [Salvucci and Goldberg 2004]. Fixa- image, and for gaze display: with gaze-blocking or without gaze- tions are clustered using a minimum-spanning tree algorithm blocking. This design gave us four conditions for displaying the based on their distance to other fixations. Only fixations clusters image in phase 2: full image with previous gaze blocking (FB), with a total gaze duration of more than 270 ms were blocked out. segmented image with gaze-blocking (SB), full image without We based this threshold on 275 ms fixations for image search gaze-blocking (F), and segmented image without gaze-blocking reported in [Rayner 1998], and based on our own experimentation. (S). These conditions directly match the proposed inspection meth- odologies discussed in the previous section. This threshold ensured that fixation clusters reflected some higher level of cognitive processes. Each fixation in the clusters is repre- 4.2 Participants sented as a 40 pixel circle in our system, centered at the gaze loca- tion. This size was chosen as it represents the 1 visual degree of the Seven men and one woman participated in the study. Participants fovea at the distance of 60 cm from the computer screen using 96 were recruited using a broadcast email solicitation within our orga- pixels per inch. Each blocked-out region was replaced with a white nization. Participants’ age ranged between 30 and 60 years. All patch over the image. were screened before the study to make sure they had normal or corrected vision by contact lenses (4 participants wore contact Structured Segmentation and Gaze-Blocking. This technique com- lenses) and that the eye tracker could properly track their eyes. bines the previous two methodologies. The image is re-presented in segments, each segment containing gaze blocks of previously 4.3 Stimuli and Tasks inspected areas (Figure 3). This retains the advantages of struc- Careful consideration was given to the design of the task stimuli. tured search while eliminating redundant examination. Motivated by the fact that most inspection tasks are not only per- ceptual but also involve higher-level cognitive processes, we Full Image. In addition to these three methodologies, we also wanted to simulate a task that involved not only perception of tar- included a baseline condition in which the full image was re-pre- gets, but also associated judgment. Furthermore, we did not want sented to the searcher (Figure 4). to require specific domain knowledge to avoid introducing domain-specific and participant-specific effects. 4 Experiment We designed two types of targets and distractors: symmetric and asymmetric (see Figure 5 shapes in black outline). For symmetric We conducted an experiment to evaluate the inspection methodol- targets a mirror image of the target is indistinguishable from the ogies. We designed a controlled experiment to explore the effects target. Asymmetric targets on the other hand, have a mirror image of using both segmentation and occlusion of well-inspected areas. that when rotated do not match the target (e.g. in Figure 5 first gray 285
  • 4. shape to left of the assymetric target). Symmetric targets differed from distractors by the size of the insections (notches shown in Figure 5) in the basic shape. To identify symmetric targets, partici- pants needed to judge details of the target. In Figure 5 for example, the participants need to judge if the circular insection was of the same size as the example target.To identify asymmetric targets, participants needed to perform mental rotation of potential targets. Each target had one to three similarly looking distractors (close distractors, targets with gray outline in Figure 5). Targets were ran- domly assigned to locations where they would not overlap other targets or distractors. Targets and close distractors were not placed on segmentation boundries. Besides close distractors, stimulus images also include a random collection of distractors (random distractors). Targets and distractors were randomly rotated in steps of 30 degrees, starting a 0 degrees.. Figure 6: Example of stimuli image used in the study. symmetric asymmetric set to 100 ms. In addition, we only displayed fixation clusters in the gaze-block conditions where gaze duration was greater than Figure 5: Examples of symmetric and asymmetric targets. 270 ms. The black shape is targets and the gray close distractors. 4.5 Procedure In addition, the task required target objects to match a particular Each participant’s session took about one hour to complete. A color. To ensure color matching was non-trivial, we placed the tar- researcher first demonstrated to the participant the basics of the gets and distractors against uniquely generated, non-uniform col- experimental task by explaining and showing a sample search task. ored background and had distractors (both close and random) col- The participant was then asked to perform the experimental image ored in the same color. For the targets, we used five different basic search tasks. The experimental session was divided into four parts, colors (blue, green, orange, purple, and red), and three shades of one for each condition (baseline-full image, segmented, gaze- each color. Figure 6 shows an example of stimuli used with a mid- blocked, and combined segmented and gaze-blocked). Ordering of shade orange color as the target color. the conditions was counterbalanced using a 4x4 Latin Square. Figure 6 further illustrates the combination of requirements and constraints of the study tasks. A legend at the top of the image In each condition, participants started the task by calibrating the describes the target shape and the color separately. In all, 24 eye tracker. Next, participants were shown a screen that described images were generated. Four images included no targets matching the current experimental condition. Pressing the space bar loaded the description in the legend. Twenty images, ten with symmetric the first stimulus image. The image would be displayed for 67.5 targets and ten with asymmetric targets, were generated with a ran- seconds (Phase 1). Next, the screen was blanked for three seconds. dom number of targets ranging between 5-20. In addition each The image was then displayed according to one of the four condi- image had between 10-40 target shapes with wrong color, 35-40 tions (Phase 2). The participant was given another 67.5 seconds to close distractors with the target correct color, 10-40 close distrac- inspect the image and find targets. In conditions including segmen- tors with wrong color. Each stimulus image had 260-300 shapes in tation of the image, each segment was shown for 7.5 seconds. total. The difference between the total number of shapes and the Another three second break was given between successive tasks, targets plus close distractors were filled with random distractors in for a total of 24 tasks per participant. random colors. All shapes had approximately the same size, around 24 pixels high and wide. A total of six images were presented in each condition. The first image was a training image to familiarize participants with the pre- 4.4 Experimental Setup sentation method and timing in the second phase. One image con- tained no targets. The no-target image and the training image were A ViewSonic 18” CRT monitor set at 1024 x 768 pixels resolution not included in the analysis. The order of images after the training was used to present task stimuli to the participants. A CRT was chosen over a LCD panel as it showed consistent colors indepen- image was randomized within each condition. Tasks were bal- dent of viewing angle. This was important because the task anced over conditions. required participants to carefully judge colors. A Tobii X120 eye tracker was positioned in front of the monitor. Participants were Before repeating the search task for the remaining conditions, par- seated approximately 60 cm from the monitor. ticipants filled out a NASA TLX questionnaire [Hart and Staveland 1988] for assessing their subjective workload. After participants The Tobii X120 was configured to collect gaze data at a rate of 60 had gone through all four conditions, they ranked the rating scales, Hz. We found that this rate meet the input requirements of our sys- as specified in the final step in the NASA TLX procedure. Finally, tem and provided more robust results compared to higher collec- the participants were asked a few debriefing questions about their tion rates. The dispersion threshold for the fixation detection algo- experience of using the four conditions. They were also asked to rithm was set to 40 pixels. The minimum duration of a fixation was rank the four conditions according to their preference. 286
  • 5. 4.6 Data Analysis Condition TP TPadj FN FNadj In all analysis, if nothing else is mentioned, we used a 2x2 full fac- Overall 86% 88% 14% 11% torial repeated measurements ANOVA. We expected that partici- Block 83% 88% 17% 11% pants’ performance would not differ during the first phase, the free viewing phase. This assumption was confirmed on all dependent No Block 88% 88% 12% 12% measures reported in the results section. Hence, we focused our Segmentation 85% 87% 15% 13% analysis on the performance during the second phase. We looked at Full Image 86% 90% 14% 10% the True Positive (TP) rate, the False Positive (FP) rate and the Block + seg 83% 86% 14% 10% False Negative (FN) rate. The false negative rate requires some additional elaboration because it consists of items viewed (gazed Block + full image 83% 90% 14% 7% at) and not judged as matching, and also of items not viewed at all. No Block + seg 87% 87% 11% 11% Thus we calculated FNV and FNNV rates (where FNV + FNNV = No Block + full image 90% 90% 8% 8% FN). We calculated the overall performance in phase 2 and an adjusted performance based on visible targets not selected for TP Table 1: TP, TPadj, FN, FNadj rates at end of phase 2. and FN. The visible targets were the total number of targets present in the image minus the targets blocked or partially blocked by the False positives classifications (FP) were rare in our experiment. gaze-blocking. The adjusted TP and FN hence reflect the number of Across all conditions, only 6% of the total number of selections targets in the second phase that the participants could see and were FPs (SD=7.6). No significant difference was found between judge in full. conditions. We analyzed patterns of gaze on targets, and found that partici- The false negative rate (FN), incorrectly identifying a true target as pants often fixated multiple times on a target. For this reason we false, was 15% (SD=10.9) across all conditions. During the second analyzed total gaze duration on the target rather than the fixation phase, participants reduced FNs by, on average, 59% (SD=21.7). durations. This method of analysis is also more consistent with We did not find any main effects for gaze display and image dis- how fixation clusters were calculated. play variables. 5 Results For many image inspection tasks, reducing the FN rate is impor- tant. In this work, we are particularly interested in reducing the FNNV rate. We compared the number of FNNV outcomes in phase 1 5.1 Performance and Error with the three different outcomes after phase 2: FNNV (still not In any inspection task, it is important to have as many true posi- viewed), FNV (viewed but not selected) and TP (viewed and tives (TPs), correctly identifying a true target, as possible. In our selected). Table 1 shows ratios for these three categories for all study the participants found on average 86% (SD=10.7) of all tar- conditions. As the table illustrates, we again did not find any main gets after both phases across all conditions. In phase 2, they effects, but we found a significant interaction for the FNNV in improved their performance on average 31% (SD=20.2) across all phase 1 transitioning to TPs durring phase 2 (F(1, 7)= 6.321, conditions. See Table 1 for a summary of TP and TP adjusted for p<0.05). In line with the previously found interaction, the combi- visible targets (TPadj) for all conditions. We did not find any main nation of gaze-blocked and no segmentation (FB) gave more TP effect on TP or TPabjfor the independent variables image display from previously not seen targets, an average of 82% compared to and gaze display, or any effect of gaze blocking or image segmen- the overall mean of 67%. tation. However, when testing the improvement of TP from phase 1 to phase 2 adjusted for visible targets, we found a border- 5.2 Revisitations of Targets and Image Search line significant interaction between gaze blocking and segmenta- Strategy tion (F(1,7)=4.589, p=0.069). As Figure 7 shows, the combination of gaze-blocking and no segments gave a higher improvement in Previous literature has shown that people revisit targets and that adjusted TP rate (49%) compared to all other combination of con- revisitation may change the person’s judgement or classification ditions. 100 [Nodine and Kundle 1987]. In our study, the gaze-blocking condi- No segmentation Segmentation 100 No segmentation Segmentation 80 80 Percent change from FN not viewed to TP Percentage increase of TP rate 60 60 40 40 20 20 0 0 BLOCK NO BLOCK BLOCK NO BLOCK Figure 7: Average improvement of TP rate adjusted for Figure 8: Average transition of FN not viewed in phase 1 to visible targets. Error bars indicate ±1 standard error TP during phase 2. Error bars indicate ± 1 standard error. 287
  • 6. Condition FNV FNNV TP 1.0 FN TP Overall 19% 13% 68% Block 19% 14% 66% 0.8 No Block 19% 11% 69% Segmentation 26% 15% 59% Survival rate of gaze 0.6 Full Image 12% 11% 77% Block + seg 29% 20% 51% 0.4 Block + full image 10% 8% 82% No Block + seg 24% 9% 67% No Block + full image 15% 13% 71% 0.2 Table 2: Ratio of FNNV from phase 1 to FNV, FNNV, or TP at end of phase 2. 0.0 0 500 1000 1500 2000 2500 3000 Gaze duration (ms) 1200 No segmentation Segmentation Figure 10: Survival rate of gaze duration on TP and FN targets during phase 1. Average gaze duration (ms) for FNs during phase I 1000 analysis indicates that participants likely used different strategies 800 for searching for targets in different inspection conditions. During the first phase of the gaze-blocking conditions, participants likely searched for easily identifiable targets and used the second phase 600 to identify targets that required a higher degree of examination to distinguish targets from distractors. For example, one participant 400 indicated in the debriefing interview that he would examine less dense clusters of objects first, and then move on to examine more dense object clusters. Another participant stated he would try to 200 identify objects “that jumped out at [him]” first. 0 5.3 Subjective Workload and Preference BLOCK NO BLOCK Although the differences between the different conditions for the Figure 9: Average duration on FN during phase 1. Error bars task load index (TLX) were small in magnitude, we found a signif- indicate ± 1 Standard Error icant main effect for the gaze display variable (F(1, 7)=6.905, p<0.05). When gaze-blocking was present, participants experi- tions make areas previously visited non-viewable if they are well enced a significantly lower workload (mean=63.6) compared to it inspected. The goal is to guide examiners to focus on finding new was not present (mean=67.5). This correspond to 6% reduction in targets. If people frequently revisit targets and change their mind, task load. We did not find any effect due to segmentation. this might negatively effect the performance of the gaze-blocked level. For this reason, we investigated how frequently participants Participants were asked to rank their preferred methodology from revisited visible targets and how often they changed judgments 1 (most preferred) to 4 (least preferred). Differences among meth- from FN to TP. We found that 51% (SD=23.1) of all new TP in odologies were small and no one methodology was overwhelming phase 2 had been viewed in phase 1. This difference was not sig- preferred across all participants. The highest overall rated method- nificant, F(1, 7)=.222, ns. There was no main effects for the two ology was no segmentation and no gaze-blocking (F: summed independent variables. score: 17), followed by no segmentation and gaze-blocking (FB: summed score: 18). However, when we correlated the preference This result shows that adding a second inspection phase is benefi- scores with the measured workload, we found a significant correla- cial since people improve performance by re-examining potential tion (r=0.45, t(30)=2.785, p<0.01). That is, participants tended to targets. Further, it appears that the gaze-blocking did not signifi- prefer conditions where they reported lowest subjective workload. cantly affect the number of targets being re-examined. The number of FNV blocked in phase 2 was small: only three of eight partici- 5.4 Fixation Patterns on Targets pants experienced any blocking of FNVs, and of these, two partici- pants had FNVs blocked in both gaze-blocking conditions, where For a gaze-blocking to be feasible, gaze durations on targets need on average 17% of FNV were blocked. to be distinguishable. When designing gaze-blocking, we relied on observations that fixations on TPs were longer than those on FNs When examining the gaze duration in phase 1 of viewed FN targets [Nodine and Kundle 1987] in the domain of radiology. We exam- which the participants changed to TP after re-examination, we saw ined gaze durations on targets, and found that in our domain too a difference in gaze duration between the two levels in the gaze TPs were gazed at longer than FNs. The average duration of FNs display variable (Figure 9). On average the gaze duration in the were 661ms (SD=628.5), and for TPs 1386 ms (SD=1032.5). This gaze-blocking conditions was 443 ms (SD=251.2), while in the difference was significant (F(1,7)=16.533, p<0.01). This result non-blocking conditions it was 746 ms (SD=493.7). This differ- shows that the duration of gaze on targets differed depending the ence was very close to significant (F(1, 7)=5.574, p=.0503). This participant’s interpretation of the target. 288
  • 7. As illustrated in Figure 10, there is a definite pattern in the data Our results showed that slightly less than half of the performance which indicates TP, on average, are dwelled on longer by examin- gain when previous gaze was blocked in phase 2 came from previ- ers compared to FN. For instance, at 550 ms, the participants had ously non-viewed objects. This is a promising result, in particular stopped looking at 50% of FNs, but only stopped looking at 32% of for situations where image analysis is done under time pressure. the TPs. The 50% gaze duration survival rate for TPs was at 1032 When time is unlimited, more targets may be viewed during the ms. On the other hand, at gaze duration of 1049 ms, 70% of the FN first phase. An interesting research issue to look at in the future is were shorter. Although both TPs and FNs has a large spread, these how people can optimize their performance using a two phase numbers reaffirm that it should be possible to use gaze duration as inspection method and what trade-offs the image analyst faces an implicit relevance feedback mechanism to support image analy- when previous gaze is blocked during a second phase. sis. Revisitation of previously looked at targets is another important We also found a significant difference in fixation duration between factor for the performance increase during phase 2. Although gaze phase 1 and phase 2 (F(1, 7)=71.597, p<0.001). During phase 2, blocking masked previously well-inspected areas, we could not participants’ fixation duration was on average 813 ms (SD=749) find a significant impact of gaze-blocking during the second phase and in phase 2 the duration was on average 1235 ms (SD=1037). in terms of revisitations. In our study, a low percentage of the pre- viously viewed and un-selected targets were blocked out, and in There was no difference between the conditions. This result is fur- particular, in gaze-blocking and full-image conditions, participants ther evidence that the participants shifted strategy between the two made up for that performance loss by finding previously-unseen phases. targets. This is a promising result, since one of our goals was to lower the number FNs during the second phase. 6 Discussion Another interesting finding in our study is that participants Intuitively, masking off where you have looked during an image adjusted their behavior during the gaze-blocking condition. During search task is both a bad idea and a compelling idea. When viewed the first phase, they adopted the strategy of scanning the image for parts of the image are blocked out, the searcher loses context and it the easiest to find targets. In the next phase, they would focus on is not possible to correct previous decisions. On the other hand, in targets that were harder to find and harder to interpret. It is possi- a complex image it can be hard to identify all occurrences, and ble that this strategy contributed to lowered workload in the gaze- blocking condition. This result is intriguing since it suggests more even to remember what you looked at already. In addition, when research on how to design image search techniques that can pro- the task involves not just perceptual processes but higher level mote a more efficient use of time and mental resources. cognitive processes, fatigue sets in quickly due to the demanding nature of the task, resulting decreased performance. We described Gaze patterns collected during study confirm that it is possible to four different inspection methods using combinations of gaze- give indications to image searchers of which parts of an image blocking and image segmentation techniques. In the study, we took they have not inspected in enough detail. As also observed in other an extreme approach to understand limitations of the methods. The studies, we found that on average, participants spent significantly performance results from our study show two things: masking an longer looking at targets they ended up selecting (TPs) than those image based on where searchers looked previously did not nega- they discarded as distractors (FNs). Because of the large individual tively impact performance, and masking gazed-at areas has poten- difference in gaze duration on TP targets and FN targets, thresholds tial. for gaze-blocking algorithms may need to be set dynamically based on each user’s behavior and preference, or by using machine The results that gaze blocking reduced participants’ workload is an learning approaches. In systems using manual pointing to for important finding. Image search can be a demanding task and if the marking selections, users’ selections can also be used to get gaze workload can be reduced, performance can be improved. We durations for TPs. In this study, we only looked at the sum of all believe it is important to find methods supporting image analysts fixations on the target. Other gaze characteristics, such as number that also reduce their workload. of fixations on target, duration between fixations, etc. can poten- tially be informative. The image search task may also influence In terms of performance, not surprisingly the gaze-blocking condi- what method the system uses to determine which parts of the tion did on average somewhat worse in true positive rate, in partic- image would be helpful for the user to look at in more detail. ular true positive rate not adjusted for visible targets. This differ- ence was not significant, and hence could just as well as have been Our study does not describe a realistic system; it merely tests the caused by other factors than those manipulated. However, when limits and the feasibility of using eye tracking to support image adjusted for targets blocked by the gaze mask, the gaze-block in search by masking viewed areas. A realistic system would need combination with whole image showed a larger increase in the TP ways to make correction, which in the study was only possible on rate and a larger reduction in FN during phase 2 than the other con- non-masked areas. A realistic system would also need to allow ditions. This result indicates that allowing participants to focus on image analysts to control the pace, segmentation, and masking of areas not previously viewed helped them find and select more of viewed areas. However, results from this study are encouraging, the remaining targets compared to the other three inspection meth- indicating that eye tracking can provide useful support during ods. This result is promising for future work. image search. Segmenting the image did not seem to affect the performance, 7 Conclusion expect in combination with gaze block. The SB condition had the overall lowest performance. Interestingly, when the transition rate In this paper, we proposed a two-phased approach to image search from FNNV to TP were highest in gaze block on full image (FB), it and investigated the performance, workload, and user preference was also lowest in SB. It appears that restricing the viewable area trade-offs of four different second phase image re-presentation too much has negative affects on performance. Future research is methodologies. Results from a controlled laboratory study showed needed to explain this phenonemon. that in a two-phase approach the gaze-blocking treatment, which 289
  • 8. masks well examined regions of the image in the second phase of gage screening due to training. Ergonomics, 52(6), 644–656. inspection, significantly reduced overall subjective workload. Gaze block on a full image increased the transition of previously MANNING, D.J., ETHELL, S.C. AND DONOVAN, T. 2004. unseen false negatives to true positives, and was nearly as pre- Detection or decision errors? Missed lung cancer from posterante- ferred as free-form search. Further, our results also suggest that rior chest radiograph. British Journal of Radiology, 77, 231-235. segmentation and gaze-blocking likely encourage searchers to MCCARLEY, J.S, KRAMER, A.F., WICKENS, C.D., VIDONI, adopt an more structured search strategy in both phases of exami- E.D. AND BOOT, W. R. 2004. Visual Skills in Airport-Security nation. Results from our study can have broad impact on the Screening. Psychological Science, 15(5), 302-306. design of future techniques and systems that aid in visual search. NAVALPAKKAM, V., KOCH, C., AND PERONA, P. 2009. Homo economicus in visual search. Journal of Vision, 9(1):31, 1- 8 Acknowledgments 16. NODINE, C. F. AND KUNDEL, H. L. 1987. Using eye move- We thank all our participants for their time and effort, and Lynn ments to study visual search and to improve tumor detection. Wilcox and Larry Rowe for supporting this research. RadioGraphics 7(6) 1241-1250. NORTON, D., AND STARK, L. 1971. Eye movements and visual 9 References perception. Scientific American, 224, 34-43. BUTLER, V. AND POOLE, R.W. 2002. Rethinking Checked-Bag- MARKUS, J.B., SOMERS, S., O’MALLEY, B.P., AND STE- gage Screening. Reason Public Policy Institute; July. 25p. Policy VENSON, G.W. 1990. Double-contrast barium enema studies: Study No. 297. effect of multiple reading on perception error. Radiology 175, 155- 156. FENGA, Q., HANDE, S., AND KAPURB, K.C. 2008. Designing POMPLUN, M., REINGOLD, E.M. AND SHEN, J. 2001. Investi- Airport Checked-Baggage-Screening Strategies Considering Sys- gating the visual span in comparative search: the effect of task dif- tem Capability and Reliability. Reliability Engineering & System ficulty and divided attention. Cognition, 8, B57-B67. Safety. 94(2) 618-627. RAYNER, K. 1998. Eye movements in reading and information FLECK., M.S. AND MITROFF, S. 2007. Rare targets are rarely processing: 20 years of research. Psychological Bulletin, 124(3), missed in correctable search. Psychological Science, 18(11), 943- 372-422. 947. SADASIVAN, S., GREENSTEIN, J. S., GRAMOPADHYE, A.K., FORLINES, C., BALAKRISHNAN, R. 2009. Improving visual DUCHOWSKI. 2005. Use of eye movement as feedforward train- search with image segmentation. In Proceedings of CHI’2009, Pp. ing for a synthetic aircraft inspection task. In Proceedings of 1093-1102. CHI’2005, 141-149. GILCHRIST, I. D. AND HARVEY, M. 2006. Evidence for a sys- SANTELLA, A. AND DECARLO, D. 2004. Robust clustering of tematic component within scan paths in visual search. Visual Cog- eye movement recordings for quantification of visual interest. In nition, 14, 704-71 Proceedings of ETRA’04. Pp27-34 GODDARD, P., LESLIE, A., JONES, A., WAKELEY, C., AND SCINTO, L. F., PILLALAMARRI, R., KARSH, R. 1986. Cogni- KABALA, J. 2001. Error in radiology. British Journal Radiology tive strategies for visual search. Acta Psychologica. 62(3), 263- 74, 949-951. 292. HAIMSON, C, BOTHELL, D., DOUGLASS, S.A., ANDERSON, SHAW, N., HENDRY, M., AND EDEN, O. 1990. Inter-observer J.R. 2004. Partitioning Visual Displays Aids Task-Directed Visual variation in interpretation of chest X-rays. Scott Med J 35, 140- Search. Human Factors, 46(3), 551-566 141. HART S.G., STAVELAND L.E. 1988. Development of a NASA- VOCK, P. 1987. Frequent errors in the interpretation of chest x-ray TLX (Task Load Index): Results of Empirical and Theoretical films. Schweiz Med Wochenschr 117, 928-932. Research.” In Hancock PS, Meshkati N, (eds). Human Mental Workload. Amsterdam: North-Holland: pp. 139–83. WOLFE, J.M., HOROWITZ, T.S., VAN WERT, M.J., KENNER, N.M., PLACE, S.S., KIBBI, N. 2007. Low target prevalence is a KOLLERA, S.M., DRURY, C.G., AND SCHWANINGER, A. stubborn source of errors in visual search tasks. Journal of Experi- 2009. Change of search time and non-search time in X-ray bag- mental Pshychology: General. 136(6), 623-638. 290