- Original article
- Open Access
Web pages: What can you see in a single fixation?
© The Author(s) 2018
- Received: 18 August 2017
- Accepted: 23 March 2018
- Published: 9 May 2018
Research in human vision suggests that in a single fixation, humans can extract a significant amount of information from a natural scene, e.g. the semantic category, spatial layout, and object identities. This ability is useful, for example, for quickly determining location, navigating around obstacles, detecting threats, and guiding eye movements to gather more information. In this paper, we ask a new question: What can we see at a glance at a web page – an artificial yet complex “real world” stimulus? Is it possible to notice the type of website, or where the relevant elements are, with only a glimpse? We find that observers, fixating at the center of a web page shown for only 120 milliseconds, are well above chance at classifying the page into one of ten categories. Furthermore, this ability is supported in part by text that they can read at a glance. Users can also understand the spatial layout well enough to reliably localize the menu bar and to detect ads, even though the latter are often camouflaged among other graphical elements. We discuss the parallels between web page gist and scene gist, and the implications of our findings for both vision science and human-computer interaction.
- At a glance
- Web page gist
- Web design
- Visual perception
A fundamental constraint on performance in any visual task is the information available at a glance. Searching for a person in a natural scene will be more difficult if we cannot detect them at a glance. Being able to at least extract target-relevant features at a glance will speed up search; otherwise, we may have to execute a number of eye movements to find the target. Our ability to quickly get the gist of a scene can also speed up search if it provides layout information, identifying likely locations to find a person.
In the context of navigating a web page (e.g., see Fig. 1), clearly much of the pertinent information comes in the form of text that may not be readable at a glance. However, if a user can quickly determine that a web page is a blog, and thus, say, an unreliable source of information about drug interactions, she can quickly navigate to another page. (Of course, in viewing a page of search results, a savvy user might also realize that a given link points to a blog based on its URL. However this cue might not always be so readily available.) To the extent that a user can determine the layout of the page at a glance, she can easily direct her attention to a paragraph of interest, click on a button relevant to her goals, or look through the menu for more choices. At the other extreme, if the user cannot get much information out of a page at a glance, she may be forced to read a significant amount of the text, or otherwise scan the page, a slow and perhaps frustrating process. The user may decide simply to navigate to a new page that is more easily comprehensible – an undesirable outcome from the point of view of the page designer, the owner of the page, and any companies with ads on that page. Understanding the information available at a glance constrains models of perception, visual representation, and attention, and informs our understanding of usability and design.
A considerable amount of information about a stimulus, such as a scene or display, is available in a single fixation. This summary information has been termed the “gist.” Colloquially, the gist is defined as “the sentence one uses to describe a stimulus.” Often this is operationalized as “the perceived contents of a scene given a certain amount of viewing time” (Fei-Fei, Iyer, Koch, & Perona, 2007), often in a single fixation (Fei-Fei et al., 2007; Oliva, 2005). We take the gist to mean the information available at a glance, i.e., in a single fixation. Such a fixation can last between 100 and 300 milliseconds (ms) (Harris, Hainline, Abramov, Lemerise, & Camenzuli, 1988; Pieters & Wedel, 2012; Wedel & Pieters, 2000), while typical fixations fall in the range of 200–250 ms (Rayner & Castelhano, 2007).
At a glance, participants can identify the category of a natural scene (e.g., beach vs. forest, indoor vs. outdoor, parking lot vs. downtown) (Ehinger & Rosenholtz, 2016; Greene & Oliva, 2009b; Joubert, Rousselet, Fize, & Fabre-Thorpe, 2007; Rousselet, Joubert, & Fabre-Thorpe, 2005), and how much room there is to navigate (Greene & Oliva, 2009a). They can determine whether a given object is present, such as an animal (Kirchner & Thorpe, 2006; Li, VanRullen, Koch, & Perona, 2002; Thorpe, Fize, & Marlot, 1996), vehicle (VanRullen & Thorpe, 2001), or a human face (Crouzet, Kirchner, & Thorpe, 2010). They can reliably distinguish between cities (e.g., Paris vs. Los Angeles) and tell what kind of intersection lies ahead (Ehinger & Rosenholtz, 2016). Furthermore, experiments in which participants freely report what they perceived in the scene, as opposed to merely carrying out a pre-defined task, have revealed the richness of the perception of lower and mid-level properties, such as the colors and textures present (Fei-Fei et al., 2007).
In addition to the extensive research on natural scenes, much of vision research has (effectively) studied vision at a glance using artificial, psychophysics-style stimuli (e.g., Gabors, simple 2D/3D shapes, synthetic textures, etc.). Many experiments studying basic visual abilities use short display times, typically only long enough for a single fixation. This includes studies of texture segmentation (Julesz, 1981; Rosenholtz & Wagemans, 2014; Treisman, 1985), popout search (Treisman & Sato, 1990), crowding (Levi, 2008), ensemble/set perception (Whitney, Haberman, & Sweeny, 2014), numerosity judgments (Feigenson, Dehaene, & Spelke, 2004), dual-task performance (VanRullen, Reddy, & Koch, 2004), iconic memory (Sperling, 1960), and perceptual organization in general (Wagemans, 2015). Experimenters use short display times not only to explicitly study at-a-glance perception; but also to study preattentive processing, or avoid complicating factors, such as fixation location. Human vision research, however, rarely extends this work to information visualizations, computer displays, and user interfaces; all of which have scene-like qualities and are practically relevant, despite being artificially designed. The goal of our research is to bridge this gap between natural and artificial stimuli by studying at a glance perception of web pages.
Research on human vision arguably suggests that perception of artificial stimuli is poorer than that of natural scenes. Synthetic stimuli and tasks appear to be more affected by attentional limitations than natural stimuli and tasks (Li et al., 2002). Researchers have suggested several explanations for this apparent difference. Our visual systems developed to process natural stimuli (Geisler, 2008). There appear to be brain areas devoted to processing stimuli like natural scenes (Epstein & Kanwisher, 1998); this specificity of neural organization possibly provides an advantage in processing those natural stimuli (VanRullen et al., 2004). In addition, web pages are often quite text-heavy; much of this text is unlikely to be readable at a glance, perhaps further impairing ability to classify a web page at a glance. One obviously cannot generalize from extracting the gist of a natural scene or of psychophysics-style stimuli to the gist of diversely designed artifacts such as web pages. Given their novelty and pervasiveness, web pages are “real-world” stimuli that require rigorous psychophysical investigation.
Beyond contributing to theories of human vision, understanding web page gist is relevant for design and usability. We can learn from web pages that are easy to comprehend at a glance in order to improve easy access to relevant information. For this reason, researchers in the HCI (Human-Computer Interaction) community have begun to study perception of web pages at a glance. However, to our knowledge all of these studies involved subjective judgments, e.g., “is this web page aesthetically pleasing,” or “does this web page appear to have high or low usability?” Researchers have found that participants form subjective impressions of the appeal of a web page in the first 50 ms of viewing, and respond consistently when shown the same stimulus later (Lindgaard, Fernandes, Dudek, & Brown, 2006). Furthermore, first impressions of visual appeal based on short (50 ms) exposures correlate well with judgments based upon longer viewing times (500 ms and further up to 10 s) (Tractinsky, Cokhavi, Kirschenbaum, & Sharfi, 2006). Users also make consistent subjective ratings about the trustworthiness and perceived usability of web pages after only 50 ms of viewing (Lindgaard, Dudek, Sen, Sumegi, & Noonan, 2011). Inspired by human vision research (see Oliva & Torralba, 2006) that suggests that low spatial frequencies are sufficient to communicate the layout of a natural scene, Thielsch & Hirschfeld (2010, 2012) found high correlation between judgments of aesthetics made on low-pass filtered web page screenshots and the original web page screenshots. Perceived usability, on the other hand, correlated better with judgments made based on high-pass filtered stimuli. Of course, just because observers can consistently make certain subjective judgments at a glance does not imply that they will be able to perform the tasks of interest in this paper. Instead of studying subjective judgments, we ask observers to perform objective tasks with web pages at a glance.
We perform several experiments to investigate what can be seen in a single fixation, 120 ms, on a web page. Display times of this magnitude are typical for similar studies with natural scenes (Fei-Fei et al., 2007). In Experiment 1, we ask whether observers can rapidly ascertain the category of a web page. This is a new question in the human vision literature. Common wisdom in HCI suggests that a user cannot acquire much semantic information, such as the category of a web page or meaning of any text, in a presentation time of less than 500 ms (e.g., Lindgaard et al., 2006). However, researchers have not actually tested this hypothesis.
In Experiment 2, we ask whether ads are detectable at a glance. This is an object detection task like the animal/no-animal task in scene perception studies (Kirchner & Thorpe, 2006; Li et al., 2002; Thorpe et al., 1996). However, since detection depends greatly on both the signal to be detected and on the background against which it appears, one cannot infer from easy animal detection that ads will be easy to detect. In particular, designers may use multiple different strategies for ad design. Some designs aim for ads to have a salient, visually distinct appearance from the rest of the web page, while other designs might disguise the ad on purpose, effectively creating camouflaged objects. In previous work, Pieters & Wedel, (2012) showed that observers can distinguish between ads and editorial articles in magazines with high accuracy (up to 85% on average) in only 100 ms. Furthermore, observers could discriminate between types of ads (i.e., for cars, financial services, food, or skincare products), at rates of 95% correct for “typical” ads, and 53% correct for “atypical” ads. This study differs from the present in several ways. We study ads embedded in web pages, as opposed to isolated full-page magazine ads. The task is to detect these embedded ads, rather than to categorize them. In addition, the ad style in magazines tends to be quite different from that in web pages.
Finally, in Experiment 3, we ask how well a user can locate the menu bar. A menu bar is essentially defined by the horizontal or vertical alignment of its elements; menu items form either a row or a column, respectively (Fig. 10). In addition, many menu bars contain menu items that have similar colors and other features, and/or those items may be contained within a rectangular box. As a result, one can think of menu localization as an initial question of what perceptual organization (alignment, similarity, and/or containment) one can perceive at a glance at a web page. Considerable work has demonstrated that observers can perform perceptual organization tasks in brief presentations (van der Helm, 2014). However, much of this work uses fairly simple and homogeneous displays, leaving open the question of what observers can perceive in web pages at a glance. Perhaps more relevant is work suggesting that observers can estimate the 3D layout of a natural scene at a glance (Greene & Oliva, 2009a, 2009b), though clearly both the task and stimuli differ greatly from detecting a web page menu.
Our particular set of tasks can be thought of as a parallel to tasks in the scene gist literature. We have a semantic task (categorization), similar to scene categorization (Biederman, 1981); an object detection task (ad detection), similar to object detection with scenes (Thorpe et al., 1996), and a layout-related task (menu localization), similar to 3D layout estimation (Oliva & Torralba, 2001). Can observers assess these mid-to-high level properties in web pages, as they can in natural scenes?
Observers viewed stimuli on a Dell E2209W LCD monitor (47.5 cm by 30 cm viewable area, 1680 × 1050 resolution) with their eyes approximately 55 cm from the center of the screen. Stimulus presentation and response collection was done using PsychToolbox-3 (Kleiner et al., 2007) in MATLAB.
In all experiments, observers responded by using the mouse to click on-screen buttons, and received no feedback. The order of presentation was randomized for each observer. Stimuli were screen shots of web pages, or modified versions thereof, as described below. Stimuli were displayed at their original 1200 × 800 resolution, and subtended an area of the screen approximately 33.75 cm × 22.5 cm (34 × 23 degrees visual angle). In other words, all web pages were displayed at the same size and resolution used to capture the screenshot, and viewed at a typical viewing distance for browsing the web. Nearly all of the text in the original screenshots was legible to any observer with normal or corrected-to-normal vision at the experimental viewing distance, with the only exceptions being (rarely) poorly rendered text, or essentially “footnote” or other unimportant text – e.g., disclaimers or trademark symbols.
As web pages are designed artifacts, the appearance of a given category is at least in part a cultural convention (Reinecke, Arbor, & Gajos, 2014). As a result, performance will likely depend upon a user’s experience with web browsing, and with whether the user comes from the same culture as the designs studied (e.g., Indian news sites might look different from American ones). While cultural peculiarities are certainly an interesting avenue of study, they are not immediately relevant to our research goal. To minimize this potential added source of variance, all observers were university students (undergraduate and graduate), with English as their first language. In total (for all experiments), we recruited 25 participants (average age = 23.56 years, standard deviation = 4.00 years, range of 18 to 35 years, 12 female). The participants, on average, spent 3 h per day surfing the Internet. Each experiment took each participant between 30 min and an hour to complete, and participants received $15 USD compensation for their time. All participants provided written informed consent prior to the experiment.
In order to rigorously study at-a-glance web page categorization, we need properly labeled web pages. One way to get labels is to do a categorization experiment with unlimited viewing time. In other words, we need to collect a corpus of web pages, define a set of unambiguous category labels, and confirm that observers agree on the ground truth labeling of those web pages’ categories.
Picking the web page categories
Selected web page categories and their descriptions. This is identical to the list that was given to subjects, and was designed to be intuitive. Each subject could consult this list at any time during the experiment
# of screenshots
1. Art place
if you want to see art, e.g., exhibition, museum, galleries
has articles with titles and dates, it usually has opinion of a person with a form that you could write your comments in
if you want to learn about the services that it provides, e.g,. consultation
4. Computer game
if you want to play a computer game either online or offline, or read the latest news about a computer game, buy or download a computer game
if you have an emergency, e.g., need advice for your kid
if you want to read daily news
7. Online tutorials
e.g., learning a course, finding a tutorial for learning html coding
if you want to buy an online product item, new or used
unions, groups of people with same interests, e.g., if you are looking for a cultural club, or book reading club
if you are looking for things to do in a destination, booking for a tour
Collecting web page screenshots
In order to gather a candidate set of web page stimuli, we first collected web pages belonging to the 10 categories. One way to do this would have been to search for keywords (e.g., “news,” “art museum”) in an online search engine. However, we did not want to bias our set of web pages to those that come up first on a web search. In particular, we wished to avoid selecting high-traffic web pages, as these might lead to anomalous responses due to familiarity with the logos, color scheme, or layout of those particular web sites. Thus, rather than using search engines to find category exemplars, we crawled the site DMOZ.org for 1795 random URLs from our selected categories. DMOZ is a widely used online repository of URLs, organized by volunteer editors into categories and sub-categories. The unbiased collection of URLs is reflected by its use in professional web page traffic-ranking services like Alexa (alexa.com) and as a database for training URL-based topic classification algorithms (Baykan, Henzinger, Marian, & Weber, 2011). We automatically captured web page screenshots in a 1200 × 800 pixel browser window using the webkit2png software package (https://github.com/adamn/python-webkit2png) in the Safari browser on an Apple desktop. We only kept the first “page,” i.e., the portion of the web page that fit in the browser; thus, some screenshots did not span the full vertical extent of the web page. Given that the first glance of a web page occurs before users can scroll further down the web page, and that these browser dimensions are typical for current display resolutions, our screenshots provide a good representation of the first 120 ms of the typical web browsing experience. By hand, we culled stimuli that appeared, upon examination, to be ambiguous in category, written in a language other than English, or not fully loaded at the time of screenshot. This narrowed our initial set of screenshots to 714.
Getting ground truth categories for screenshots
In order to more objectively assess whether observers would agree with the labeled category for each screenshot, given unlimited viewing time, we asked 6 participants in a pilot experiment to categorize each screenshot into one of the 10 categories. Of the initial 714 screenshots, we randomly selected a subset of 379, such that the experiment took approximately one hour. Each participant categorized all 379 screenshots. Presentation order was randomized across participants.
Half of the participants were male, and all were native English-speaking university students. We discarded the data from one observer because of apparent difficulty understanding the instructions. A row of response buttons, one for each category, appeared below each screenshots. Although participants were allowed unlimited viewing time, they were encouraged not to spend more than several seconds on any particular web page. This was to avoid deciding upon the category based on small or technical details in the text; we wanted to avoid having participants overthink the categorizations, such as deciding that a blog is a tourism web page because many of the posts involve travel.
Prior to the experiment, we gave the observers a verbal overview of each category. The participants were also instructed to consult with a list of all 10 categories and their short descriptions (see Table 1), while making their decisions. This list was displayed on a separate screen. The same was true for the at-a-glance categorization experiments described below.
For 283 web pages (about 75%), all 5 observers agreed with our categorization, and for 334 (about 88%) at least 4 of 5 agreed. In further experiments, we use only web pages for which at least 4 out of these 5 observers agreed with our initial categorization. This set of web pages has the following number of exemplars per category: art place, 30; blog, 32; company, 33; computer game, 33; helpline, 32; news, 34; online tutorial, 28; shopping, 34; society, 37; and tourism, 41. For some examples of these stimuli, see Fig. 4.
We recruited 10 participants for this task (5 female).
Each trial consisted of three steps. First, participants were asked to fixate on a cross that appeared on the screen for 900 ms. We asked participants to fixate the center of the page, consistent with many previous scene perception experiments. For web pages, previous work has examined where people look while surfing the Internet, and found that for information foraging tasks (e.g., “which car has the best performance, a Porsche, BMW, or Audi?,” subjects mainly look at the center of the web pages in the first second of viewing (Buscher, Cutrell, & Morris, 2009). Future work could examine web page perception when fixating at typical fixations during a particular task.
Participants first did a short training session of 30 trials to orient them to the experiment. The screenshots used in the training came from a separate set than used in the main experiment.
Mean (in percent correct, top value in each cell) and standard deviation (in pp., bottom value in each cell) of performance in experiments 1A and 1B, computed across subjects (N = 10 per experiment) per category. The mean values are identical to the diagonals of the confusion matrices in Figs. 3 and 6
Intact text (Exp. 1A)
M = 35
SD = 10
Scrambled text (Exp. 1B)
What cues in the stimuli might support such high performance? Interestingly, several participants indicated to us that they could read words within the screenshots, and that they used this information in doing the categorization task. We explored this possibility in Experiment 1B.
Was some of the web page text at least partially readable? Did reading text help observers determine the category? A common assumption in HCI design is that text is not readable at a glance (Lindgaard et al., 2006). Research in human vision, however, suggests that participants could plausibly have read some of the text. Humans can read at least 12 words per second (83.3 ms per word) when presented foveally, one after another (Potter, 1984). Humans can also read words presented peripherally for short times, provided that both the font and the spacing between letters is large enough (Latham & Whitaker, 1996). If the spacing is insufficient, a phenomenon known as crowding strongly limits peripheral reading (Pelli & Tillman, 2008), and limits it far more than a lack of acuity (Rosenholtz, 2016). In addition, the participants need not literally read the text for it to be useful; they need only infer that words look more like words that would appear on, say, an art site than on a helpline page.
In order to test the hypothesis that observers are using text to classify web pages at a glance, we changed the text on a set of web pages to be unreadable. We then measured to what degree categorization performance degraded. If observers do not read any text at a glance, performance should be unaffected. On the other hand, if participants can read some of the text, there should be a significant effect of performance.
Scrambling web page text
Previous work has used a number of techniques to render text unreadable, e.g., on a web page, including “Greeking” text (Tullis, 1998) or converting it into a language that our participant population is unable to read, such as Finnish. We chose instead to use image operations to flip the text about a horizontal axis, i.e., flip (using minimal bounding boxes around words). This enabled us to change all text, including text embedded in images, while minimally affecting font and layout of the page.
We recruited 10 new participants (5 female) for this task.
The procedure was identical to that in Experiment 1A, except that all of the text in each screenshot was scrambled (except for the initial training, which included 15 trials with scrambled screenshots as well as 15 trials with unmodified screenshots).
Importantly, we found that scrambling the text significantly decreased performance by 4.34% (p < 4E-4, collapsing over all subjects and categories); the categorization task is more difficult when the text is unreadable. This implies that participants in Experiment 1A were in fact inferring the category in part from readable text. Since performance in the scrambled condition was well above chance, however, readable text was clearly not the only cue. Further examining the individual web page categories, we found that performance was significantly better with intact than scrambled text for three categories: company (15.5 pp), news (22.14 pp), and society (11.62 pp) (p < 5E-3 per category, see Table 2 for standard deviation). Thus, readable text provided a better cue for some categories than for others; we examine this finding in more detail in the discussion.
It is possible that some visual processing occurs after the stimulus is removed, due to afterimages or iconic memory, despite the appearance of the response screen. This would effectively increase the stimulus display time, beyond the length of a typical fixation. To control for this possibility, we ran an experiment identical to Experiment 1A, except that immediately after each web page image, a noise mask appeared. The mask was made by taking a random web page from the unused subset of the original 714, transforming the image into the Fourier domain, randomizing the phase, and using the inverse Fourier transform to convert back to the image domain. The mask thus had the same spatial frequency content and color distribution as a real web page, but random phase.
We recruited 7 new participants (4 female) for this experiment.
The procedure was identical to Experiment 1A, except that a mask appeared on the screen for 120 ms after the stimulus in the same position as the web page was on the screen.
All 7 observers performed above chance in all categories (p < 1E-5). Furthermore, there was no significant decrease in overall performance between the performance in this experiment than Experiment 1A (overall performance was 2.05 pp. lower in the masked experiment, p < 0.152), and there were no significant differences in performance for any categories. These results indicate that performance was not significantly affected by the mask, which presumably halts visual processing after 120 ms.
In addition to perceiving the category of a real world scene, observers can also extract sufficient information at a glance to recognize a few objects (Fei-Fei et al., 2007; Oliva, 2005). Can participants also recognize elements of a web page at a glance? Advertising is one interesting and ubiquitous type of web content. Web pages often display ads to generate revenue, and advertisers in many cases want users to click on these ads. As a result, designers of both web pages and ads take care in setting the location and style of an ad to maximize the chances that a visitor will click on the ad. Users, on the other hand, may desire not to click on irrelevant ads, or perhaps learn to ignore ads completely. This might lead advertisers to use various techniques to “trick” users into clicking on the ad, for example when advertisers try to combat “banner blindness” (Benway, 1997).
In Experiment 2, we ask whether participants could detect the presence of an ad embedded in a web page at a glance.
The same 10 participants in Experiment 1B completed this experiment in a separate session.
From our original set of 714 screenshots, we excluded 334 of the stimuli from the previous experiments. We then manually selected the first 50 with at least one visible ad, and the first 50 without. To qualify as having an ad, the screenshot had to display the entire ad, fully loaded, and the ad could not link to the same web site as the current page. Thus, we excluded ads for a different product sold on the same shopping site, e.g., a bank web page containing an ad about its own banking service. Overall, the experimental design was similar to Experiment 1A, except that the response screen had only two buttons, (one green for “ad,” one red for “no ad”), underneath the text, “Did you see any ad(s)?”
Participants did this task randomly interleaved with the one in Experiment 3 (finding navigational menus). After displaying each web page screenshot, the screen displayed (with equal probability) one of two questions: “Was an ad was present?” or “Where was the menu?” The participant answered only a single question per trial. The participant did not know which question would be asked until after the screenshot was removed from the screen. We made the task uncertain for two main reasons. First, this uncertainty better mimics standard web page viewing conditions. A typical user does not first approach a web page with the sole intention of finding any ads or isolating the menu; rather he has a higher level goal, which might at some point require locating a particular graphical element. While having two possible tasks is not identical to natural conditions, it is a step in that direction. Furthermore, we wanted to avoid participants deploying covert spatial attention or making eye movements to, say, a probable menu location, which would be more likely if the task were known. To put it more directly: an observer could simply look at the top of the web page and easily detect the presence of the menu, responding “top” if present and “left” if not present. Using our method, we can better test whether ads and menus are normally available at a glance when fixating the center of a web page, instead of testing how well observers can deploy overt or covert attentional mechanisms.
We have demonstrated that in a single 120 ms fixation, participants can quickly categorize a web page into one of the ten common categories. Furthermore, it seems that participants are, at least in part, using text in the web pages to do the task. Future work should probe the role of text further, and ask whether participants mainly use text near the point of fixation, or whether they can read (or at least infer category from) bigger text that appears more peripherally. Clearly, however, the content of text is not the whole story; the bulk of the performance is driven by other cues, which most likely include text quantity and font, page layout, organization, and presence and content of images, among other factors. Our current work does not delineate the relative contribution of these cues, and thus a formal investigation of them is an important direction of future work. Furthermore, participants can discriminate pages that contain ads from those that do not, as well as localize elements of the layout, namely the menu bar. Ad detection is more difficult than menu localization, possibly resulting from designers’ different purposes for the two elements. Further research is needed to pin down which menu or ad styles are more visible than others, and why.
These results speak to the capabilities of the human visual system. Both recognizing scenes and recognizing web pages presumably result from similar processes that bring together extraction of a general-purpose visual representation with higher-level inference informed by previous experience. The general-purpose representation likely developed for ecologically important tasks like scene perception, but also likely supports understanding of web pages, in part because design develops to make use of existing visual processing architecture in the human visual system.
While much work has been done to study the perception of natural scenes at a glance, our study shows the importance of extending it to design research. Beyond providing the first quantitative study of rapid web page perception, we challenge previous assumptions about what can and cannot be perceived in a glance at a web page. Furthermore, researching web page perception is timely; as we become more enmeshed in the virtual world, more of our “natural vision” will be filled with artificial displays like web pages. Our expectation is that research on at-a-glance perception of digital displays and designs will become more prevalent and applicable as digital interfaces become more pervasive in daily life.
Our results also touch on practical issues. If a user clicks on a link in a web page, and the first glance or two does not suggest the correct category, she may quickly leave the page to find another one. For a well-designed page, viewing the URL at the top of the page would be far less efficient for classifying the page than getting the gist at a glance. Designers could improve existing applications by, for example, displaying text to be easily readable at a glance and suggestive of the web page’s category, or by including an easily comprehensible image that better cues the category. In working to improve their design, they could also use our technique of rapid presentation coupled with an objective categorization task to test the placement of their menus, ads, and graphical elements. While we do not test usability directly, we argue that having a quickly comprehendible web page would promote overall ease of use. This logic potentially extends to mobile devices: one can consider the first viewport of a mobile page (before scrolling) to be analogous to the first glance of a web page.
One could ask many other interesting questions about web page perception at-a-glance. What other elements can users quickly identify? How much layout information do users get at a glance? What design elements underlie the ability to accurately categorize a page, identify an ad, or find a menu? Are there computational models that can predict the results from objective tasks like ours? How does web page perception depend on the viewer’s age, experience, or visual impairment? Our hope is, beyond presenting our specific experimental results, to understand perception of designed, real-world stimuli using the rich experimental paradigms of vision science.
We would like to thank Dr. Krista Ehinger and Dr. Benjamin Wolfe for their input to this work. We also thank Kevin Kwok (firstname.lastname@example.org) and Guillermo Webster (email@example.com) for providing us with text-flipping software.
This study was partly funded by a Google Award granted to Dr. Ruth Rosenholtz.
Availability of data and materials
As stated in the manuscript, the data and code are available at: https://github.com/ali-design/WebpageGist
All of the authors made substantial contributions to conception and design, data analysis and interpretation, drafting and revising it critically for important intellectual content, and upon final approval of the version will take public responsibility for appropriate portions of the content; and agreed to be accountable for all aspects of the work in ensuring that questions related to the accuracy or integrity of any part of the work are appropriately investigated and resolved. All authors read and approved the final manuscript.
Ethics approval and consent to participate
This study’s approval and consent was under MIT-IRB approval.
Consent for publication
The authors declare that they have no competing interests.
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Open AccessThis article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made.
- Baykan, E., Henzinger, M., Marian, L., & Weber, I. (2011). A Comprehensive Study of Features and Algorithms for URL-Based Topic Classification. ACM Transactions on the Web, 5(3), 15:1--15:29. https://doi.org/10.1145/1993053.1993057
- Benway, J. P. (1997). Banner blindness: The irony of attention grabbing on the World Wide Web. Proceedings of the Human Factors and Ergonomics Society Annual Meeting, 42(5), 463–467 https://doi.org/10.1177/154193129804200504.View ArticleGoogle Scholar
- Biederman, I. (1981). On the semantics of a glance at a scene in. In M. Kubovy, & J. Pomerantz (Eds.), Perceptual Organization. Hillsdale, NJ: Erlbaum.Google Scholar
- Buscher, G., Cutrell, E., & Morris, M. R. (2009). What do you see when you’re surfing? Using eye tracking to predict salient regions of web pages. Proceedings of the 27th SIGCHI Conference on Human Factors in Computing Systems, 21–30 https://doi.org/10.1145/1518701.1518705.
- Crouzet, S. M., Kirchner, H., & Thorpe, S. J. (2010). Fast saccades toward faces: face detection in just 100 ms. Journal of Vision, 10(4), 16.1–17. https://doi.org/10.1167/10.4.16
- Ehinger, K. A., & Rosenholtz, R. (2016). A general account of peripheral encoding also predicts scene perception performance. Journal of Vision, 16(2016), 1–19 https://doi.org/10.1167/16.2.13.doi.Google Scholar
- Epshtein, B., Eyal, O., Yonatan, W., Ofek, E., & Wexler, Y. (2010). Detecting Text in Natural Scenes with Stroke Width Transform. Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, (d), 2963–2970 https://doi.org/10.1109/CVPR.2010.5540041.
- Epstein, R., & Kanwisher, N. (1998). A cortical representation of the local visual environment. Nature, 392(6676), 598–601 https://doi.org/10.1038/33402.View ArticlePubMedGoogle Scholar
- Fei-Fei, L., Iyer, A., Koch, C., & Perona, P. (2007). What do we perceive in a glance of a real-world scene? Journal of Vision, 7(1), 10 https://doi.org/10.1167/7.1.10.View ArticlePubMedGoogle Scholar
- Feigenson, L., Dehaene, S., & Spelke, E. (2004). Core systems of number. Trends in Cognitive Sciences, 8(7), 307–314 https://doi.org/10.1016/j.tics.2004.05.002.View ArticlePubMedGoogle Scholar
- Geisler, W. S. W. (2008). Visual Perception and the Statistical Properties of Natural Scenes. Annual Review of Psychology, 59(1), 167–192 https://doi.org/10.1146/annurev.psych.58.110405.085632.View ArticlePubMedGoogle Scholar
- Greene, M. R., & Oliva, A. (2009a). Recognition of natural scenes from global properties: seeing the forest without representing the trees. Cognitive Psychology, 58(2), 137–176 https://doi.org/10.1016/j.cogpsych.2008.06.001.View ArticlePubMedGoogle Scholar
- Greene, M. R., & Oliva, A. (2009b). The Briefest of Glances: The Time Course of Natural Scene Understanding. Psychological Science, 20(4), 464–472 https://doi.org/10.1111/j.1467-9280.2009.02316.x.View ArticlePubMedPubMed CentralGoogle Scholar
- Harris, C. M., Hainline, L., Abramov, I., Lemerise, E., & Camenzuli, C. (1988). The distribution of fixation durations in infants and naive adults. Vision Research, 28(3), 419–432 https://doi.org/10.1016/0042-6989(88)90184-8.View ArticlePubMedGoogle Scholar
- Joubert, O. R., Rousselet, G. A., Fize, D., & Fabre-Thorpe, M. (2007). Processing scene context: fast categorization and object interference. Vision Research, 47(26), 3286–3297 https://doi.org/10.1016/j.visres.2007.09.013.View ArticlePubMedGoogle Scholar
- Julesz, B. (1981). Textons, the elements of texture perception, and their interactions. Nature, 290(5802), 91–97 https://doi.org/10.1038/290091a0.View ArticlePubMedGoogle Scholar
- Kirchner, H., & Thorpe, S. J. (2006). Ultra-rapid object detection with saccadic eye movements: Visual processing speed revisited. Vision Research, 46(11), 1762–1776 https://doi.org/10.1016/j.visres.2005.10.002.View ArticlePubMedGoogle Scholar
- Kleiner, M., Brainard, D. H., Pelli, D. G., Broussard, C., Wolf, T., & Niehorster, D. (2007). What’s new in Psychtoolbox-3? Perception, 36, S14 https://doi.org/10.1068/v070821.Google Scholar
- Latham, K., & Whitaker, D. (1996). A comparison of word recognition and reading performance in foveal and peripheral vision. Vision Research, 36(17), 2665–2674 https://doi.org/10.1016/0042-6989(96)00022-3.View ArticlePubMedGoogle Scholar
- Levi, D. (2008). Crowding-An essential bottleneck for object recognition: A mini-review. Vision Research, 48(5), 635–654 Retrieved from http://www.ncbi.nlm.nih.gov/pmc/articles/PMC2268888/.View ArticlePubMedPubMed CentralGoogle Scholar
- Li, F. F., VanRullen, R., Koch, C., & Perona, P. (2002). Rapid natural scene categorization in the near absence of attention. Proceedings of the National Academy of Sciences of the United States of America, 99(14), 9596–9601 https://doi.org/10.1073/pnas.092277599.View ArticlePubMedPubMed CentralGoogle Scholar
- Lindgaard, G., Dudek, C., Sen, D., Sumegi, L., & Noonan, P. (2011). An exploration of relations between visual appeal, trustworthiness and perceived usability of homepages. ACM Trans. Comput.-Hum. Interact., 18(1), 1–30 https://doi.org/10.1145/1959022.1959023.View ArticleGoogle Scholar
- Lindgaard, G., Fernandes, G., Dudek, C., & Brown, J. (2006). Attention web designers: You have 50 milliseconds to make a good first impression! Behaviour & Information Technology, 25(2), 115–126 https://doi.org/10.1080/01449290500330448.View ArticleGoogle Scholar
- Oliva, A. (2005). Gist of the scene. In Neurobiology of Attention, (pp. 251–256) https://doi.org/10.1016/B978-012375731-9/50045-8.View ArticleGoogle Scholar
- Oliva, A., & Torralba, A. (2001). Modeling the shape of the scene: A holistic representation of the spatial envelope. International Journal of Computer Vision, 42(3), 145–175.View ArticleGoogle Scholar
- Oliva, A., & Torralba, A. (2006). Building the gist of a scene: The role of global image features in recognition. Progress in Brain Research, 155(B), 23–36 https://doi.org/10.1016/S0079-6123(06)55002-2.View ArticlePubMedGoogle Scholar
- Palmer, J. W. (2002). Web Site Usability, Design, and Performance Metrics. Information Systems Research, 13(2), 151–167 https://doi.org/10.1287/isre.184.108.40.206.View ArticleGoogle Scholar
- Pelli, D. G., & Tillman, K. a. (2008). The uncrowded window of object recognition. Nature Neuroscience, 11(10), 1129–1135 https://doi.org/10.1038/nn.2187.View ArticlePubMedPubMed CentralGoogle Scholar
- Pieters, R., & Wedel, M. (2012). Ad Gist: Ad Communication in a Single Eye Fixation. Marketing Science, 31(February 2015), 59–73 https://doi.org/10.1287/mksc.1110.0673.View ArticleGoogle Scholar
- Pirolli, P., & Card, S. (1999). Information foraging. Psychological Review, 106(4), 643–675 https://doi.org/10.1037/0033-295X.106.4.643.View ArticleGoogle Scholar
- Potter, M. C. (1984). Rapid Serial Visual Presentation (RSVP): A Method for Studying Language Processing. In New Methods in Reading Comprehension Research, (pp. 91–118).Google Scholar
- Rayner, K., & Castelhano, M. S. (2007). In M. Wedel, & R. Pieters (Eds.), Eye movements during reading, scene perception, visual search, and while looking at print advertisements, (vol. 2175) https://doi.org/10.4324/9780203809617.Google Scholar
- Reinecke, K., Arbor, A., & Gajos, K. Z. (2014). Quantifying visual preferences around the world. Proceedings of the 32nd Annual ACM Conference on Human Factors in Computing Systems - CHI ‘14, 11–20 https://doi.org/10.1145/2556288.2557052.
- Rosenholtz, R. (2016). Capabilities and Limitations of Peripheral Vision. Annual Review of Vision Science, 2(1), 437–457 https://doi.org/10.1146/annurev-vision-082114-035733.View ArticlePubMedGoogle Scholar
- Rosenholtz, R., & Wagemans, J. (2014). Texture Perception (pp. 1–24). Oxford University Press https://doi.org/10.1093/oxfordhb/9780199686858.013.058.
- Rousselet, G. A., Joubert, O. R., & Fabre-Thorpe, M. (2005). How long to get to the “gist” of real-world natural scenes? Visual Cogn, 12(6), 852–877 https://doi.org/10.1080/13506280444000553.View ArticleGoogle Scholar
- Shrestha, S., Lenz, K., Chaparro, B. S., & Owens, J. W. (2007). “F” Pattern Scanning of Text and Images in Web Pages. Human Factors and Ergonomics Society, 55(18), 1200–1204 https://doi.org/10.1177/154193120705101831.View ArticleGoogle Scholar
- Sperling, G. (1960). The information available in brief visual presentations. Psychological Monographs: General and Applied, 74(11), 1–29 https://doi.org/10.1037/h0093759.View ArticleGoogle Scholar
- Still, A. W., & White, A. P. (1981). The approximate randomization test as an alternative to the F test in analysis of variance. British Journal of Mathematical and Statistical Psychology, 34(2), 243–252 https://doi.org/10.1111/j.2044-8317.1981.tb00634.x.View ArticleGoogle Scholar
- Thielsch, M. T., & Hirschfeld, G. (2010). High and low spatial frequencies in website evaluations. Ergonomics, 53(8), 972–978 https://doi.org/10.1080/00140139.2010.489970.View ArticlePubMedGoogle Scholar
- Thielsch, M. T., & Hirschfeld, G. (2012). Spatial frequencies in aesthetic website evaluations – explaining how ultra-rapid evaluations are formed. Ergonomics, 55(7), 731–742 https://doi.org/10.1080/00140139.2012.665496.View ArticlePubMedGoogle Scholar
- Thorpe, S. J., Fize, D., & Marlot, C. (1996). Speed of processing in the human visual system. Nature, 381(6582), 520–522 https://doi.org/10.1038/381520a0.View ArticlePubMedGoogle Scholar
- Thorpe, S. J., Gegenfurtner, K. R., Fabre-Thorpe, M., & Bülthoff, H. H. (2001). Detection of animals in natural images using far peripheral vision. European Journal of Neuroscience, 14(5), 869–876 https://doi.org/10.1046/j.0953-816x.2001.01717.x.View ArticlePubMedGoogle Scholar
- Tractinsky, N., Cokhavi, A., Kirschenbaum, M., & Sharfi, T. (2006). Evaluating the consistency of immediate aesthetic perceptions of web pages. International Journal of Human Computer Studies, 64(11), 1071–1083 https://doi.org/10.1016/j.ijhcs.2006.06.009.View ArticleGoogle Scholar
- Treisman, A. (1985). Preattentive processing in vision. Computer Vision, Graphics and Image Processing, 31(2), 156–177 https://doi.org/10.1016/S0734-189X(85)80004-9.View ArticleGoogle Scholar
- Treisman, A., & Sato, S. (1990). Conjunction Search Revisited. Journal of Experimental Psychology: Human Perception and Performance, 75(3), 459–478 https://doi.org/10.1037/0096-15220.127.116.119.Google Scholar
- Tullis, T. S. (1998). A method for evaluating Web page design concepts. In CHI 98 conference summary on Human factors in computing systems - CHI ‘98, (pp. 323–324). New York, New York, USA: ACM Press https://doi.org/10.1145/286498.286793.View ArticleGoogle Scholar
- van der Helm, P. A. (2014). In J. Wagemans (Ed.), Simplicity in Perceptual Organization. Oxford University Press https://doi.org/10.1093/oxfordhb/9780199686858.013.052.
- VanRullen, R., Reddy, L., & Koch, C. (2004). Visual search and dual tasks reveal two distinct attentional resources. Journal of Cognitive Neuroscience, 16(1), 4–14 https://doi.org/10.1162/089892904322755502.View ArticlePubMedGoogle Scholar
- VanRullen, R., & Thorpe, S. J. (2001). The Time Course of Visual Processing: From Early Perception to Decision-Making. Journal of Cognitive Neuroscience, 13(4), 454–461 https://doi.org/10.1162/08989290152001880.View ArticlePubMedGoogle Scholar
- Wagemans, J. (2015). Historical and Conceptual Background: Gestalt Theory. The Oxford Handbook of Perceptual Organization, 3–20 https://doi.org/10.1017/CBO9781107415324.004.
- Wedel, M., & Pieters, R. (2000). Eye Fixations on Advertisements and Memory for Brands: A Model and Findings. Marketing Science, 19(4), 297–312 https://doi.org/10.1287/mksc.19.4.297.11794.View ArticleGoogle Scholar
- Whitney, D., Haberman, J., Sweeny, T. D. (2014). From textures to crowds: multiple levels of summary statistical perception. The New Visual Neurosciences. MIT Press, 695–710.Google Scholar