[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
Next Article in Journal
A Novel Signal Separation Method Based on Improved Sparse Non-Negative Matrix Factorization
Next Article in Special Issue
Visual Analysis of Research Paper Collections Using Normalized Relative Compression
Previous Article in Journal
Compact Belief Rule Base Learning for Classification with Evidential Clustering
Previous Article in Special Issue
Dissecting Deep Learning Networks—Visualizing Mutual Information
You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Gaze Information Channel in Cognitive Comprehension of Poster Reading

1
College of Intelligence and Computing, Tianjin University, Yaguan Road 135, Tianjin 300350, China
2
Institute of Informatics and Applications, University of Girona, 17003 Girona, Spain
*
Author to whom correspondence should be addressed.
Entropy 2019, 21(5), 444; https://doi.org/10.3390/e21050444
Submission received: 18 February 2019 / Revised: 23 April 2019 / Accepted: 23 April 2019 / Published: 28 April 2019
(This article belongs to the Special Issue Information Theory Application in Visualization)
Figure 1
<p>The poster materials used in our eye tracking experiment. From left to right, in top row, posters 1, 2, and 3, in bottom row, posters 4, 5, 6, and 7. Marked in red, the Areas of Interests (AOIs) in which each poster is divided, and that are not displayed to the participants.</p> ">
Figure 2
<p>The information diagram represents the relationship between Shannon’s information measures.</p> ">
Figure 3
<p>The main elements of an information channel.</p> ">
Figure 4
<p>The scanpaths for participant 2 (<b>left</b>) and participant 5 (<b>right</b>) for posters 2, 4 and 7.</p> ">
Figure 5
<p>The heatmaps for participant 2 for posters 2, 4 and 7.</p> ">
Figure 6
<p>The heatmaps for participant 5 for posters 2, 4 and 7.</p> ">
Figure 7
<p>The AOI bin charts for participant 2 (<b>left</b>) and participant 5 (<b>right</b>) for posters 2, 4 and 7.</p> ">
Figure 8
<p>The channel measures for all participants and poster 7. From left to right and top to down, the equilibrium distribution <span class="html-italic">π</span> (<b>a</b>), <span class="html-italic">H<sub>s</sub></span>, <span class="html-italic">H<sub>t</sub></span>, <span class="html-italic">H</span>(<span class="html-italic">X</span>,<span class="html-italic">Y</span>), <span class="html-italic">I</span>(<span class="html-italic">X</span>;<span class="html-italic">Y</span>) (<b>b</b>), <math display="inline"><semantics> <mrow> <mi>H</mi> <mo>(</mo> <mi>Y</mi> <mo>|</mo> <mi>x</mi> <mo>)</mo> </mrow> </semantics></math> (<b>c</b>), and <span class="html-italic">I</span>(<span class="html-italic">x</span>;<span class="html-italic">Y</span>) (<b>d</b>). AOI1 is the title section, AOI2 is the intro section, AOI3 is the method section, AOI4 is the algorithm section, AOI5 is the results section, and AOI6 is the conclusions section.</p> ">
Figure 9
<p>The gaze channel for participant 5 and poster 7, between the AOIs with equilibrium distribution p(AOI) (<b>a</b>) and (<b>c</b>), and with conditional probabilities (<b>b</b>). The probability distributions values are found in <a href="#entropy-21-00444-t003" class="html-table">Table 3</a> and <a href="#entropy-21-00444-t004" class="html-table">Table 4</a>.</p> ">
Figure 10
<p>From left to right and top to down, the equilibrium distribution <span class="html-italic">π</span> (<b>a</b>), <span class="html-italic">H<sub>s</sub></span>, <span class="html-italic">H<sub>t</sub></span>, <span class="html-italic">H</span>(<span class="html-italic">X</span>,<span class="html-italic">Y</span>), <span class="html-italic">I</span>(<span class="html-italic">X</span>;<span class="html-italic">Y</span>) (<b>b</b>), <math display="inline"><semantics> <mrow> <mi>H</mi> <mo>(</mo> <mi>Y</mi> <mo>|</mo> <mi>x</mi> <mo>)</mo> </mrow> </semantics></math> (<b>c</b>), and <span class="html-italic">I</span>(<span class="html-italic">x</span>;<span class="html-italic">Y</span>) (<b>d</b>) in gaze information channel of participants 2 and 5 for poster 7. The numerical values are found in <a href="#entropy-21-00444-t004" class="html-table">Table 4</a>.</p> ">
Figure 11
<p>The stacked <span class="html-italic">H<sub>t</sub></span>, <span class="html-italic">H<sub>s</sub></span> (<b>a</b>), and <span class="html-italic">H<sub>t</sub></span>, <span class="html-italic">I</span>(<span class="html-italic">X</span>;<span class="html-italic">Y</span>) (<b>b</b>), for all participants.</p> ">
Figure 12
<p>The stacked <span class="html-italic">H<sub>t</sub></span>, <span class="html-italic">H<sub>s</sub></span> (<b>a</b>), and <span class="html-italic">H<sub>t</sub></span>, <span class="html-italic">I</span>(<span class="html-italic">X</span>;<span class="html-italic">Y</span>) (<b>b</b>), for all tested posters.</p> ">
Versions Notes

Abstract

:
Today, eye trackers are extensively used in studying human cognition. However, it is hard to analyze and interpret eye movement data from the cognitive comprehension perspective of poster reading. To find quantitative links between eye movements and cognitive comprehension, we tracked observers’ eye movement for reading scientific poster publications. We model in this paper eye tracking fixation sequences between content-dependent Areas of Interests (AOIs) as a Markov chain. Furthermore, we use the fact that a Markov chain is a special case of information or communication channel. Then, the gaze transition can be modeled as a discrete information channel, the gaze information channel. Next, some traditional eye tracking metrics, together with the gaze entropy and mutual information of the gaze information channel are calculated to quantify cognitive comprehension for every participant. The analysis of the results demonstrate that the gaze entropy and mutual information from individual gaze information channel are related to participants’ individual differences. This is the first study that eye tracking technology has been used to assess the cognitive comprehension of poster reading. The present work provides insights into human cognitive comprehension by using the novel gaze information channel methodology.

1. Introduction

As we all know, the eye is an important organ of the human being. It is often said that the eyes are the window of the soul, reflecting the thoughts of us human beings, and revealing the way in which the participants observe the scene. With more and more researchers using eye tracking technology as a research tool, eye tracking is a promising method in academic and industrial research. It has the potential to provide insights into a lots of issues in the visual and cognitive fields: education [1,2,3], medicine [4,5,6,7], assistive technology for people with a variety of debilitating conditions [8,9,10], better interface design [11,12,13], marketing and media [14,15,16]. Furthermore, as an important psychological experiment research method, eye movement provides a new perspective and way for educational technology research [17,18,19]. Actually, eye tracking has always been an important human–computer interaction method for making decisions [20,21,22].
Importantly, research based on the idea of using eye tracking as an instructional tool is still in its infancy. There is an urgent need to quantitatively compare eye movement metrics [23]. Several eye tracking metrics have been developed. The first are scanpaths, represented by an ordered sequence of fixations, for which vector and string-based editing methods have been developed to compute similarity [24,25,26,27]. The second are heatmaps, represented by Gaussian Mixture Models (GMMs) indicating the frequency (or probability) of fixation location [28,29]. A third type of sequential fixation pattern analysis is the transition matrix, which is rarely used as a quantitative measure [30]—see also the recent survey by Shiferaw et al. [31]. In this paper, we model eye tracking fixation sequences of Areas of Interests (AOIs) as a Markov chain. Furthermore, we model the gaze transition as a gaze information channel, introduced in [32]. We extend here the gaze information channel in [32] with a more complete description and interpretation of the metrics of the channel, and by showing how it is well adapted for clustering, allowing thus to analyze collective behavior. We also notice the importance of normalization of mutual information that was not considered in [32], when comparing different channels’ data.
While efforts are made to teach the elements of writing a scientific article in many graduate school curricula, much less attention is paid to teaching those skills necessary to read scientific posters, even though these arguably are the most common and most rapid ways to disseminate new findings. Especially for graduate students who are committed to scientific research, reading related research papers is an extremely important skill, and it is also a reflection of research ability. Actually, posters provide a coherent and efficient way to convey core ideas expressed in scientific papers, as described in [33]. Thus, how to quickly grasp the core idea of a scientific paper is also an essential ability for them. Furthermore, poster as a form of academic expression represents a concise and visual summary of one’s research. Its purpose is to be accessible and to drive attention to the research, and get the main point of the research across to as many people as possible through a concise and artistically attractive manner [34,35,36]. In other words, poster is one style of the most vivid and short scientific papers, which can best reflect a reader’s scientific reading skills and thinking process. Qiang et al. [33] used probabilistic graphical models to learn scientific poster design patterns, from existing posters, and they proposed an algorithm that considered both information conveyed and aesthetics to generate the poster layout. They used subjective evaluation of readability, informativeness and aesthetics to compare different designs of a poster. However, there is to our knowledge no study that has investigated by using eye tracking the cognitive comprehension of poster reading. Is it possible to use eye tracking data to quantify cognitive comprehension during reading poster from participants? This article reports on our efforts to answer this question.
Ten participants’ eye movement data of reading published posters were recorded individually using an eye tracker under controlled laboratory conditions. The tested posters are divided into content-dependent Area of Interests (AOIs), following the sections of a poster defined by the authors, as shown in Figure 1. The gaze information channel was used to analyze and interpret the eye tracking data. Some traditional eye tracking metrics, together with the gaze entropy and mutual information, are calculated to quantify cognitive comprehension of poster reading for every participant.

2. Background

Ponsoda et al. [37] introduced probability vectors and transition matrices by classifying the directions of saccade. Interestingly, their matrices were based on transition between the eight main saccade directions rather than between the Areas of Interests (AOIs), which are now more commonly used. Although Ponsoda et al. compared the matrices with a statistical method, they did not model the sequence of saccade directions as a Markov chain.
Ellis and Stark [38] compared the airline pilot transition matrices by dividing cockpit display traffic information (CDTI) into eight AOIs. They introduced first-order (fixation) transition matrices and converted them to conditional probability matrices. Then, conditional entropy was calculated using the conditional probability, or transition, matrices. Its value provided a measure of the statistical dependency in the spatial pattern of fixations represented by the transition matrix.
Liechty et al. [39] used Hidden Markov Models (HMMs) to distinguish between local and global visual attention states in eye movement data. Instead of applying the transition matrix Markov model as we do in this paper, they used HMMs to distinguish between fixations, similar to Velichkovsky et al. [40], who proposed classifying attention as ambient or focal.
Hwang et al. [41] did not construct a transition matrix between AOIs or in a grid, but considered position translation within a generated saliency map for a given scene, and introduced transitional semantic guidance calculations to evaluate gaze transition. This method can be seen as a hybrid between the transformation matrix construction and the scan path comparison where the transition matrices were replaced by semantic maps. Because the saliency maps generated by each scene may be different, one drawback of this approach is the comparison between scenes. This problem can be solved by building content-independent transition matrices.
Bonev et al. [42] built a Markov chain between the nodes of a regular grid matrix, the elements of the matrix being the normalized mutual information defined by the covariances of the Gaussian distribution of the attributes of the image. They obtained the equilibrium distribution of this Markov chain and defined the entropy of this equilibrium distribution as the complexity of the image. Thus, this complexity measure was defined independently of any observational task, only depending on grid and Gaussian distribution of attributes of the image. Then, Bonev et al. studied the correlation of this complexity with the sequences of long and short saccades. In our case, Markov chain transition probabilities matrix is defined from the observation trajectories, and thus it depends on the task.
Besag and Mondal [43] verified the feasibility of modeling gaze transition as a first-order Markov process. According to modeling eye movement transitions between areas of interest (AOIs) as a Markov chain, Krejtz et al. [44,45] calculated stationary entropy H s and transition entropy H t to measure the complexity of the Markov process. Raptis et al. [46] divided the images into three AOIs and used the gaze transition entropy proposed by Krejtz et al. [44] as a tool to quantify differences on visual search patterns among individuals within visual pattern recognition tasks of varying complexity.

3. Information Measures and Information Channel

In this section, we briefly introduce the most basic information measures of information theory and main elements of the information channel [47]. Since its inception by Shannon [48], information theoretic measures and concepts, which include as one of the main tools the information or communication channel, have been successfully used in many fields. For their application in visualization, image processing, and pattern recognition, see [49,50].

3.1. Basic Information-Theoretic Measures

Let X be a discrete random variable with alphabet X and probability distribution { p ( x ) } , where p ( x ) = P r { X = x } , x X . In this paper, { p ( x ) } will be denoted by p ( X ) .
The entropy H ( X ) of a discrete random variable X is defined by
H ( X ) = - x X p ( x ) log p ( x ) ,
where the summation is over the corresponding alphabet and the convention 0 log 0 = 0 is taken. In this paper, logarithms are taken in base 2 and, as a consequence, entropy is expressed in bits. The entropy H ( X ) gives the average uncertainty (or amount of information) of a random variable X.
The joint entropy H ( X , Y ) of a pair of discrete random variables X and Y with a joint probability distribution p ( X , Y ) = { p ( x , y ) } is defined by
H ( X , Y ) = - x X y Y p ( x , y ) log p ( x , y ) ,
where p ( x , y ) = P r [ X = x , Y = y ] is the joint probability of x and y.
The conditional entropy H ( Y | X ) of a random variable Y given a random variable X is defined by
H ( Y | X ) = x X p ( x ) H ( Y | X = x ) = x X p ( x ) ( - y Y p ( y | x ) log p ( y | x ) ) = - x X y Y p ( x , y ) log p ( y | x ) ,
where p ( y | x ) = P r [ Y = y | X = x ] is the conditional probability of y given x. H ( Y | X ) measures the average uncertainty associated with Y if we know the outcome of X.
The mutual information I ( X ; Y ) between two random variables X and Y is defined by
I ( X ; Y ) = H ( X ) + H ( Y ) - H ( X , Y ) = H ( X ) - H ( X | Y ) = H ( Y ) - H ( Y | X ) = x X y Y p ( x , y ) log p ( x , y ) p ( x ) p ( y ) .
Mutual information (MI) represents that knowledge of Y decreases the uncertainty of X, and vice versa. I ( X ; Y ) is a measure of the shared information or dependence between X and Y.
The relationship between Shannon’s information measures can be illustrated by a Venn diagram, as shown in Figure 2. The information diagram represents the relationship between Shannon’s information measures. Observe that I ( X ; Y ) and H ( X , Y ) are represented, respectively, by the intersection and the union of the information in X (represented by H ( X ) ) with the information in Y (represented by H ( Y ) ). H ( X | Y ) is represented by the difference between the information in X and the information in Y, and vice versa for H ( Y | X ) .

3.2. Information Channel

Communication or information channel is a system in which the output depends probabilistically on its input [47,51]. The conditional entropy and mutual information can be thought of in terms of a communication or information channel X Y whose output Y depends probabilistically on its input X. This information channel is characterized by a transition probability matrix which determines the conditional distribution of the output given the input [47,51]. Given that X and Y are two random variables, we can establish an information channel between X and Y. The diagram in Figure 3 shows the main elements of the information channel:
  • p ( X ) and p ( Y ) represent the probability distributions of input and output variables X and Y, respectively.
  • Probability transition matrix p ( Y | X ) composed of conditional probabilities p ( y | x ) , which gives the output distribution p ( Y ) given the input distribution p ( X ) . Each row of p ( Y | X ) can be seen as a probability distribution, denoted by p ( Y | x ) .
All of these elements are connected by the Bayes theorem relating marginal probabilities p ( X ) and p ( Y ) , conditional probabilities p ( y | x ) and p ( x | y ) , and joint probabilities p ( x , y ) : p ( x , y ) = p ( x ) p ( y | x ) = p ( y ) p ( x | y ) .

4. Method

In this section, we introduce how to set up the gaze information channel.

4.1. Gaze Information Channel

Gaze information channel has been proposed in our previous work [32]. A Markov chain is a stochastic model that describes a series of possible events X 1 , X 2 , X 3 , , in which the probability of each event depends only on the state of the previous event, or Markov property. If the state space is finite, the transition probability distribution can be represented by a matrix, called the transition matrix. A time-invariant Markov chain is characterized by its initial state and a probability transition matrix P = [ p i j ] [52]. A Markov chain { X i } is fully determined by the initial state X 0 and the transition matrix P = [ p i j ] , p i j = P r { X n + 1 = j | X n = i } , i , j { 1 , 2 , , s } , where s is the number of states.
A Markov chain is said to be irreducible if its state space is a single communicating class; in other words, if it is possible to get to any state from any other state. It is aperiodic if all its states are aperiodic, that is, the return to any state is not constrained to a number of steps multiple of any integer >1. An irreducible and aperiodic Markov chain has a positive stationary distribution, the stationary distribution is unique, and from any starting distribution, the distribution of X n tends to the stationary distribution as n . The stationary distribution can be calculated by Equation (5):
π P = π .
The stationary distribution represents the frequency of visits of each state.
In this paper, we divide a tested poster into s content-dependent AOIs. The set of AOIs can be represented as S = { 1 , 2 , , s } , and the gaze switching process can be described as a stochastic process { X t } , t = 1 , 2 , , n , x 1 , x 2 , , x n S . In [44], the Markov property has been tested. Once the stochastic process is modeled as a Markov process, we obtain the transition matrix P = [ p i j ] s × s and the stationary or equilibrium probability π .
Similar to the work in [53] in the 3D scene visibility context, and as proposed in [32] to study Van Gogh’s painting, we extend the Markov chain model [44] for gaze transitions when reading posters to an information channel, X Y , where X and Y are discrete random variables with alphabet X = Y = 1 , 2 , , s , corresponding to the AOIs. In this case, input variables p ( X ) and output variables p ( Y ) of gaze information channel represent the same regions with the same probabilities π . The conditional probability p ( j | i ) in the gaze information channel corresponds to the p i j of transition matrix P in the Markov chain. Contrary to the case in [32], where the AOIs where arbitrarily fixed for a painting, we consider the AOIs in the posters as being defined by the authors in their poster design, that is, the different sections that are contained in a poster.
The basic elements of the gaze information channel are thus the following ones:
  • The conditional probability p ( j | i ) is given by p i j , which represents the estimated probability of transitioning from i t h AOI to any j t h AOI given i t h AOI as the starting point. Matrix elements p i j are set to the number of transitions from i t h source AOI to j t h destination AOI for each participant and then the matrix is normalized relative to each source AOI (i.e., per row), p i j = n i j j = 1 n i j , i , j S . Conditional probabilities fulfill j Y p ( j | i ) = 1 , i X , that is, j = 1 s p i j = 1 , i { 1 , , s } .
  • The marginal probabilities of input X and output Y, p ( i ) and p ( j ) , are both given by the stationary probability π , π = ( π 1 , π 2 , , π s ) , giving the frequency of visits of each AOI.

4.2. Entropy and Mutual Information in Gaze Information Channel

From the previous definitions and Equations (1)–(4), Shannon’s information measures can be defined for the gaze information channel. We first introduce the entropy of the input (and also output), random variables with stationary distribution,
H s = H ( X ) = H ( Y ) = - i = 1 s π i log π i .
As the equilibrium distribution represents the average number of visits in each AOI, H s indicates the average uncertainty of gaze position between the AOIs. Low H s values of stationary entropy means that the observer prefers some AOIs over the other ones, while high values mean that visual attention is balanced between AOIs.
The conditional entropy of i t h row, H ( Y | i ) , is defined as
H ( Y | i ) = - j = 1 s p i j log p i j .
It gives the uncertainty that the next fixation would be the j t h AOI if it were presently in the i t h AOI.
The conditional entropy H t of the information channel is the average of row entropies
H t = H ( Y | X ) = i = 1 s π i H ( Y | i ) = - i = 1 s π i j = 1 s p i j log p i j .
It represents the average uncertainty of a transition between two areas of interest, or average uncertainty that remains about the destination AOI when the source AOI is known.
The joint entropy H ( X , Y ) of the information channel is the entropy of the joint distribution of X and Y
H ( X , Y ) = H ( X ) + H ( Y | X ) = H s + H t = i = 1 s j = 1 s π i p i j log π i p i j
and gives the total uncertainty of the channel. The mutual information of i t h row, I ( i ; Y ) , is given by
I ( i ; Y ) = j = 1 s p i j log p i j π j
and represents the degree of correlation between AOI i and all the AOIs. The measures I ( i ; Y ) and H ( Y | i ) show in general opposite behavior patterns. A high value of H ( Y | i ) represents a high degree of uncertainty about next area of interest, while a high value of I ( i ; Y ) indicates the next AOI is known with high probability.
The mutual information I ( X ; Y ) is given by
I ( X ; Y ) = H ( X ) + H ( Y ) - H ( X , Y ) = i = 1 s π i I ( i ; Y ) = i = 1 s j = 1 s π i p i j log p i j π j
and represents the total correlation, or information shared, between the AOI’s.

5. Experiment and Data Collection

5.1. Materials

To set up the test, we selected, with permission of authors, seven image processing research posters as the testing materials. All students participating in the experiment had sufficient knowledge background to understand the tested posters, and confirmed that they had never seen the materials before. The posters AOIs followed the sections defined by the authors of the posters. For the sake of display in this paper, we combine all tested posters into one image, Figure 1. In order to make the reader of this paper more aware of the AOIs, we use the red block diagram to mark them in Figure 1, but in the eye tracking experiment, the red block diagram will not be displayed to participants. Moreover, we blurred the author and institutional information.

5.2. Participants

A total of 10 master in computer science students (male: 5, female: 5) from Tianjin University (Tianjin, China) participated in the eye tracking experiment. Their ages range from 22 to 28 years (average: 23.75, standard deviation: 1.5). All the participants can understand well English and have normal color vision. They had enough background to understand the posters, although they had not seen them before. Before the experiment, all participants signed a consent form.

5.3. Procedure

Equipment calibration was completed prior to the experiment. Then, the participants were instructed to view the poster as though they were reading papers as usual. These tested materials were presented for 60 s. Everyone was seated in an office chair, and asked to lean forward to rest his/her chin comfortably, with his/her head is 60 cm distant from the computer screen. During the eye tracking, there was no interaction between the operator and the participants. The posters were presented one after another. After the experiment, each participant was asked to review individually the poster once again and explain the core idea of the tested poster.

5.4. Apparatus and Site Setting

The SMI iViewETG2.7 eye tracker (Sensomotoric Instruments, Teltow, Germany) and BeGaze 3.7 software (Sensomotoric Instruments, Teltow, Germany) were utilised for data collection and for computing eye gaze metrics. The participants wore the eye tracker and looked at the high resolution ( 1920 × 1080 ) 24 inch LCD monitor that displayed the tested posters. The experiment was conducted in a quiet room. The curtains of the room were pulled to avoid uncontrollable light and reflection.

5.5. Data Collection

Eye movements were recorded with an SMI iViewETG2.7 eye tracking system. The raw video data was produced by the iViewETG2.7, and then video data was input to the eye tracking analysis software BeGaze 3.7 to edit AOIs, produce some visualizations (scanpath, heatmap and the bin charts of AOIs), and generate a series of fixations. Each fixation contains four parameters: the start time, the duration, and the X and Y position on the screen. The following analysis is based on this format of eye tracking data.

6. Results Analysis

6.1. Traditional Metrics

We first applied some traditional metrics and visualizations on the collected eye tracking data. These include: scanpaths with fixation count, heatmaps, bin charts, and the bin charts of AOIs. Since the limited space of this article, we just take participants 2 and 5, selected at random, as an example to present the result analysis for the convenience of display. Figure 4 shows the scanpaths of the tested posters from participant 2 and participant 5, respectively. The diameter of fixations, for all scanpaths, is set as 80 px = 500 ms. For the sake of space, in this section, we focus on only two participants and three posters based on the different number of AOIs to show the analysis of the results. Figure 4 left shows the scanpaths of participant 2; we can observe that the participant 2 is more focused on the result section of these posters because there are more fixations and duration in the image and table section. In contrast, the scanpaths of Figure 4 right of the participant 5 is free and more random. We can observe that this participant is not interested in the image and table area of the results section. Participant 5 is more focused on the overall reading and understanding.
Figure 5 and Figure 6 present the heatmaps, from participant 2 and participant 5, respectively. Obviously, the result of the heatmap is similar to the result of the scanpath, as it is another representation of the same data. We can compare in Figure 5 and Figure 6 the participants’ attention distribution within the poster. Observe that ranges are not unified, thus the fixation times should be compared in Figure 4. The heatmap of the participant 2 shows that the participant 2 is more focused on the result section of the posters. In contrast, the heat map of Figure 6 of the participant 5 is free and more discrete, and focused on the text. We can see that participant 5 is not interested in the image and table detail area of the posters.
Figure 7 presents the bin charts of AOIs from participant 2 and participant 5, respectively. It shows the relative visual intake time of which AOI the observer falls on at each time. We can find that participant 2 prefers AOI 3 (results section) for posters 2 and 4, and AOI 5 (results section) for poster 7. The left bin charts show a large number of red areas for posters 2 and 4 (AOI 3) and blue areas for poster 7 (AOI 5), which are the AOIs corresponding to the result part in the test posters. Participant 5 is more focused too on AOI 3 for posters 2 and 4 (results section), as the red area from AOI bin chart is large in Figure 7 right, but it is focused on AOI 3 (method section) and AOI 6 (conclusions section) for poster 7 because the red area and cyan area from AOI bin chart is very large. This shows there can be individual differences between participants in reading the same poster.

6.2. Entropy and Mutual Information in Gaze Information Channel

We consider each poster divided into content-dependent AOIs. As described in Section 4, we compute the entropy and mutual information of the gaze information channel to quantify cognitive comprehension for each participant.

6.2.1. Transition Matrices

In order to better understand the process of participant’s eye movement, we first analyze the gaze transition matrix when the participant views the tested posters. Table 1 provides the transition matrix before normalization of three tested posters (posters 1, 2, 3, with 3 AOIs) for all participants. That is, we accumulate in a single matrix all the transitions by all participants for these three posters. It can be observed that there are about 1200 fixations in total in the transition matrix, and the numbers (bold gray value in table) on the diagonal of the transition matrix are larger. This is consistent with a common sense interpretation, as the participant, before shifting to another area, will explore the area it is in until he/she has an understanding of it. Thus, participants’ cognitive process creates these transition matrix data shown in Table 1. Similarly, Table 2 shows the transition matrix before normalization of three tested poster (posters 4,5,6, with four AOIs) by all participants.

6.2.2. Comparison between Two Participants for Poster 7

Here, we show first in Figure 8 the results for poster 7 for all participants and then we compare more finely for participants 2 and 5. Observe from Figure 8a that the AOIs more visited by all participants are AOI5 (results section), AOI6 (conclusions section), and AOI3 (method section), although the most visited area depends on the participant. The majority of participants prefer, or visit it often, AOI5 (results section), others AOI6 (conclusions section), and finally others AOI3 (method section). Figure 8b shows the main measures of the channel for each participant, some of them are similar for several participants, although from Figure 8c,d, we can conclude that the exploration strategy can be in general different for each participant.
Next, Table 3 shows the transition probabilities of the participants 2 and 5 for the poster with more areas of interest, poster 7 with six AOIs. See Figure 9 for an illustration of the gaze channel for participant 5. Observe that, in Table 3, the values of p i i are the highest transition probabilities, which is consistent with the above transition matrix analysis. This is similar to The tempest painting example presented in [44]. As observed before, this means that, before switching to another AOI, the observer firstly moves the gaze within the current AOI. As shown in Table 3, we can clearly find that there is no direct transition between AOI 2 and AOI 6 when viewing the tested poster. The reason might be that the AOI 2 (introduction section of the poster) is far apart from AOI 6 (the conclusion section of the poster).
Table 4 and Figure 10 show the values for the equilibrium distribution, H s , H t , H ( Y | x ) , H ( X , Y ) , I ( X ; Y ) and I ( x ; Y ) , for the gaze information channel for participants 2 and 5. The gaze entropy H s is the entropy of the equilibrium distribution π , which indicates how frequently each AOI is visited. Note that currently in our gaze channel model, as in Markov chain model, we do not support fixation time, thus number of visits does not directly translate into spent time, although it can be considered as an approximation. From Table 4 and Figure 10, we can find that the AOIs that the participants prefer, AOI 5 (results section) for participant 2, and AOI 3 (method section) and AOI 6 (conclusions section) for participant 5, have the larger equilibrium distribution π i value. This is consistent with Figure 7 charts for poster 7. A higher value of H s means that the participant visited more equally each AOI. A lower value of H s is obtained when the number of visits in each AOI is not balanced, possibly because the participant spent more time concentrated on a certain region. It can be seen from Table 4 and Figure 10 that the entropy H s of the participant 5 is greater than for the participant 2. This means that the participant 5 pays more attention to overall reading and spent time more equally among AOIs than the participant 2. This conclusion is consistent with the previous scanpath analysis from Figure 4.
H t reflects the randomness of gaze transition among the different AOIs. Higher H t values mean that there are frequent transition among AOIs, while lower H t values indicate more careful observation of AOIs [44]. H ( Y | i ) measures the randomness of the gaze transfer from the i-th AOI. A lower value of H ( Y | i ) indicates that the observer is more clear about the next AOI in the following view. It may also represent that the i-th AOI provides the observer with significant clues to understand the test poster. From Table 4 and Figure 10, we can find that, for participant 2, H ( Y | 1 ) has the highest value, which means that when in AOI1 (title section of the poster), the observer moves randomly (or evenly) towards any of the other neighbour AOIs. For participant 5, H ( Y | 2 ) has the highest value, which indicates that this participant moves evenly from AOI 2 (intro section) to any AOI of the poster. Moreover, we can also see that I ( 3 ; Y ) has the lowest value, which represents that the information shared between AOI3 (method section) and all the AOIs is minimum. H ( X , Y ) = H s + H t measures the total uncertainty, or total randomness of fixations distribution and gaze transition. The lowest value of H ( X , Y ) is obtained when the participant 2 views the poster. Compared with the participant 5’s scanpath in Figure 4, the scanpath with lowest H ( X , Y ) has higher fixation length and less gaze transitions.
As expected, we can observe in Table 4 and Figure 10 that I ( i ; Y ) and H ( Y | i ) show in general opposite behavior. Higher values of I ( i ; Y ) correspond to lower values of H ( Y | i ) and viceversa. The values of I ( 4 ; Y ) for participant 2 and I ( 1 ; Y ) for participant 5 are higher than the other values of I ( i ; Y ) . This indicates that next AOIs when leaving AOI4 (algorithm section) for the participant 2, and leaving AOI1 (title section) for participant 5, were well defined, as a high value of I ( i ; Y ) means that the next AOI is known with high probability. This behavior can be re-confirmed in the corresponding scanpaths in Figure 4. Furthermore, from Table 4 and Figure 10 we can see that participant 5 has the highest I ( X ; Y ) value. Mutual information I ( X ; Y ) expresses the degree of dependence between the AOIs. It might mean that participant 5 has a more precise strategy or more clues in exploring the tested poster. However, this is in apparent contradiction to the fact that total uncertainty of participant 5 is higher than for participant 2. To be able to compare the mutual information between the two participants, we should first normalize it. Several normalization proposals exist in the literature [54]. If we consider for instance the one defined in [47] as a correlation coefficient ρ = I ( X ; Y ) H ( X ) = I ( X ; Y ) H s , the value of ρ for participant 2 is 0 . 643 , and for participant 5 is 0 . 644 , practically the same. Thus, in this case, we can not discover any difference based on mutual information.

6.2.3. Averaging Results for All Posters and Participants

We can find in the Appendix the Table A1, Table A2, Table A3, Table A4, Table A5 and Table A6, with the values for all participants and posters of I ( X ; Y ) , H s , H t and H ( X , Y ) , and I ( X ; Y ) normalized by H s and H ( X , Y ) , respectively. For instance, Table A1 lists the mutual information I ( X ; Y ) of all participants when they view all tested posters, the average value and standard deviation for each poster is given in the last two rows. It can be observed clearly that the MI values for tested poster 7 (with six AOIs) are much larger for all participants in general than for the other posters, which may indicate that the degree of dependence or correlation between AOIs of poster 7 is much stronger. We observe also that, although values of MI for different posters might be significantly different, the differences are reduced when considering the average MI value. These facts are confirmed looking at the normalized MI (see Table A5 and Table A6).
We have summarized Table A1, Table A2, Table A3 and Table A4 in Figure 11 and Figure 12. This allows readers to more intuitively observe the quantitative gaze collection of all participants. Figure 11 shows the stacked H t , H s , H ( X , Y ) and I ( X ; Y ) in the gaze information channel from all participants when they view all tested posters. From the stacked H s and H t bar chart in Figure 11a, we see that, for every participant, the values of joint entropy H ( X , Y ) (marked in gray color) approximately equal the total of H s and H t . Their total is equal for each separated transition matrix, Figure 11 shows that using averages is a valid approach. The joint entropy H ( X , Y ) measures the total uncertainty, which gives the uncertainty when every participant views the tested poster. At the same time, we can find that the values of the conditional entropy or transfer entropy H t (given by the crimson color bar) are close for all participants. This phenomenon illustrates, for all participants, when they begin to reading the test poster, they always like to switch between the different AOIs to better understand the context of the poster. This is consistent with the property of H t which reflects the randomness of gaze transition among the different AOIs.
From the right stacked H t and I ( X ; Y ) chart in Figure 11, we can see that H s (as marked in blue color) is approximately equal to the H t plus I ( X ; Y ) (see previous remark about totals). Mutual information (MI) I ( X ; Y ) in gaze information channel represents the degree of dependence or correlation between the set of AOIs. Furthermore, H s , which is the entropy of the equilibrium distribution π , measures how much equally the AOIs have been visited. From the blue bars in Figure 11a, it is clear that the participants 3, 5, 8, 9 spent more balanced time in each AOI when they read the tested poster since their H s is larger compared with the participants 1, 7, 10. This means that the participants 1, 7, 10 possibly spent more time concentrated on certain regions of the tested poster.
Figure 12 also shows the stacked H t , H s , H ( X , Y ) and I ( X ; Y ) in gaze information channel for all tested posters. According to Figure 12, we could consider the posters into three groups, the first one with poster 1, with the lowest value of H ( X , Y ) and H s , a second group with posters 2–6, with similar value of H ( X , Y ) and H s , and a third one with poster 7, with highest value of H ( X , Y ) and H s . Looking at Figure 1, we observe that poster 1 has one AOI that does not practically include relevant information, AOI3, this explains the lower values for this poster, as this AOI will be mostly ignored by participants. On the other extreme, poster 7 with six areas of interest is the more complex of all them. It also has the highest mutual information, and also, from Table A5 and Table A6, the highest normalized mutual information. It might mean that, although it is a more complex poster than the other ones, it is well structured and readers establish a coherent reading strategy.
Looking now at Figure 12b, we can observe the differences between the posters in the second group, from 2 to 6. All of them have similar H s value, but, in poster 2, the distribution is different. For poster 2, the mutual information I ( X ; Y ) is higher (and correspondingly the entropy H t is lower) than for posters 3–6. This is further checked by taking a look at Table A1. It means that this poster is easier to read or to interpret than posters 3–6. It can also be seen from Table 5, where we have classed the results of the explaining the core idea stage after the experiment into two groups: expressing the core ideas basically (called basic group), and saying only some keywords (called keywords group), Table 5 gives the participants from both groups for all tested posters. Although due to the low number of participants we can not draw any conclusive result, it seems that higher mutual information in posters 2 and 7 is related to a higher cognitive comprehension. It might work in an indirect way, that is, higher MI means more coherent exploration strategies that facilitate the comprehension of the poster.
Having a look at Figure 1, we see that poster 2 contains just text in the middle AOI, being probably easier the flow from graphics to text and graphics again than in the other posters. In addition, we see that posters 4–6, although they contain four areas of interest, one of them contains very little relevant information to understand the posters, thus, although we should in principle expect more information and uncertainty with four areas than with three, the results are similar. Observe that, for the analysis of posters 2–6, we do not need to consider the normalized mutual information, as we had to do in Section 6.2.2, as we compare posters with similar values of H s .

7. Discussion

We consider the information channel metrics as complementary to classic metrics for eye tracking. Actually, the information channel models the eye tracking process from an information theoretic perspective, extending the Markov chain model introduced by Krejtz et al. [44,45], and reviewed in [31]. The information channel interpretation of data communication has been successful in many areas of science, and specifically in visual computing, and we believe it also has a place in understanding eye tracking.
In particular, as already observed for Markov chain model, for stationary entropy H s and transition entropy H t , greater stationary entropy H s means that the participant visited more equally the AOIs, while higher transition entropy H t denotes more uncertainty and more frequent transition between AOIs. In terms of reading a poster, it can give information on the strategy of an observer. With only H s and H t , which are the metrics for the Markov channel, it is difficult to discriminate the behaviour of observers. Our model introduces the additional metrics H ( X , Y ) , H ( Y | x ) , I ( X ; Y ) , and I ( x ; Y ) , interpreted as the total uncertainty, the uncertainty from a given AOI, the total information shared between the AOIs, and the information shared between an AOI and all AOIs, respectively. For instance, observe from Figure 10 how we can clearly differentiate the behaviour of two observers, by using H ( Y | x ) and I ( x ; Y ) , and in less amount using H ( X , Y ) , metrics that are only available once you extend the Markov chain model of eye tracking to gaze information channel.
The information channel paradigm also has the advantage of easily clustering or classification, see Table A1, Table A2, Table A3 and Table A4, and its visualization in Figure 11 and Figure 12. Given a group of observers and a poster, the transition matrices in the information channels corresponding to one class can be averaged to obtain the information channel of the class, to help understand the behaviour of that class. However, we can also obtain the average of a single observer for the different posters, by averaging the measure values obtained. The averaged results give us hints about the behaviour of observers for poster reading, and the different difficulty of reading each poster for all the observers. We believe that, in addition to help understand the cognitive process of poster reading, clues can be gathered for improving the poster design.
One weak point of the information channel model for eye tracking trajectories is that, as in the Markov chain model, the channel depends on the AOIs defined, so that changing the areas of interest the information channel measures values change. This is the same situation encountered in [53]. However, changing AOIs does not need repeating the observations, but just recomputing to which AOI belong the hit points of gaze trajectories; thus, computing the channel for different configurations of AOIs could be straightforward. The criterion of maximizing mutual information I ( X ; Y ) gain, or minimizing its loss, for optimal subdivision or clustering [53,55], could also be used in the gaze information channel. In this paper, we have used the sections of a poster, as defined by the poster authors, as AOIs, which we thus consider semantically meaningful, although the maximization of I ( X ; Y ) could help further in the design of the poster sections.

8. Conclusions and Future Work

To find quantitative links between eye movements and cognitive comprehension, we tracked 10 observers’ eye movements for reading published posters. We model eye tracking fixation sequences between content-dependent Areas of Interests (AOIs) as a Markov chain. Furthermore, we use the fact that an irreducible and aperiodic Markov chain is a special case of information or communication channel, where input and output are the same random variable, and equal to the equilibrium distribution. Thus, the gaze transition can be modeled as a discrete information channel, the gaze information channel. Next, some traditional eye tracking metrics, together with the gaze entropy and mutual information of the gaze information channel are calculated to quantify cognitive comprehension for every participant. As far as we know, this is the first study to use the eye tracking technology to assess cognitive comprehension when reading a scientific poster. The present work provides insights into quantitative cognitive comprehension. Although promising, there are limitations (such as a limited number of participants) to this paper that need to be addressed in continuing work. In the future, we will continue to explore the unique significance of human visual search patterns, which need to be paired with behavioral or cognitive metrics. As MI seems to be related to coherent strategies in reading a poster, we will check the difference in the gaze channel measurements for different poster design for the same content, similar to [33]. We will study the best division in AOIs, driven by the maximum transfer of information, or MI. We will also extend the information channel paradigm to the work of Ponsoda et al. [37], that is, the Markov chain of gaze displacement directions will be extended to an information channel, as we have done here with the trajectories. In addition, our current gaze channel model does not support fixation time, thus although the number of visits given by the equilibrium distribution can be a rough approximation of spent time in each AOI, for a more complete analysis we plan to integrate the fixation time into the model.

Author Contributions

Q.H. implemented the approach and performed the experiments. M.S. introduced the use of the information channel to study gaze. L.M. assisted experiment execution and recorded related information from all participants. All authors have read and approved the final manuscript.

Funding

This work has been funded by the National Natural Science Foundation of China under grants No.61471261 and No.61771335, and by grants TIN2016-75866-C3-3-R from the Spanish Government.

Acknowledgments

We thank Miquel Feixas for his valuable insight. We thank also anonymous reviewers for their valuable comments on previous versions of this manuscript.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Table A1. Mutual information, I ( X ; Y ) , values in gaze information channel for all posters and all participants.
Table A1. Mutual information, I ( X ; Y ) , values in gaze information channel for all posters and all participants.
MITested
Poster1
Tested
Poster2
Tested
Poster3
Tested
Poster4
Tested
Poster5
Tested
Poster6
Tested
Poster7
Average
Value
Standard
Deviation
Participant 10.4220.4890.5840.4740.5480.4740.7160.5300.091
Participant 20.4690.6080.3770.5450.4110.3980.9080.5310.172
Participant 30.7820.5570.5330.3940.5120.4280.9690.5960.191
Participant 40.6640.6410.5820.4110.3510.5871.0430.6110.207
Participant 50.6550.5420.4220.4150.6710.4490.9580.5870.180
Participant 60.4970.7860.3010.4290.5440.4810.9560.5710.208
Participant 70.5750.5080.5120.4390.4640.2940.6090.4860.095
Participant 80.6380.5440.6110.4120.4520.4061.0620.5890.211
Participant 90.5120.4980.4700.4270.4970.3971.0230.5460.198
Participant 100.5810.4150.4370.5140.2880.3700.4840.4410.089
Average Value0.5800.5590.4830.4460.4740.4280.873
Standard Deviation0.1030.0970.0950.0470.1030.0740.189
Table A2. H s values in gaze information channel for all posters and all participants.
Table A2. H s values in gaze information channel for all posters and all participants.
H s Tested
Poster1
Tested
Poster2
Tested
Poster3
Tested
Poster4
Tested
Poster5
Tested
Poster6
Tested
Poster7
Average
Value
Standard
Deviation
Participant 10.7110.8330.9390.9301.0210.9240.9480.9010.093
Participant 20.8890.9420.7441.150.7550.881.4120.9670.221
Participant 31.0081.0511.0771.0871.0221.1151.2931.0930.088
Participant 40.9040.9361.0610.9030.9300.9801.3231.0050.139
Participant 50.8511.0251.0500.9701.2290.9151.4871.0750.201
Participant 60.9190.9180.9410.9301.0231.1281.3601.0310.152
Participant 70.8360.9981.0300.9180.9290.9991.0040.9590.063
Participant 80.9141.0661.0110.8081.1841.1201.3861.0700.174
Participant 90.7981.0921.0550.9961.0150.7801.5691.0440.242
Participant 100.8221.0931.0940.9590.7270.9341.0090.9480.126
Average Value0.8650.9951.0000.9650.9840.9781.279
Standard Deviation0.0770.0820.0990.0910.1520.1090.206
Table A3. H t values in gaze information channel for all posters and all participants.
Table A3. H t values in gaze information channel for all posters and all participants.
H t Tested
Poster1
Tested
Poster2
Tested
Poster3
Tested
Poster4
Tested
Poster5
Tested
Poster6
Tested
Poster7
Average
Value
Standard
Deviation
Participant 10.2890.3440.3550.4560.4730.4670.2320.3740.088
Participant 20.4200.3330.3670.6160.3440.4660.5040.4360.094
Participant 30.2260.4940.5430.6930.5090.6870.3240.4970.160
Participant 40.2400.2950.4780.4920.5800.3930.2800.3940.118
Participant 50.1960.4830.6270.5550.5580.4530.5290.4860.129
Participant 60.4220.1320.6390.5050.4790.6470.4040.4610.161
Participant 70.2610.4910.5180.4800.4650.7040.3960.4740.123
Participant 80.2760.5230.3980.3960.7320.7140.3240.4800.169
Participant 90.2860.5920.5820.5690.5180.3830.5370.4950.108
Participant 100.2380.6780.6580.4450.4390.5570.5250.5060.139
Average Value0.2850.4370.5170.5210.5100.5470.406
Standard Deviation0.0730.1520.1080.0840.0970.1240.108
Table A4. H ( X , Y ) values in gaze information channel of all posters and all participants.
Table A4. H ( X , Y ) values in gaze information channel of all posters and all participants.
H(X,Y)Tested
Poster1
Tested
Poster2
Tested
Poster3
Tested
Poster4
Tested
Poster5
Tested
Poster6
Tested
Poster7
Average
Value
Standard
Deviation
Participant 11.0001.1761.2931.3851.4941.3911.1801.2740.155
Participant 21.3101.2751.1111.7651.0991.3461.9161.4030.293
Participant 31.2341.5441.6201.7811.5321.8021.6171.5900.175
Participant 41.1441.2311.5391.3951.5101.3731.6031.3990.155
Participant 51.0471.5081.6781.5261.7871.3672.0161.5610.287
Participant 61.3411.0501.5801.4351.5021.7751.7641.4920.234
Participant 71.0971.4891.5481.3981.3931.7031.4001.4330.172
Participant 81.1901.5891.4091.2041.9151.8351.7101.5500.270
Participant 91.0831.6831.6371.5661.5321.1642.0151.5390.316
Participant 101.0601.7711.7521.4041.1661.4921.5341.4540.250
Average Value1.1511.4321.5171.4861.4931.5251.685
Standard Deviation0.1090.2240.1840.1700.2320.2230.268
Table A5. Normalized mutual information by H s in gaze information channel of all posters and all participants.
Table A5. Normalized mutual information by H s in gaze information channel of all posters and all participants.
Normalized MITested
Poster1
Tested
Poster2
Tested
Poster3
Tested
Poster4
Tested
Poster5
Tested
Poster6
Tested
Poster7
Average
Value
Standard
Deviation
Participant 10.5940.5870.6220.5100.5370.5130.7550.5880.079
Participant 20.5280.6450.5070.4740.5440.4520.6430.5420.071
Participant 30.7760.5300.4950.3620.5010.3840.7490.5420.151
Participant 40.7350.6850.5490.4550.3770.5990.7880.5980.138
Participant 50.7700.5290.4020.4280.5460.4910.6440.5440.118
Participant 60.5410.8560.3200.4610.5320.4260.7030.5480.166
Participant 70.6880.5090.4970.4780.4990.2940.6070.5100.113
Participant 80.6980.5100.6040.5100.3820.3630.7660.5480.141
Participant 90.6420.4560.4450.4290.4900.5090.6520.5170.086
Participant 100.7070.3800.3990.5360.3960.3960.4800.4710.110
Average Value0.6680.5690.4840.4640.4800.4430.679
Standard Deviation0.0840.1270.0900.0480.0650.0840.089
Table A6. Normalized mutual information by H ( X , Y ) in gaze information channel of all posters and all participants.
Table A6. Normalized mutual information by H ( X , Y ) in gaze information channel of all posters and all participants.
Normalized MITested
Poster1
Tested
Poster2
Tested
Poster3
Tested
Poster4
Tested
Poster5
Tested
Poster6
Tested
Poster7
Average
Value
Standard
Deviation
Participant 10.4220.4160.4520.3420.3670.3410.6070.4210.092
Participant 20.3580.4770.3390.3090.3740.2960.4740.3750.073
Participant 30.6340.3610.3290.2210.3340.2380.5990.3880.165
Participant 40.5800.5210.3780.2950.2320.4280.6510.4410.152
Participant 50.6260.3590.2510.2720.3750.3280.4750.3840.130
Participant 60.3710.7490.1910.2990.3620.2710.5420.3980.189
Participant 70.5240.3410.3310.3140.3330.1730.4350.3500.109
Participant 80.5360.3420.4340.3420.2360.2210.6210.3900.149
Participant 90.4730.2960.2870.2730.3240.3410.5080.3570.094
Participant 100.5480.2340.2490.3660.2470.2480.3160.3150.113
Average Value0.5070.4100.3240.3030.3180.2890.523
Standard Deviation0.0990.1450.0830.0420.0580.0740.103

References

  1. Was, C.; Sansosti, F.; Morris, B. Eye-Tracking Technology Applications in Educational Research; IGI Global: Hershey, PA, USA, 2016. [Google Scholar]
  2. Prieto, L.P.; Sharma, K.; Wen, Y.; Dillenbourg, P. The Burden of Facilitating Collaboration: Towards Estimation of Teacher Orchestration Load Using Eye-tracking Measures; International Society of the Learning Sciences, Inc. [ISLS]: Albuquerque, NM, USA, 2015. [Google Scholar]
  3. Ellis, E.M.; Borovsky, A.; Elman, J.L.; Evans, J.L. Novel Word Learning: An Eye-tracking Study. Are 18-month-old Late Talkers Really Different From Their Typical Peers? J. Commun. Disord. 2015, 58, 143–157. [Google Scholar] [CrossRef]
  4. Fox, S.E.; Faulkner-Jones, B.E. Eye-Tracking in the Study of Visual Expertise: Methodology and Approaches in Medicine. Frontline Learn. Res. 2017, 5, 29–40. [Google Scholar] [CrossRef]
  5. Jarodzka, H.; Boshuizen, H.P. Unboxing the Black Box of Visual Expertise in Medicine. Frontline Learn. Res. 2017, 5, 167–183. [Google Scholar] [CrossRef]
  6. Fong, A.; Hoffman, D.J.; Zachary Hettinger, A.; Fairbanks, R.J.; Bisantz, A.M. Identifying Visual Search Patterns in Eye Gaze Data; Gaining Insights into Physician Visual Workflow. J. Am. Med. Inform. Assoc. 2016, 23, 1180–1184. [Google Scholar] [CrossRef]
  7. McLaughlin, L.; Bond, R.; Hughes, C.; McConnell, J.; McFadden, S. Computing Eye Gaze Metrics for the Automatic Assessment of Radiographer Performance During X-ray Image Interpretation. Int. J. Med. Inform. 2017, 105, 11–21. [Google Scholar] [CrossRef]
  8. Holzman, P.S.; Proctor, L.R.; Hughes, D.W. Eye-tracking Patterns in Schizophrenia. Science 1973, 181, 179–181. [Google Scholar] [CrossRef]
  9. Pavlidis, G.T. Eye Movements in Dyslexia: Their Diagnostic Significance. J. Learn. Disabil. 1985, 18, 42–50. [Google Scholar] [CrossRef]
  10. Zhang, L.; Wade, J.; Bian, D.; Fan, J.; Swanson, A.; Weitlauf, A.; Warren, A.; Sarkar, N. Cognitive Load Measurement in A Virtual Reality-based Driving System for Autism Intervention. IEEE Trans. Affect. Comput. 2017, 8, 176–189. [Google Scholar] [CrossRef] [PubMed]
  11. Vidal, M.; Bulling, A.; Gellersen, H. Pursuits: Spontaneous Eye-based Interaction for Dynamic Interfaces. GetMobile Mob. Comput. Commun. 2015, 18, 8–10. [Google Scholar] [CrossRef]
  12. Strandvall, T. Eye Tracking in Human-computer Interaction and Usability Research. In IFIP Conference on Human-Computer Interaction; Springer: Berlin/Heidelberg, Germany, 2010; pp. 936–937. [Google Scholar]
  13. Wang, Q.; Yang, S.; Liu, M.; Cao, Z.; Ma, Q. An Eye-tracking Study of Website Complexity from Cognitive Load Perspective. Decis. Support Syst. 2014, 62, 1–10. [Google Scholar] [CrossRef]
  14. Schiessl, M.; Duda, S.; Tholke, A.; Fischer, R. Eye tracking and Its Application in Usability and Media Research. MMI-interaktiv J. 2003, 6, 41–50. [Google Scholar]
  15. Steiner, G.A. The People Look at Commercials: A Study of Audience Behavior. J. Bus. 1966, 39, 272–304. [Google Scholar] [CrossRef]
  16. Lunn, D.; Harper, S. Providing Assistance to Older Users of Dynamic Web Content. Comput. Hum. Behav. 2011, 27, 2098–2107. [Google Scholar] [CrossRef]
  17. Van Gog, T.; Scheiter, K. Eye Tracking as A Tool to Study and Enhance Multimedia Learning; Elsevier: Amsterdam, The Netherlands, 2010. [Google Scholar]
  18. Navarro, O.; Molina, A.I.; Lacruz, M.; Ortega, M. Evaluation of Multimedia Educational Materials Using Eye Tracking. Procedia-Soc. Behav. Sci. 2015, 197, 2236–2243. [Google Scholar] [CrossRef] [Green Version]
  19. Van Wermeskerken, M.; van Gog, T. Seeing the Instructor’s Face and Gaze in Demonstration Video Examples Affects Attention Allocation but not Learning. Comput. Educ. 2017, 113, 98–107. [Google Scholar] [CrossRef]
  20. Stuijfzand, B.G.; van der Schaaf, M.F.; Kirschner, F.C.; Ravesloot, C.J.; van der Gijp, A.; Vincken, K.L. Medical Students’ Cognitive Load in Volumetric Image Interpretation: Insights from Human-computer Interaction and Eye Movements. Comput. Hum. Behav. 2016, 62, 394–403. [Google Scholar] [CrossRef]
  21. Ju, U.; Kang, J.; Wallraven, C. Personality Differences Predict Decision-making in An Accident Situation in Virtual Driving. In Proceedings of the 2016 IEEE Virtual Reality, Greenville, SC, USA, 19–23 March 2016; pp. 77–82. [Google Scholar]
  22. Chen, X.; Starke, S.D.; Baber, C.; Howes, A. A Cognitive Model of How People Make Decisions through Interaction with Visual Displays. In Proceedings of the 2017 CHI Conference on Human Factors in Computing Systems, Denver, CO, USA, 6–11 May 2017; pp. 1205–1216. [Google Scholar]
  23. Duchowski, A.T.; Driver, J.; Jolaoso, S.; Tan, W.; Ramey, B.N.; Robbins, A. Scanpath Comparison Revisited. In Proceedings of the 2010 Symposium on Eye-Tracking Research & Applications, Austin, TX, USA, 22–24 March 2010; pp. 219–226. [Google Scholar]
  24. De Bruin, J.A.; Malan, K.M.; Eloff, J.H.P. Saccade Deviation Indicators for Automated Eye Tracking Analysis. In Proceedings of the 2013 Conference on Eye Tracking South Africa, Cape Town, South Africa, 29–31 August 2013; pp. 47–54. [Google Scholar]
  25. Peysakhovich, V.; Hurter, C. Scanpath visualization and comparison using visual aggregation techniques. J. Eye Mov. Res. 2018, 10, 1–14. [Google Scholar]
  26. Mishra, A.; Kanojia, D.; Nagar, S.; Dey, K.; Bhattacharyya, P. Scanpath Complexity: Modeling Reading Effort Using Gaze Information. In Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017. [Google Scholar]
  27. Li, A.; Zhang, Y.; Chen, Z. Scanpath Mining of Eye Movement Trajectories for Visual Attention Analysis. In Proceedings of the 2017 IEEE International Conference on Multimedia and Expo (ICME), Hong Kong, China, 10–14 July 2017; pp. 535–540. [Google Scholar]
  28. Grindinger, T.; Duchowski, A.T.; Sawyer, M. Group-wise Similarity and Classification of Aggregate Scanpaths. In Proceedings of the 2010 Symposium on Eye-Tracking Research & Applications, Austin, TX, USA, 22–24 March 2010; pp. 101–104. [Google Scholar]
  29. Isokoski, P.; Kangas, J.; Majaranta, P. Useful Approaches to Exploratory Analysis of Gaze Data: Enhanced Heatmaps, cluster Maps, and Transition Maps. In Proceedings of the 2018 ACM Symposium on Eye Tracking Research & Applications, Warsaw, Poland, 14–17 June 2018; p. 68. [Google Scholar]
  30. Gu, Z.; Jin, C.; Dong, Z.; Chang, D. Predicting Webpage Aesthetics with Heatmap Entropy. arXiv 2018, arXiv:1803.01537. [Google Scholar]
  31. Shiferaw, B.; Downey, L.; Crewther, D. A review of gaze entropy as a measure of visual scanning efficiency. Neurosci. Biobehav. Rev. 2019, 96, 353–366. [Google Scholar] [CrossRef]
  32. Ma, L.J.; Sbert, M.; Xu, Q.; Feixas, M. Gaze Information Channel. In Pacific Rim Conference on Multimedia; Springer: Cham, Switzerland, September 2018; pp. 575–585. [Google Scholar]
  33. Qiang, Y.; Fu, Y.; Guo, Y.; Zhou, Z.H.; Sigal, L. Learning to Generate Posters of Scientific Papers. In Proceedings of the Thirtieth AAAI Conference on Artificial Intelligence, Phoenix, AZ, USA, 12–17 February 2016. [Google Scholar]
  34. Bavdekar, S.B.; Vyas, S.; Anand, V. Creating Posters for Effective Scientific Communication. J. Assoc. Phys. India 2017, 65, 82–88. [Google Scholar]
  35. Berg, J.; Hicks, R. Successful Design and Delivery of A Professional Poster. J. Am. Assoc. Nurse Pract. 2017, 29, 461–469. [Google Scholar] [CrossRef] [PubMed]
  36. Rezaeian, M.; Rezaeian, M.; Rezaeian, M. How to Prepare A Poster for A Scientific Presentation. Middle East J. Fam. Med. 2017, 7, 133. [Google Scholar] [CrossRef]
  37. Ponsoda, V.; Scott, D.; Findlay, J.M. A Probability Vector and Transition Matrix Analysis of Eye Movements During Visual Search. Acta Psychol. 1995, 88, 167–185. [Google Scholar] [CrossRef]
  38. Ellis, S.R.; Stark, L. Statistical Dependency in Visual Scanning. Hum. Factors 1986, 28, 421–438. [Google Scholar] [CrossRef]
  39. Liechty, J.; Pieters, R.; Wedel, M. Global and Local Covert Visual Attention: Evidence from A Bayesian Hidden Markov Model. Psychometrika 2003, 68, 519–541. [Google Scholar] [CrossRef]
  40. Helmert, J.R.; Joos, M.; Pannasch, S.; Velichkovsky, B.M. Two Visual Systems and Their Eye Movements: Evidence from Static and Dynamic Scene Perception. In Proceedings of the 2005 Annual Meeting of the Cognitive Science Society, Stresa, Italy, 21–23 July 2005; p. 27. [Google Scholar]
  41. Hwang, A.D.; Wang, H.C.; Pomplun, M. Semantic Guidance of Eye Movements in Real-world Scenes. Vision Res. 2011, 51, 1192–1205. [Google Scholar] [CrossRef] [PubMed]
  42. Bonev, B.; Chuang, L.L.; Escolano, F. How do Image Complexity, Task Demands and Looking Biases Influence Human Gaze Behavior? Pattern Recognit. Lett. 2013, 34, 723–730. [Google Scholar] [CrossRef]
  43. Besag, J.; Mondal, D. Exact Goodness-of-Fit Tests for Markov Chains. Biometrics 2013, 69, 488–496. [Google Scholar] [CrossRef] [PubMed]
  44. Krejtz, K.; Duchowski, A.; Szmidt, T.; Krejtz, I.; González Perilli, F.; Pires, A.; Vilaro, A.; Villalobos, N. Gaze Transition Entropy. ACM Trans. Appl. Percept. 2015, 13, 4. [Google Scholar] [CrossRef]
  45. Krejtz, K.; Szmidt, T.; Duchowski, A.; Krejtz, I.; Perilli, F.G.; Pires, A.; Vilaro, A.; Villalobos, N. Entropy-based Statistical Analysis of Eye Movement Transitions. In Proceedings of the 2014 Symposium on Eye Tracking Research and Applications, Safety Harbor, FL, USA, 26–28 March 2014; pp. 159–166. [Google Scholar]
  46. Raptis, G.E.; Fidas, C.A.; Avouris, N.M. On Implicit Elicitation of Cognitive Strategies using Gaze Transition Entropies in Pattern Recognition Tasks. In Proceedings of the 2017 CHI Conference Extended Abstracts on Human Factors in Computing Systems, Denver, CO, USA, 6–11 May 2017; pp. 1993–2000. [Google Scholar]
  47. Cover, T.M.; Thomas, J.A. Elements of Information Theory; John Wiley and Sons: Hoboken, NJ, USA, 1991; pp. 33–36. [Google Scholar]
  48. Shannon, C.E. A Mathematical Theory of Communication. Bell Syst. Tech. J. 1948, 27, 379–423. [Google Scholar] [CrossRef]
  49. Chen, M.; Feixas, M.; Viola, I.; Bardera, A.; Shen, H.W.; Sbert, M. Information Theory Tools for Visualization; CRC Press: Boca Raton, FL, USA, 2016. [Google Scholar]
  50. Ruiz, F.E.; Perez, P.S.; Bonev, B.I. Information Theory in Computer Vision and Pattern Recognition; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2009. [Google Scholar]
  51. Yeung, R.W. Information Theory and Network Coding; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2008. [Google Scholar]
  52. Gagniuc, P.A. Markov Chains: From Theory to Implementation and Experimentation; John Wiley & Sons: Hoboken, NJ, USA, 2017. [Google Scholar]
  53. Feixas, M.; del Acebo, E.; Bekaert, P.; Sbert, M. An Information Theory Framework for the Analysis of Scene Complexity. Comput. Gr. Forum 1999, 18, 95–106. [Google Scholar] [CrossRef] [Green Version]
  54. Hu, B.-G. Information Theoretic Learning in Pattern Classification. In Proceedings of the ICONIP Tutorial, Guangzhou, China, 14–18 November 2017. [Google Scholar]
  55. Tishby, N.; Pereira, F.C.; Bialek, W. The Information Bottleneck Method. In Proceedings of the 37th Annual Allerton Conference on Communication, Control and Computing, Monticello, IL, USA, 22–24 September 1999; pp. 368–377. [Google Scholar]
Figure 1. The poster materials used in our eye tracking experiment. From left to right, in top row, posters 1, 2, and 3, in bottom row, posters 4, 5, 6, and 7. Marked in red, the Areas of Interests (AOIs) in which each poster is divided, and that are not displayed to the participants.
Figure 1. The poster materials used in our eye tracking experiment. From left to right, in top row, posters 1, 2, and 3, in bottom row, posters 4, 5, 6, and 7. Marked in red, the Areas of Interests (AOIs) in which each poster is divided, and that are not displayed to the participants.
Entropy 21 00444 g001
Figure 2. The information diagram represents the relationship between Shannon’s information measures.
Figure 2. The information diagram represents the relationship between Shannon’s information measures.
Entropy 21 00444 g002
Figure 3. The main elements of an information channel.
Figure 3. The main elements of an information channel.
Entropy 21 00444 g003
Figure 4. The scanpaths for participant 2 (left) and participant 5 (right) for posters 2, 4 and 7.
Figure 4. The scanpaths for participant 2 (left) and participant 5 (right) for posters 2, 4 and 7.
Entropy 21 00444 g004
Figure 5. The heatmaps for participant 2 for posters 2, 4 and 7.
Figure 5. The heatmaps for participant 2 for posters 2, 4 and 7.
Entropy 21 00444 g005
Figure 6. The heatmaps for participant 5 for posters 2, 4 and 7.
Figure 6. The heatmaps for participant 5 for posters 2, 4 and 7.
Entropy 21 00444 g006
Figure 7. The AOI bin charts for participant 2 (left) and participant 5 (right) for posters 2, 4 and 7.
Figure 7. The AOI bin charts for participant 2 (left) and participant 5 (right) for posters 2, 4 and 7.
Entropy 21 00444 g007
Figure 8. The channel measures for all participants and poster 7. From left to right and top to down, the equilibrium distribution π (a), Hs, Ht, H(X,Y), I(X;Y) (b), H ( Y | x ) (c), and I(x;Y) (d). AOI1 is the title section, AOI2 is the intro section, AOI3 is the method section, AOI4 is the algorithm section, AOI5 is the results section, and AOI6 is the conclusions section.
Figure 8. The channel measures for all participants and poster 7. From left to right and top to down, the equilibrium distribution π (a), Hs, Ht, H(X,Y), I(X;Y) (b), H ( Y | x ) (c), and I(x;Y) (d). AOI1 is the title section, AOI2 is the intro section, AOI3 is the method section, AOI4 is the algorithm section, AOI5 is the results section, and AOI6 is the conclusions section.
Entropy 21 00444 g008
Figure 9. The gaze channel for participant 5 and poster 7, between the AOIs with equilibrium distribution p(AOI) (a) and (c), and with conditional probabilities (b). The probability distributions values are found in Table 3 and Table 4.
Figure 9. The gaze channel for participant 5 and poster 7, between the AOIs with equilibrium distribution p(AOI) (a) and (c), and with conditional probabilities (b). The probability distributions values are found in Table 3 and Table 4.
Entropy 21 00444 g009
Figure 10. From left to right and top to down, the equilibrium distribution π (a), Hs, Ht, H(X,Y), I(X;Y) (b), H ( Y | x ) (c), and I(x;Y) (d) in gaze information channel of participants 2 and 5 for poster 7. The numerical values are found in Table 4.
Figure 10. From left to right and top to down, the equilibrium distribution π (a), Hs, Ht, H(X,Y), I(X;Y) (b), H ( Y | x ) (c), and I(x;Y) (d) in gaze information channel of participants 2 and 5 for poster 7. The numerical values are found in Table 4.
Entropy 21 00444 g010
Figure 11. The stacked Ht, Hs (a), and Ht, I(X;Y) (b), for all participants.
Figure 11. The stacked Ht, Hs (a), and Ht, I(X;Y) (b), for all participants.
Entropy 21 00444 g011
Figure 12. The stacked Ht, Hs (a), and Ht, I(X;Y) (b), for all tested posters.
Figure 12. The stacked Ht, Hs (a), and Ht, I(X;Y) (b), for all tested posters.
Entropy 21 00444 g012
Table 1. Transition matrix before normalization of three tested poster (posters 1, 2, 3, with three Areas of Interests (AOIs)) by all participants observed. The highest values of each row, in bold, correspond to transitions within the same AOI.
Table 1. Transition matrix before normalization of three tested poster (posters 1, 2, 3, with three Areas of Interests (AOIs)) by all participants observed. The highest values of each row, in bold, correspond to transitions within the same AOI.
Tested PostersTested Poster 1Tested Poster 2Tested Poster 3
AOI1AOI2AOI3AOI1AOI2AOI3AOI1AOI2AOI3
AOI1419250AOI1337424AOI1241541
Transition matrixAOI22451819AOI24113842AOI24553750
AOI321964AOI3641506AOI31045269
Table 2. Transition matrix before normalization of three tested poster (posters 4, 5, 6 with four AOIs) by all participants observed. The highest values of each row, in bold, correspond to transitions within the same AOI.
Table 2. Transition matrix before normalization of three tested poster (posters 4, 5, 6 with four AOIs) by all participants observed. The highest values of each row, in bold, correspond to transitions within the same AOI.
Tested PostersTested Poster 4Tested Poster 5Tested Poster 6
AOI1AOI2AOI3AOI4AOI1AOI2AOI3AOI4AOI1AOI2AOI3AOI4
AOI1503221AOI1712310AOI1563210
T r a n s i t i o n m a t r i x AOI231253520AOI219149450AOI225159490
AOI345063621AOI354369029AOI383870038
AOI4012138AOI4103163AOI4143652
Table 3. Transition probability of tested poster 7 (with 6 AOIs) by participants 2 and 5.
Table 3. Transition probability of tested poster 7 (with 6 AOIs) by participants 2 and 5.
ObserversParticipant 2Participant 5
AOI1 AOI2 AOI3 AOI4 AOI5 AOI6AOI1 AOI2 AOI3 AOI4 AOI5 AOI6
AOI10.750 0.125 0.000 0.000 0.125 0.000AOI10.000 1.000 0.000 0.000 0.000 0.000
AOI20.000 0.818 0.182 0.000 0.000 0.000AOI20.111 0.333 0.556 0.000 0.000 0.000
T r a n s i t i o n p r o b a b i l i t y AOI30.000 0.125 0.813 0.063 0.000 0.000AOI30.000 0.111 0.806 0.083 0.000 0.000
AOI40.000 0.000 0.000 0.750 0.000 0.250AOI40.000 0.000 0.133 0.800 0.067 0.000
AOI50.000 0.000 0.000 0.000 0.894 0.106AOI50.000 0.000 0.000 0.000 0.778 0.222
AOI60.026 0.000 0.000 0.026 0.180 0.769AOI60.000 0.031 0.000 0.000 0.031 0.934
Table 4. Equilibrium distribution, H s , H t , H ( Y | x ) , H ( X , Y ) , I ( X ; Y ) and I ( x ; Y ) in gaze information channel of participants 2 and 5 for poster 7.
Table 4. Equilibrium distribution, H s , H t , H ( Y | x ) , H ( X , Y ) , I ( X ; Y ) and I ( x ; Y ) in gaze information channel of participants 2 and 5 for poster 7.
ObserversParticipant 2Participant 5
π (0.056, 0.076, 0.11, 0.028, 0.458, 0.271)(0.010, 0.088, 0.353, 0.147, 0.088, 0.314)
H s 1.4121.487
H t 0.5040.529
H ( Y | x ) (0.736, 0.474, 0.602, 0.562, 0.338, 0.698)(0.000, 0.937, 0.625, 0.628, 0.529, 0.277)
H ( X , Y ) 1.9162.016
I ( X ; Y ) 0.9080.958
I ( x ; Y ) (1.851, 2.030, 1.729, 2.452, 0.498, 0.613)(2.428, 0.965, 0.643, 1.207, 1.616, 0.961)
Table 5. The number of participants classed in two groups according to their answer after experiment, together with mutual information (MI), MI normalized by H s , and MI normalized by H ( X , Y ) .
Table 5. The number of participants classed in two groups according to their answer after experiment, together with mutual information (MI), MI normalized by H s , and MI normalized by H ( X , Y ) .
Tested PosterTested
Poster 1
Tested
Poster 2
Tested
Poster 3
Tested
Poster 4
Tested
Poster 5
Tested
Poster 6
Tested
Poster7
Basic Group4766445
Keywords Group6344665
MI0.5800.5590.4830.4460.4740.4280.873
MI normalized by H s 0.6680.5690.4840.4640.4800.4430.679
MI normalized by H ( X , Y ) 0.5070.4100.3240.3030.3180.2890.523

Share and Cite

MDPI and ACS Style

Hao, Q.; Sbert, M.; Ma, L. Gaze Information Channel in Cognitive Comprehension of Poster Reading. Entropy 2019, 21, 444. https://doi.org/10.3390/e21050444

AMA Style

Hao Q, Sbert M, Ma L. Gaze Information Channel in Cognitive Comprehension of Poster Reading. Entropy. 2019; 21(5):444. https://doi.org/10.3390/e21050444

Chicago/Turabian Style

Hao, Qiaohong, Mateu Sbert, and Lijing Ma. 2019. "Gaze Information Channel in Cognitive Comprehension of Poster Reading" Entropy 21, no. 5: 444. https://doi.org/10.3390/e21050444

APA Style

Hao, Q., Sbert, M., & Ma, L. (2019). Gaze Information Channel in Cognitive Comprehension of Poster Reading. Entropy, 21(5), 444. https://doi.org/10.3390/e21050444

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop