Toward Fully Automated Person-Independent Detection of Mind Wandering Robert Bixler1, Sidney D’Mello1,2 Departments of Computer Science1 and Psychology2, University of Notre Dame, Notre Dame, IN 46556 {rbixler, sdmello}@nd.edu

Abstract. Mind wandering is a ubiquitous phenomenon where attention involuntary shifts from task-related processing to task-unrelated thoughts. Mind wandering has negative effects on performance, hence, intelligent interfaces that detect mind wandering can intervene to restore attention to the current task. We investigated the use of eye gaze and contextual cues to automatically detect mind wandering during reading with a computer interface. Participants were pseudo-randomly probed to report mind wandering instances while an eye tracker recorded their gaze during a computerized reading task. Supervised machine learning techniques detected positive responses to mind wandering probes from gaze and context features in a user-independent fashion. Mind wandering was predicted with an accuracy of 72% (expected accuracy by chance was 62%) when probed at the end of a page and an accuracy of 59% (chance was 50%) when probed in the midst of reading a page. Possible improvements to the detectors and applications are discussed. Keywords: gaze tracking, mind wandering, affect detection, user modeling

1

Introduction

Most of us have had the experience of reading, listening to a lecture, or engaging in a personally-relevant task only to realize that our attention has gradually drifted away from the task at hand to off-task thoughts, such as dinner, childcare, or everyday worries and anxieties. It has been estimated that these automatic attentional shifts toward internal task-unrelated thoughts, referred to as mind wandering (MW, or zoning out), occur between 20-40% of the time depending on the task and the environmental context [12, 19, 24]. For example, one recent large-scale study tracked MW in 5,000 individuals with random prompts from an iPhone app and made the surprising discovery that people reported MW for 46.9% of the prompts [12]. MW is not merely incidental but also negatively effects performance for tasks requiring conscious control because an individual cannot simultaneously focus on both the task at hand and task-unrelated thoughts. Research has indicated that MW leads to performance failures during a number of tasks, such as increased error rates during signal detection tasks [22], lower recall during memory tasks [20] and poor comprehension during reading tasks [9]. For computer-supported tasks, this suggests that there is an opportunity for an intelligent interface to attempt to reorient attention to

the task at hand when MW occurs. This paper makes a first step towards this goal by automatically detecting MW in near real-time in a manner that generalizes to new users. 1.1

Related Work

User state estimation is a broad topic applicable to a variety of areas such as affective computing, attention-aware computing, augmented cognition, and social signal processing. MW detection is a unique instance of user state estimation, specifically the detection of attentional states (or the lack thereof in the case of MW). Interest in modeling user attentional states has spurred research into how interface design decisions affect attention. For example, Muir et al. [14] used eye movements to explore how attention was affected by adaptive hints in an educational game. There have also been efforts at attentional state estimation. Yonetani et al. [26] analyzed the relationship between eye gaze and saliency dynamics of a video in order to determine if a user was in a high level attention condition or a low level attention condition with an accuracy of 80.6%. Additionally, Navalpakkam et al. [15] created hidden markov models that accurately modeled attention while users chose to read one of eight articles displayed on a computer screen. Although a growing number of studies have investigated attention, with the two exceptions noted below, none have considered MW behaviors. MW is a distinct aspect of attentional state as it is defined by involuntary lapses in attention. While numerous studies have investigated MW [9, 18, 23], research on realtime detection of MW is in its infancy. Drummond and Litman [8] made one of the first attempts to automatically detect MW (operationalized as zoning out) with acoustic-prosodic (e.g., pitch) features. Their model attempted to discriminate “high” versus “low” zoning out while users were engaged in a spoken dialogue with an intelligent tutoring system. Their accuracy of 64% reflects an important first step in MW detection. However, their validation approach did not ensure independence of training and testing sets, so generalization to new users is unknown. Furthermore, their MW detector is limited to speech-enabled interfaces. In this paper, we consider eye gaze as a possible modality for MW detection. Eye movements are attractive since they are ubiquitous in most visual interfaces. There has also been previous research linking eye movements and the occurrence of MW. For example, research has found that individuals blink more often [24] and are less likely to fixate, re-fixate, and look backward through previously read text [17] when MW compared to normal reading. These studies, however, did not attempt to automatically detect MW from eye gaze and although eye gaze has been used for attentional state estimation (as cited above), there is no research directed at MW detection. We recently took a step in this direction by making an initial attempt to use eye gaze data to detect mind wandering during reading [6]. The best performing models (supervised learners) yielded a detection accuracy of 60% on a down-sampled corpus containing 50% “yes” and 50% “no” responses. Although these results are promising, they are limited by the fact that classification accuracy was not very impressive. Furthermore, the data set was downsampled prior to classification, thereby interfering with the natural distribution of MW responses (35%).

1.2

Current Work

This study reports the development and validation of one of the first (aside from exceptions above) fully automated user-independent detector of MW during reading. We focus on a computerized reading task since reading is a critical component of many real-world tasks and is negatively affected by MW [9]. Our approach to MW detection entails collecting eye gaze data and self-reports of MW while users read texts on a computer screen (see Figure 1). We then extract features from the eye gaze signal and contextual cues (see below) and use supervised classification techniques to build models that discriminate instance of MW from normal reading. The models are constructed and validated in a user-independent fashion, so we will have some confidence that they generalize to new users. The present research is novel in a number of respects. First, previous work on attentional state estimation has not considered MW, and other than the Drummond & Litman study [8] and our preliminary attempt [6], Figure 1. Gaze fixations (circles) and saccades this work represents the first large(lines) of a sample page. Diameter of circle is scale attempt at fully automated userproportional to fixation duration independent detection of MW. Second, it significantly expands upon our preliminary work with a much larger and more diverse data set. Third, we considered an enhanced set of gaze features in order to improve classification accuracy when compared to our preliminary attempt using an impoverished feature set. Fourth, eye gaze features were complemented with contextual cues, such as high-level text characteristics (e.g., difficulty) and reading behaviors (e.g., reading rate), as context might help disambiguate noisy gaze signals.

2

Data Collection

2.1

Participants

Participants were 178 undergraduate students from two U.S. universities that participated for course credit. 93 students were from a medium-sized private mid-western university while 85 were from a large public university in the mid-south. The average age of participants was 20 years (SD = 3.6). Demographics included 62.7% female, 49% Caucasian, 34% African American, 7% Asian, 6% Hispanic, and 4% “Other”. 2.2

Texts and Experimental manipulations

Participants read four different texts on research methods topics (i.e., experimenter bias, replication, causality, and dependent variables). On average the texts contained 1500 words (SD = 10) and were split into 30-36 pages with approximately 60 words per page. Texts were presented on a computer screen with size 36 Courier New font.

There were two manipulations: difficulty and value. The difficulty manipulation consisted of presenting either an easy or a difficult version of each text. The value manipulation pertained to questions from “high-value” texts which counted three times more toward the test score than questions for the “low-value” texts. Participants were instructed they would be required to read additional material if they did not do well enough on the posttest, incentivizing them to do well on the posttest (and thus the “high-value” questions) to avoid an unappealing additional task. The difficulty and value manipulations were part of a larger research study and are less relevant here. 2.3

Mind Wandering Probes

Auditory probes were used to measure Table 1. Incidence of mind wandering MW. Although misreports are possiResponse Type Yes No Total ble, there is no clear alternative for 209 651 860 End-of-page tracking such a highly internal phe1278 2839 4117 Within-page nomena, and auditory probes are the 1487 3490 4977 Total standard and validated method for collecting online MW reports [13, 23]. Nine pseudorandom pages in each text were identified as “probe pages.” An auditory probe (i.e., a beep) was triggered on probe pages at a randomly chosen time interval 4 to 12 seconds from the time the page appeared. These probes were considered to be within-page probes. An end-of-page probe was triggered if the participant tried to advance to the next page before the within-page probe was triggered. Participants were instructed to indicate if they were MW or not by pressing keys marked “Yes” or “No,” respectively. The instructions defined MW as consisting of having “no idea what you just read” and realizing that “you were thinking about something else altogether.” Table 1 provides a summary of analyzed MW reports. The MW probabilities of 24% for end-of-page and 31% for within-page probes are similar to previous studies [19, 23]. 2.4

Procedure

All procedures were approved by the ethics board of both Universities prior to any data collection. After signing an informed consent, participants were seated in front of either a Tobii TX 300 or Tobii T60 eye tracker depending on the university (both were in binocular mode). The Tobii eye trackers are remote eye trackers, so participants could read freely without any restrictions on head movement. Participants completed a brief 60-second standard calibration procedure. Participants were then instructed how to respond to the MW probes based on instructions from previous studies [9]. Next, they then read four texts for an average of 32.4 mins (SD = 9.09) on a page-by-page basis, using the space bar to navigate forward. They completed a posttest after reading all four texts and were fully debriefed.

3

Supervised Classification

3.1

Feature Engineering

Gaze fixations (points where gaze was maintained on the same location) were estimated from raw gaze data using a fixation filter from OGAMA, an open source gaze analyzer [25]. The series of gaze fixations and saccades (eye movements between fixations) were segmented into windows of varying length (4 secs, 8 secs, 12 secs), each ending with a MW probe. The windows ended immediately before the auditory probe was triggered in order to avoid confounds associated with motor activities in preparation for the key press in response to the probe. Windows that contained less than five fixations or windows that were shorter than four seconds were eliminated because these windows did not contain sufficient data to compute gaze features. Three sets of gaze features were computed: 30 global features, 19 local features, and 12 context features, yielding 61 features overall. Global features were independent of the actual words being read and fell into three categories. Fixation duration was the length of time in milliseconds of a fixation. Saccade duration was the time in milliseconds between two subsequent fixations. Saccade length was the distance in pixels between two subsequent fixations. For each of these three categories, we computed the number of events (fixations, saccades, etc), min, max, mean, median, standard deviation, skew, kurtosis, and range, thereby yielding 27 features. Three additional global features were computed, totaling 30 global features overall. Fixation dispersion was the root mean square of the distances of each fixation from the average position of all fixations in the window. Reading depth was the ratio of the actual reading time to an expected reading time calculated by multiplying the number of words in the window by 200ms, an estimated average fixation length in milliseconds during reading [16]. The last global feature was the fixation duration/saccade duration ratio. In contrast to global features, local features were sensitive to the words being read. The first set of local features consisted of different fixation types. First pass fixations were the first fixation on each word during the first pass through the text. Regression fixations were fixations on words that had already been passed. Single fixations were fixations on words that were only fixated on once. Gaze fixations were consecutive fixations on the same word. Non word fixations were fixations that were not on a word. Specific local features extracted from these different types of fixations included proportion of each fixation type compared to total number of fixations, and the mean and standard deviation of duration of each fixation type, totaling 15 local features. Four additional local fixations pertained to the extent to which well-known relationships between characteristics of words (in each window) and gaze fixations were observed. These included correlations between fixation durations and: (a) word length (number of characters), (b) hypernym depth, which is the semantic specificity of a word (i.e. “crimson” is more specific than “red”, which is more specific than “color”), (c) global frequency of a word, which is the overall frequency of the word in English as measured by the CELEX corpus [1], and (d) Synset size of a word, which was the number of synonyms of a word. The idea behind these features is that known relationships during normal reading, such as a negative correlation between word length and

fixation duration, should break down during MW compared to normal reading. Taken together, this yielded 19 local features in all. Context features captured the context of the reading task and included timing features and the difficulty and value of the text (see previous section). Session time, text time, and page time were the elapsed time between the MW probe and the beginning of the session, text, and page, respectively. Session page number and text page number were the number of pages read from the beginning of the session and text, respectively. Average page time was the average amount of time spent reading of all previous pages. Previous page time was the time spent reading the previous page. Previous page time ratio was the ratio of the previous page time to the average page time. Current difficulty and current value were the difficulty and value of the current text, respectively. Previous difficulty and previous value were the difficulty and value of the previous text, respectively. In all, there were 11 context features. 3.2

Model Building and Validation

Twenty supervised machine learning algorithms from WEKA [10] were used to build models discriminating MW (responding “Yes” to a MW probe) from normal reading (responding “No” to a MW probe). These included default WEKA implementations of: instance-based or lazy-learners (e.g., K-nearest neighbor), Bayesian probabilistic models, regression models, support vector machines, rule-based classifiers, decision trees, etc. We consider a large set of classifiers because we have no a priori prediction about the type of model that is best suited for this classification task. Models were built from datasets with a number of varying parameters in order to identify the most accurate models as well as to explore how different factors affect classification accuracy. First, data sets included either end-of-page or within-page MW reports as defined above (see Data Collection section). These report types were analyzed separately because they occur at different moments during reading and might potentially be associated with different gaze characteristics. Second, the features in each model were varied as each type of feature potentially captures a different aspect of gaze behaviors and we were interested in determining which feature set was most diagnostic of MW. Accordingly, models were built using global features, local features, context features, or a combination of the three. Third, models were built either with or without feature selection. Feature selection was performed in order to remove the negative influence of features that convey the same information (i.e. number of fixations and number of saccades) and restrict the feature space in order to address the “curse of dimensionality” [7]. Features that were strongly correlated with other features but weakly correlated with MW reports were discarded using correlation-based feature selection (CFS) [11]. Fourth, we calculated features using four different window lengths to ascertain the amount of gaze data needed to predict MW. Windows that were either 4, 8, or 12 seconds before each MW probe were considered. Fifth, we varied the minimum number of fixations that were required in each window before it was included in the data set. A lack of fixations could indicate gaze tracking problems, prolonged eye closure, off-screen gaze, etc. To account for this,

each window was first required to have at least 5 fixations. In addition, windows were required to have a minimum of 1, 2, or 3 fixations per second. Finally, the data was modified in five ways encompassing various combinations of outlier treatment and downsampling. The data was: unmodified (raw), trimmed, winsorized, trimmed and downsampled, or winsorized and downsampled. Outlier treatment was performed because outliers can cause model instability – especially for parametric models. Trimming consisted of removing values greater/lower than 3 standard deviations above/below the mean, while winsorization consisted of replacing those values with the corresponding value +3 or -3 standard deviations above/below the mean. Downsampling was also varied in data sets that were trimmed or winsorized as there was an uneven class distribution (“No” MW responses accounted for 70% of all responses), which can have adverse effects on classification accuracy. Instances from the most common MW response (i.e., “No” responses) were removed at random until there were an equal number of “Yes” and “No” responses in the training set. Importantly, downsampling was only applied to the training data. A leave-several-participant-out validation method was used to ensure that data from each participant was exclusive to either the training or testing set. Data from a random 66% of the participants were placed in the training set, while data from the remaining 34% were placed in the testing set. Feature selection was done on the training set after this split, to ensure independence from the testing set. This process was repeated 20 times for each model and classification accuracy was averaged across these iterations. The kappa metric [5] was used to evaluate model performance as it corrects for random guessing when there are uneven class distributions as with the current data. The kappa metric is calculated using the formula K = (Observed Accuracy - Expected Accuracy) / (1 - Expected Accuracy), where Observed Accuracy is equivalent to recognition rate and Expected Accuracy is computed from the confusion matrix to account for the pattern of misclassifications. Kappas of 0, 1, > 0, and < 0 indicate chance, perfect, above chance, and below chance agreement, respectively.

4

Results

The parameters and results for the best (highest kappa) models for either end-of-page and within-page MW reports are listed in Table 2. Both were Bayesian models (Naive Bayes for end-of-page and Bayes Net for within-page), which suggests that these models might be suitable for this type of data and classification task. Table 2. Results for best models. Standard deviations in parenthesis.

Type of Probe

Features

Data

Win. Size(s)

Fix /S

Kappa

Acc.

Exp. Acc.

End-ofpage

Global, Local, Context

Trimmed

8

0

.28 (.08)

72% (5%)

61% (5%)

Withinpage

Global with CFS

Winsorized

12

3

.17 (.08)

59% (5%)

50% (3%)

Table 3. Confusion matrices for best models Confusion matrices for each Actual Classified Prior model are shown in Table 3. Yes No From this table, we note an Yes .54 (hit) .46 (miss) .23 approximately 50% chance of End-ofNo .23 (FA) .77 (CR) .77 accurately classifying MW page responses (hits) versus incorYes .61 (hit) .39 (miss) .36 rectly classifying MW as nor- WithinNo .42 (FA) .58 (CR) .64 mal reading (misses) for the page end-of-page MW responses. Note. Values are proportionalized and averaged over iterations. The hits: misses ratio is higher FA = false alarm; CR = correct rejection for the within-page MW responses (roughly 60% hits vs. misses). Note, however, that in both cases hit rates greatly exceed the prior probabilities of “Yes” MW responses. Correct rejections (correctly classifying normal reading or CRs) and false-alarms (incorrectly classifying normal reading as MW or FAs) for both models were in line with prior probabilities of “No” MW responses, but we note a higher proportion of CRs to FAs for end-ofpage responses.

4.1

Parameter Comparison

Kappa Value

As certain operations such as Global 0.2 outlier treatment or downsampling increase complexity in realLocal time systems, we investigated the effect of each parameter on kapContext 0.1 pa values across the best performing classifier for each paGlobal + rameter configuration. The only Local + Context parameter with a clear trend was 0 feature type, which is shown in End-of-page Within-page Figure 2. Models built with just global features outperformed Figure 2. Effect of feature type on kappa value models built with just local features or models built with just context features. However, models built with a combination of features resulted in a small improvement over global features alone. 4.2

Feature Selection

The 10 top ranked features for the best models are shown in Table 4. It is interesting to note that the top four features for the end-of-page model were context features. Furthermore, maximum saccade length, mean saccade length, saccade length skew, and saccade length range were among the top features for both models, indicating that saccade length is an important indicator of MW.

Table 4. Top 10 ranked features for each model

Rank 1 2 3 4 5 6 7 8 9 10

5

End-Of-Page Previous Value Previous Difficulty Difficulty Value Maximum Saccade Length Saccade Length Range Page Number Saccade Length SD Mean Saccade Length Saccade Length Skew

Within-Page Maximum Saccade Length Median Saccade Length Fixation Duration/Saccade Duration Saccade Length Range Mean Saccade Length Saccade Length Skew Median Fixation Duration Mean Fixation Duration Mean Saccade Duration Minimum Saccade Duration

General Discussion

Mind wandering (MW) is a highly frequent phenomenon that has a significant negative impact on performance. This suggests that systems that support critical tasks such as learning, vigilance, and decision making should have some mechanism for tracking and responding to wandering minds. As an initial step in this direction, the purpose of this paper was to build a system capable of automatically detecting MW using eye gaze in a manner that generalizes to new users. Main Findings. A number of conclusions can be drawn from our results. First we have shown that it is possible to detect MW during reading by analyzing eye gaze features and aspects of the reading context. We were able to unobtrusively detect MW while reading both within a page and at the end of a page with accuracies of 59% and 72% respectively. Importantly, our user-independent validation method provides evidence that our models generalize to new users. Our work expanded on previous research by analyzing a richer eye gaze feature set that was complemented with an entire new set of context features which resulted in improved MW classification accuracy. Generalization was improved by obtaining participants from two universities with very different demographic characteristics. Second, classification rates were much higher for end-of-page MW reports compared to within-page MW reports. One possibility is that gaze patterns are distinct within a page vs. at the end of a page, thereby resulting in different classification accuracies for the same features. Alternatively, MW itself could adopt different forms when occurring within vs. at the end of a page. Analyzing this temporal aspect could yield to a deeper scientific understanding of MW itself, which would then be leveraged towards improving MW classification accuracy. Third, feature type had a profound effect on classification accuracy. Global features resulted in much higher classification accuracy when compared to local features, thereby suggesting that it might be more important to track overall gaze patterns (global features) rather than focusing on the specific words being read (local features). This is a significant finding because the global features are easier to compute and are more likely to generalize to different tasks beyond reading. Additionally, although

context features were among the highest ranked features in the end-of-page model, context-only models performed poorly, indicating that alone they are not sufficient for proper classification. Future research is needed to understand how a global MW model built in one task context (e.g., reading) can generalize to closely related (e.g., textdiagram integration) or unrelated (e.g., watching a film) contexts. Applications. The present findings are applicable to any user interface that contains a task involving reading comprehension of primarily textual information. MW reduces the ability to learn information from a text, so it has negative effects for any learning task involving text comprehension, which is the standard way to learn. In addition to reading, it is possible that gaze-based MW detection during a wider array of tasks and contexts could be attempted as well. Attentional state estimation has already been studied in a variety of areas, and any interface that would benefit from modeling attentional states would likely also benefit from modeling MW, which is an involuntary lapse in attention. Further research is needed to better understand the need and potential of automatic MW detection systems. Limitations. It is important to discuss a number of limitations with our study. First, the cost of high quality eye trackers limits the scalability of eye gaze as a MW detection modality. That being said, it is possible that this will be resolved in the short term due to the steadily decreasing cost of eye tracking technology such as Eye Tribe ($99) and Tobii EyeX ($195) eye trackers, or web-cam based eye tracking [21]. Second, although the participants were more diverse than in the two previous attempts at MW detection, the sample was still restricted to data from undergraduate students collected in a lab setting. Hence, quantifying performance on a more diverse population and in more diverse settings would boost some of our claims of generalizability. Third, although self-reports of MW have been validated in a number of studies [22, 23], they rely on participants to respond accurately and honestly. However, at the current moment there is no clear alternative for tracking MW. One alternative that we are exploring involves experimentally inducing MW and tracking associated gazepatterns. Finally, our highest accuracy of 72% can be considered to be moderate at best. However, this MW detection accuracy reflects a considerable improvement over chance, and is comparable with current user state detection systems when evaluated in a user-independent fashion [2–4]. Furthermore, MW is an internal, evasive, and noisy phenomenon, and generalizability was emphasized over accuracy by collecting naturalistic instances of MW and using a stringent validation method which guarantees independent training and testing sets. Accuracy would ostensibly be higher with models optimized for individual users. Future work. Future work could be focused in several areas. First, additional eye movement features could be considered. Possible features that were not included in the present study are eye blink features, pupillometry features, and temporal dynamics of gaze trajectories. Second, it is possible that easily collected individual differences such as a user’s predisposition for MW or baseline measures of reading behavior could be used to improve detection rates. Third, instead of using the “probe-caught” method used in this study, a more naturalistic “self-caught” method could be employed where participants are instructed to self-monitor for MW and report whenever they do so. Finally, possible interventions to restore attention when MW is detected

could be explored, including pausing the session, displaying the missed content in an alternate format, or even merely notifying the user that they were MW. Concluding Remarks. In summary, the present study demonstrated that gaze data coupled with contextual cues can be effective in automatically detecting MW during reading. Importantly, our approach was relatively unobtrusive (remote gaze trackers are no-contact sensors), allowed for unrestricted head and body movement, involved an ecologically-valid reading activity, and used a supervised classification method that is likely to generalize to new individuals. Our results indicate that future attempts to detect MW using eye gaze data should at the very least include global features, which can be easily computed with low-cost eye tracking, though this is an empirical question that awaits further research. Acknowledgment. This research was supported by the National Science Foundation (NSF) (ITR 0325428, HCC 0834847, DRL 1235958). Any opinions, findings and conclusions, or recommendations expressed in this paper are those of the authors and do not necessarily reflect the views of NSF.

References 1. Baayen, R.H. et al.: The CELEX Lexical Database, Release 2 [CD-ROM]. Philadelphia, PA: Linguistic Data Consortium, University of Pennsylvania. (1995). 2. Bednarik, R. et al.: What Do You Want to Do Next: A Novel Approach for Intent Prediction in Gaze-Based Interaction. In: Proceedings of the Symposium on Eye Tracking Research and Applications. pp. 83–90 ACM, New York, NY, USA (2012). 3. Biedert, R. et al.: A Robust Realtime Reading-Skimming Classifier. In: Proceedings of the Symposium on Eye Tracking Research and Applications. pp. 123–130 ACM, New York, NY, USA (2012). 4. Calvo, R.A., D’Mello, S.: Affect Detection: An Interdisciplinary Review of Models, Methods, and Their Applications. IEEE Transactions on Affective Computing. 1, 1, 18–37 (2010). 5. Cohen, J.: A Coefficient of Agreement for Nominal Scales. Educational and Psychological Measurement. 20, 1, 37–46 (1960). 6. D’Mello, S. et al.: Automatic Gaze-Based Detection of Mind Wandering During Reading. Educational Data Mining. (2013). 7. Domingos, P.: A Few Useful Things to Know About Machine Learning. Communications of the ACM. 55, 10, 78–87 (2012). 8. Drummond, J., Litman, D.: In the Zone: Towards Detecting Student Zoning Out Using Supervised Machine Learning. In: V. Aleven, et al. (eds.) Intelligent Tutoring Systems. LNCS, vol 6095, pp. 306-308. Springer-Verlag (2010). 9. Feng, S. et al.: Mind Wandering While Reading Easy and Difficult Texts. Psychonomic Bulletin & Review. 1–7 (2013). 10. Hall, M. et al.: The WEKA Data Mining Software: an Update. ACM SIGKDD Explorations Newsletter. 11, 1, 10–18 (2009). 11. Hall, M.A.: Correlation-Based Feature Selection for Discrete and Numeric Class Machine Learning. In: Proceedings of the Seventeenth International Conference on Machine Learning, pp. 359-366. (2000). 12. Killingsworth, M.A., Gilbert, D.T.: A Wandering Mind is an Unhappy Mind. Science. 330, 6006, 932–932 (2010).

13. Mooneyham, B.W., Schooler, J.W.: The Costs and Benefits of Mind-Wandering: A Review. Canadian Journal of Experimental Psychology/Revue Canadienne de Psychologie Expérimentale. 67, 1, 11 (2013). 14. Muir, M., Conati, C.: An Analysis of Attention to Student – Adaptive Hints in an Educational Game. In: Cerri, S.A. et al. (eds.) Intelligent Tutoring Systems. pp. 112–122 Springer Berlin Heidelberg (2012). 15. Navalpakkam, V. et al.: Attention and Selection in Online Choice Tasks. In: Masthoff, J. et al. (eds.) User Modeling, Adaptation, and Personalization. pp. 200–211 Springer Berlin Heidelberg (2012). 16. Rayner, K.: Eye Movements in Reading and Information Processing: 20 Years of Research. Psychological Bulletin. 124, 3, 372 (1998). 17. Reichle, E.D. et al.: Eye Movements During Mindless Reading. Psychological Science. 21, 9, 1300–1310 (2010). 18. Schooler, J.W. et al.: Meta-Awareness, Perceptual Decoupling and the Wandering Mind. Trends in Cognitive Sciences. 15, 7, 319–326 (2011). 19. Schooler, J.W. et al.: Zoning Out While Reading: Evidence for Dissociations Between Experience and Metaconsciousness. Thinking and Seeing: Visual Metacognition in Adults and Children. 203–226 (2004). 20. Seibert, P.S., Ellis, H.C.: Irrelevant Thoughts, Emotional Mood States, and Cognitive Task Performance. Memory & Cognition. 19, 5, 507–513 (1991). 21. Sewell, W., Komogortsev, O.: Real-Time Eye Gaze Tracking with an Unmodified Commodity Webcam Employing a Neural Network. In: CHI’10 Extended Abstracts on Human Factors in Computing Systems, pp. 3729-3744. ACM, Atlanta, Georgia, USA (2010). 22. Smallwood, J. et al.: Subjective Experience and the Attentional Lapse: Task Engagement and Disengagement During Sustained Attention. Consciousness and Cognition. 13, 4, 657– 690 (2004). 23. Smallwood, J. et al.: When Attention Matters: The Curious Incident of the Wandering Mind. Memory & Cognition. 36, 6, 1144–1150 (2008). 24. Smilek, D. et al.: Out of Mind, Out of Sight Eye Blinking as Indicator and Embodiment of Mind Wandering. Psychological Science. 21, 6, 786–789 (2010). 25. Voßkühler, A. et al.: OGAMA (Open Gaze and Mouse Analyzer): Open-Source Software Designed to Analyze Eye and Mouse Movements in Slideshow Study Designs. Behavior Research Methods. 40, 4, 1150–1162 (2008). 26. Yonetani, R. et al.: Multi-mode Saliency Dynamics Model for Analyzing Gaze and Attention. In: Proceedings of the Symposium on Eye Tracking Research and Applications. pp. 115–122 ACM, New York, NY, USA (2012).

Toward Fully Automated Person-Independent Detection ...

instructed they would be required to read additional material if they did not do ..... Killingsworth, M.A., Gilbert, D.T.: A Wandering Mind is an Unhappy Mind. Science. 330, ... Canadian Journal of Experimental Psychology/Revue Canadienne de ...

223KB Sizes 3 Downloads 354 Views

Recommend Documents

FULLY AUTOMATED NON-NATIVE SPEECH ...
grating corresponding L1 phone models as alternative HMM paths. We introduce graphemic contraints in the confusion extraction pro- cess: the phonetic ...

FULLY AUTOMATED NON-NATIVE SPEECH ...
ABSTRACT. This paper presents a fully automated approach for the recognition of non-native speech based on acoustic model modification. For a na-.

Automated Down Syndrome Detection Using ... - Semantic Scholar
these anatomical landmarks and texture features based on LBP ... built using PCA describing shape variations in the training data. After PCA, we can represent ...

AUTOMATED SUPER-RESOLUTION DETECTION OF ...
call for flexible detection approaches such as deformable mod- els, rod-shaped .... of the segment center, θ is its angle with the x-axis, L = 2l the segment length, and ..... We plan to pursue this work along several lines, including extension to 3

Automated Down Syndrome Detection Using ... - Semantic Scholar
*This project was supported by a philanthropic gift from the Government of Abu Dhabi to Children's National Medical Center. Its contents are solely the responsibility of the authors and ..... local and global facial textures. The uniform LBP, origina

A fully automated method for quantifying and localizing ...
aDepartment of Electrical and Computer Engineering, University of Pittsburgh, .... dencies on a training set. ... In the current study, we present an alternative auto-.

Towards fully automated axiom extraction for finite ...
Key words: Many-valued logics, tableaux, automated theorem proving. 1 Introduction. This note will report on the first developments towards the implementation.

Development of a fully automated system for delivering ... - Springer Link
Development of a fully automated system for delivering odors in an MRI environment. ISABEL CUEVAS, BENOÎT GÉRARD, PAULA PLAZA, ELODIE LERENS, ...

Towards fully automated axiom extraction for finite ...
This note will report on the first developments towards the implementation of a fully automated system for the extraction of adequate proof-theoretical.

Fully Automated Non-Native Speech Recognition Using ...
tion for the first system (in terms of spoken language phones) and a time-aligned .... ing speech recognition in mobile, open and noisy environments. Actually, the ...

A fully automated method for quantifying and localizing ...
machine learning algorithms including artificial neural networks (Pachai et al., 1998) .... attenuated inversion recovery (fast FLAIR) (TR/TE= 9002/56 ms Ef; TI=2200 ms, ... imaging data to predefined CHS visual standards and representative of ...

VAMO: Towards a Fully Automated Malware ... - Semantic Scholar
[10] A. K. Jain, M. N. Murty, and P. J. Flynn. Data clustering: a review. ACM Comput. Surv., 31(3):264–323, 1999. [11] J. Jang, D. Brumley, and S. Venkataraman.

VAMO: Towards a Fully Automated Malware ... - Semantic Scholar
Dept. of Computer Science. University of Georgia. Athens .... 11, 15, 18]) on M to partition it in a number of malware clusters, (b) use VAMO to build a reference.

Automated Detection of Engagement using Video-Based Estimation of ...
Abstract—We explored how computer vision techniques can be used to detect ... supervised learning for detection of concurrent and retrospective self-reported engagement. ...... [49] P. Ekman and W. V. Friesen, Facial Action Coding System: A ... [On

Automated Detection of Sensor Detachments for ...
module on an Android mobile smartphone using the Au-. toSense body-area sensor network and the mStress mobile inferencing framework [2]. AutoSense [3] is ...

Real-time automated 3D sensing, detection, and ...
May 1, 2006 - integral imaging for 3D sensing, visualization, and recognition of biological ..... techniques based on the edge map may fail to segment these images ... calculating statistical parameters of the microorganisms, the data can be ...

Automated Detection of Engagement using Video-Based Estimation of ...
Abstract—We explored how computer vision techniques can be used to detect engagement while ... supervised learning for detection of concurrent and retrospective self-reported engagement. ...... [Online]. Available: http://msdn.microsoft.com/en-us/l

Real-time automated 3D sensing, detection, and ...
May 1, 2006 - optical imaging techniques for real-time automated sensing, visualization, ...... 32], however more advanced methods such as bivariate region snake in Section 3 can be applied. ..... used for illustration in the figures hereafter.

Towards Automated Detection and Regulation of ...
approach to the writing process developed by Rijlaarsdam and Bergh [3, 4]. Re- searchers have also proposed some automated systems to help students ...

Automated Detection of Stable Fracture Points in ...
3. FRACTURE POINT DETECTION IN. INDIVIDUAL 2D SLICES. We used simple but useful concepts from curvature scale-space theory and graph theory to detect .... vp. (9). We further assume that the estimation/prediction of the velocity or change in position

Automated Physiological-Based Detection of Mind ...
6. Andreassi, J.L.: Psychophysiology: Human behavior and physiological response. Rout- ledge (2000). 7. Smallwood, J., Davies, J.B., Heim, D., Finnigan, F., ...