Next Article in Journal
Executive Profile of the Logopenic Variant of Primary Progressive Aphasia: Comparison with the Semantic and Non-Fluent Variants and Alzheimer’s Disease
Next Article in Special Issue
Dyslexia: Causes and Concomitant Impairments
Previous Article in Journal
Application of Proteomics Analysis and Animal Models in Optic Nerve Injury Diseases
Previous Article in Special Issue
Revisiting Multifactor Models of Dyslexia: Do They Fit Empirical Data and What Are Their Implications for Intervention?
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Accessible Dyslexia Detection with Real-Time Reading Feedback through Robust Interpretable Eye-Tracking Features

1
School of Electrical Engineering, University of Belgrade, Bulevar Kralja Aleksandra 73, 11120 Belgrade, Serbia
2
Innovation Center, School of Electrical Engineering, Bulevar Kralja Aleksandra 73, 11120 Belgrade, Serbia
3
Faculty of Technical Sciences, University Singidunum, Danijelova 32, 11000 Belgrade, Serbia
4
Faculty of Philosophy, University of Belgrade, Čika-Ljubina 18-20, 11000 Belgrade, Serbia
*
Authors to whom correspondence should be addressed.
Brain Sci. 2023, 13(3), 405; https://doi.org/10.3390/brainsci13030405
Submission received: 19 January 2023 / Revised: 8 February 2023 / Accepted: 24 February 2023 / Published: 26 February 2023
(This article belongs to the Special Issue Developmental Dyslexia: Theories and Experimental Approaches)

Abstract

:
Developing reliable, quantifiable, and accessible metrics for dyslexia diagnosis and tracking represents an important goal, considering the widespread nature of dyslexia and its negative impact on education and quality of life. In this study, we observe eye-tracking data from 15 dyslexic and 15 neurotypical Serbian school-age children who read text segments presented on different color configurations. Two new eye-tracking features were introduced that quantify the amount of spatial complexity of the subject’s gaze through time and inherently provide information regarding the locations in the text in which the subject struggled the most. The features were extracted from the raw eye-tracking data (x, y coordinates), from the original data gathered at 60 Hz, and from the downsampled data at 30 Hz, examining the compatibility of features with low-cost or custom-made eye-trackers. The features were used as inputs to machine learning algorithms, and the best-obtained accuracy was 88.9% for 60 Hz and 87.8% for 30 Hz. The features were also used to analyze the influence of background/overlay color on the quality of reading, and it was shown that the introduced features separate the dyslexic and control groups regardless of the background/overlay color. The colors can, however, influence each subject differently, which implies that an individualistic approach would be necessary to obtain the best therapeutic results. The performed study shows promise in dyslexia detection and evaluation, as the proposed features can be implemented in real time as feedback during reading and show effectiveness at detecting dyslexia with data obtained using a lower sampling rate.

1. Introduction

The definition of dyslexia has been a topic in many research studies [1,2,3], but regardless of the definition, its detrimental effects on academic performance [4,5] and quality of life [6,7,8] have been consistently proven. Multiple studies have the goal of improving the understanding of dyslexia from multiple different aspects, including neural activity [9,10], visual sampling strategy [11,12,13], emotional response [14,15], educational performance [16,17], social status [18,19], and others [20,21,22,23]. However, considering the nature of experimental and study requirements, implementations of the study results for assistive and treatment purposes often prove to be a challenge in itself [24,25,26].
Although different aspects of dyslexia are studied in the literature, in practice, the diagnosis of dyslexia is commonly based on standardized oral or written tests conducted by specialists. These tests take into consideration background information, intelligence, oral and reading language skills, as well as other relevant tests [27]. With the advancement of technologies, however, different diagnostical procedures are being proposed in the literature, focusing on different measurable and quantifiable criteria based on biomarker approaches [28,29].
The visual sampling strategy of dyslexia (i.e., gaze patterns) is often analyzed in dyslexia research for the purposes of dyslexia diagnosis and evaluation. The irregularity in gaze patterns of dyslexic subjects has long been analyzed in studies [30,31], but technological advancement has made dyslexia eye-tracking studies more common and the experimental setup scenarios more diverse. In Rello et al. [32], the authors have reportedly performed a first study focusing on dyslexia classification based on the eye event (fixations and saccades) parameters during reading. A Tobii 1750 eye tracker was used to measure the eye movements of 98 subjects (48 dyslexic), and a support vector machine (SVM) was used to differentiate dyslexic and control subjects based on derived features and participant characteristics. This study included native Spanish speakers and reported an accuracy of 80.18% and has paved the way for many future similar studies. In [33], the authors monitor the eye movements of 185 subjects (97 dyslexic) using a goggle-based Ober-2 eye-tracker. The study included participants with ages ranging from 9 to 10, presented them a text on a piece of paper, and derived 168 eye-tracking features from the recorded data. They implemented recursive feature elimination (RFE) to reduce the number of features and trained an SVM classifier to detect dyslexic reading patterns. The results were an accuracy of 95.6%, showing the methodology as quite effective. Prabha et al. in [34] also analyzed the dataset obtained in [33], focusing on only the features extracted from fixations. The authors implanted an RFE feature selection algorithm and several different machine learning (ML) algorithms, including SVM, k-nearest neighbors (KNN), and random forest (RF). They obtained an accuracy of 95% and continued their work through several other publications, analyzing the same dataset [35,36,37]. They tried different ML algorithms and eye-tracking feature inputs, obtaining a slightly higher accuracy of 96% in [36], and providing a method for clustering the dyslexic group into high and low dyslexics, providing reference ranges for fixation and saccade duration to estimate the severity of dyslexia [37].
In [38], a study including 69 children (32 dyslexic), ages 8.5–12.5, was performed. The study included native Greek speakers, and the participants were given two paragraphs of text to read while their eye movements were recorded with a custom eye tracker. The study employed several ML algorithms and a multitude of features, obtaining an accuracy of 97% with only three features: saccade length, the number of short forward movements, and the number of repeatedly fixated words. In [39], the authors used an EyeLink 1000 eye tracker to monitor the eye movements of 165 (30 dyslexic) Finnish students with an average age of 12.5. The authors implemented two ML algorithms, SVM and RF, and obtained an accuracy of 89.7% using a novel, transition-matrix-based feature extraction. Observing the students during both reading and non-reading tasks, El Hmimdi et al. analyzed 87 subjects (46 dyslexic) recruited from schools in Paris [40]. Two non-reading and one reading test were performed, and the eye movements were monitored using a Pupil Core head-mounted video-oculography device. Based on a variety of eye-tracking features and ML algorithms, the authors report an accuracy of 81.25% from the reading tests and 81.25% and 77.3% from the two non-reading tests. A study including 30 Serbian children (15 dyslexic) was performed in [41], focusing on the spatio-temporal characteristics of gaze during reading. The experiment included the reading of 13 text segments, each presented on a different background/overlay color, and used a SMI RED-m portable remote eye tracker to monitor the participants’ eye movements. The study implemented several ML algorithms using various features individually as well as feature groups as inputs and obtained an accuracy of 94% for a combination of features, with an accuracy of 91% for a single feature describing the spatial complexity of gaze.
Recent studies also report attempts at detecting dyslexia from eye tracking during reading without performing feature extraction rather than working with minimally processed data using convolutional neural networks (CNNs). In [42], the authors used the dataset from [33] and implemented a one-dimensional neural network that performed dyslexia detection on the sequences of x and y gaze coordinates. The study reports an accuracy of 96.6%, showing that a more holistic approach can indeed be used to detect dyslexic tendencies. The authors in [43] also perform dyslexia classification using a CNN, but use 2D images as inputs instead. The study was performed using data gathered in [44] and used color coding to represent gaze data in the form of colored images. These images were then used to train and evaluate a deep two-dimensional CNN, obtaining an accuracy of 87% and proving that the spatial characteristics of gaze carry enough information for deep learning algorithms to provide accurate dyslexia detection. A method that also used CNN and the spatial characteristics of gaze for dyslexia detection was presented in [45]. The method does include preprocessing steps and feature extraction, but it does not use eye-event parsing, and it was implemented and tested on two datasets (trained on one and tested on the other) from [44] and [33], achieving accuracies of 85.6% and 82.9%, respectively.
The abundance of eye-tracking studies in dyslexia points to the fact that a dyslexia diagnosis can indeed be successfully made based on eye tracking during reading. However, implementation of the scientific results in terms of eye tracking during reading, in practice, is not abundantly present. There are tools reported in the scientific literature that focus on the digitalization of dyslexia diagnosis [46,47], some of which do use eye tracking [38], but considering the variety of study conditions including eye trackers used, text presentation, native language, participant age, and many others, this makes practical applications somewhat complicated and difficult. Providing feedback during reading in terms of exact points in the text that caused struggles would also be quite difficult, considering that few eye-tracking processing techniques can pinpoint the locations in the text that were problematic to read.
Finally, examining the impact of colors on reading performance is challenging in research as well as in practice [48]. The use of colored overlays, lenses, plastic sheets, and visual changes in the presentation of text, such as background color and font size, can relieve dyslexia symptoms [49,50]. Some studies have suggested that certain colors (like peach, orange, yellow, turquoise, and so on) may be more effective than others in reducing visual stress and improving reading performance in dyslexics [44,51]. Therefore, the effect of colors should certainly be taken into consideration when designing studies dealing with the analysis of reading difficulties.
The goal of the study was the creation of eye-tracking features that could differentiate between dyslexic and control subjects, work on low-frequency eye trackers, and be independent of eye movement parsing (fixations and saccades). The proposed features inherently provide locations in the text where the struggle during reading was detected, making real-time feedback possible. Furthermore, the study performs an analysis of the influence of color configuration (CC) on the performance of subjects, showing the importance of the presentation of the text and its possible impact on the dyslexia diagnosis and monitoring success.

2. Materials and Methods

2.1. Dataset

The data were acquired in a previous study [44] and included an experiment where Serbian children were instructed to read 13 text segments, each presented on a computer screen with different CC. The data analyzed in this study were gathered from 30 subjects (15 dyslexic), with ages ranging from 7 to 13, and the distribution of age is the same for both the dyslexic and control groups. Each text segment (2–3 sentences) is presented on a different CC, with the first always being black text on a white background and the others having a pseudo-random order, including the colors red, blue, yellow, orange, turquoise, and purple as both backgrounds for black text, and a transparent overlay over black text on a white background. The analyzed data included 378 trials, with a trial representing the reading of one text segment from one participant. The 378 trials were obtained from a total of 30   subjects × 13   CC = 390 trials, after excluding 12 trials owing to a short reading time (reading time less than 5 s), exhibiting insufficient focus on the text. The subjects were monitored during reading using a multimodal sensor hub [20], but only the eye-tracking data were used in this study. An SMI RED-m portable remote eye-tracker (iMotions, Copenhagen, Denmark) was used to record the eye movements and the data were acquired at a sampling frequency of 60 Hz.

2.2. Feature Extraction

The feature extraction process is designed to rely only on the sequence of x and y coordinates of gaze, without the information on eye event occurrence, i.e., fixations and saccade events. Firstly, the data points that had no information on the gaze coordinates for both eyes were removed, leaving only the data where the gaze could be located on the screen (the average percentage of data per trial that contained no gaze coordinates was 4.5% of total trial data for the dyslexic group and 2.5% for the neurotypical group). Secondly, two different methods for feature extraction, with the same goal but different characteristics, were implemented to detect the subjects’ struggle during reading (feature 1 and feature 2). The feature extraction was performed only on valid gaze data.
Feature 1 was based on the detection of the self-intersection (SI) of gaze lines during reading. Every four consecutive data points in the x–y coordinate plane are observed, and it was determined whether the gaze line connecting these points has an SI. If there is an intersection, the number of changes in the direction of gaze movement on the y-axis is monitored for the following 250 ms (approximately an average fixation duration [52]). Each change is counted, and a vertical alteration score (VAS)—indicating the number of changes in the direction of movement on the y-axis—is added to the total VAS of feature 1. So, feature 1 starts off at the value of 0 at the beginning of the trial, and each time an SI is detected in four consecutive data points, the feature can increase based on the VAS in the time period that follows the SI event. This effectively causes the feature value to progressively rise during the trial, but uses the SI event as a form of a trigger, and the VAS of the included time intervals as the quantification of struggle during reading.
Feature 2 follows the same principle as feature 1, but instead of using an SI of four consecutive data points as a trigger to add the VAS, it uses the number of changes in the gaze direction on the y-axis of four consecutive data points (observing the VAS in four consecutive data points). If there are two changes in the direction of movement on the y-axis in the four observed consecutive data points, the VAS of the gaze obtained from the 250 ms following the trigger event is added to the total VAS of feature 2.
An example of feature 1 and feature 2 extracted for the reading of one line of text for a dyslexic subject is given in Figure 1.
Figure 1 clearly shows a much higher number of trigger events in feature 2, as the change in the y direction is much more common than SI in four consecutive points. Consequentially, feature 2 will have much higher values than feature 1, detecting more events and thus having more opportunity to increase based on the VAS in the observed time interval of 250 ms. An example of the extraction of feature 1 and feature 2 for one line of text for a control subject is given in Figure 2.
In contrast to the dyslexic subject, the control subject has a lower number of trigger events and overall lower feature values. The VAS is not close to 0, especially when the trigger event is based on the VAS detection in four consecutive data points, because vertical changes in gaze movement are not indicative of a dyslexic reading pattern per se, but the number of changes after trigger events is expected to be lower in controls because dyslexic subjects do have a more spatially complex gaze pattern [41,45].
To investigate the stability of the proposed features and whether they could be used with data acquired at a lower sampling rate, the feature extraction process was repeated on gaze data downsampled from 60 Hz to 30 Hz. This would result in four total features observed in the paper: feature 1 on 60 Hz, feature 1 on 30 Hz, feature 2 on 60 Hz, and feature 2 on 30 Hz. Each of these was analyzed independently as inputs to various machine learning algorithms and through statistical analysis.
Additionally, the total reading time of each trial was used as input to the ML algorithms as a baseline feature for dyslexia classification.

2.3. Machine Learning and Statistical Analysis

Several different ML algorithms were implemented in this study, and each of them had only a single feature at a single sampling frequency as an input. Leave-one-subject-out evaluation was implemented to obtain results that would be similar those in practice, taking all the data from one subject and placing it in the test set and using the rest of the data as training. This was repeated 30 times (each subject was placed in the test set), and the predictions and labels of all the test sets were concatenated and used to provide classification metrics. The algorithms used were logistic regression (LR), SVM, KNN, and RF, and the metrics used were accuracy (ACC), sensitivity (Se), specificity (Sp), F1-score, and the area under the receiver operating curve (AUROC). For each training fold, a grid search for hyperparameter optimization was performed on the training set, using fivefold cross-validation.
The statistical analysis of the data includes a comparison between the dyslexic and control group on all collective CC, and for each CC separately, using the Mann–Whitney statistical test. An analysis of the impact of each CC on the dyslexic group was performed by comparing each pair of the CC within the dyslexic group, also using the Mann–Whitney test. The effect of the CC on the separability of classes was also analyzed by displaying the boxplots for each feature for the dyslexic and control groups when each of the individuals was evaluated on their best-performing CC and on their worst-performing CC.
The processing, ML algorithm implementation, statistical analysis, and visualization of the data were performed in the Python programming language, using the sklearn and matplotlib libraries [53,54,55].

3. Results

3.1. Classification Results

The results of the performed cross-validation for each feature and for all observed ML algorithms are shown in Table 1.
The evaluation results show several important findings. Firstly, regardless of the ML algorithm and the proposed features used as the input, the classification accuracy is above 82%, indicating that the features are overall effective at separating the dyslexic and control groups. The results also show an improved performance when compared with the considered baseline (total reading time). Secondly, the best results, in terms of ACC, are obtained for the VAS event detection feature at 60 Hz and for the LR and SVM, being 88.9%. Finally, the accuracies obtained for the features extracted at 30 Hz are not much lower than the values obtained for features extracted at 60 Hz, with the best result being an ACC of 87.8% for the VAS event detection feature and the LR algorithm. This is also equal to the best-obtained accuracy for the SI event detection feature, giving a slight advantage to the VAS detection in the classification performance.

3.2. Color Configuration—Reading Performance Results

The results of the statistical analysis are congruent with the overall classification results. There is a statistically significant difference (p < 0.001) between the dyslexic and control group when all CCs are observed together for each of the two features at both frequencies. When comparing the dyslexic and control group on each CC separately, there is a statistically significant difference (p < 0.001) for all CCs.
Comparing the CCs within the dyslexic group suggested no statistically significant difference. This, however, does not mean that the influence of color is negligible, just that there is no single CC that is universally better than the others. The influence of CC can be seen in Figure 3 for the SI event detection and in Figure 4 for VAS event detection features, at both 30 and 60 Hz, when the dyslexic and control groups are evaluated on each individual subject’s best-/worst-performing CC.

3.3. Computational Resources

The feature extraction was executed on an Intel(R) Core(TM) i3-7100 CPU @ 3.90 GHz processor with 16 GB of RAM. The quotient of the total time needed for feature extraction and the total reading time of all subjects is 7.21‰ for the SI event feature and 7.33‰ for the VAS event feature at 60 Hz. Effectively, this means that, for 1 s of reading time, ≈7 ms of processing time is needed, making real-time processing and feedback possible. The processing time is halved for feature extractions performed for data downsampled to 30 Hz.

4. Discussion

When observing the obtained classification ACC, it is not above 90%, which is reported in several studies, e.g., [38,42], but it definitely falls in the range of ACCs (80% to 96.6% [32,33,38,40,41,42]) that are common for dyslexia detection based on eye tracking during reading. The results of the authors’ previous study [41] showed an ACC of 94% obtained when using a combination of features. However, when observing a single feature, the best ACC was 91% (which is 2% higher than the results obtained in this study). Using a single feature is an important goal in the context of accessible dyslexia detection and monitoring, because providing simple feedback and a quantitative measure of dyslexic tendencies is easier on a single-feature scale. Furthermore, the features proposed in [41] introduce a concept of the spatial characteristics of the gaze, but rely on the parsing of the gaze into fixations and saccades, which is common in the literature. One more important aspect of the proposed features is that they have trigger events and continual scoring, making real-time feedback possible, and the locations at which the feature gained most of its value (struggles in reading) are traceable.
The features proposed in this paper rely on spatial characteristics, which are shown to be indicative of dyslexic behavior in different languages [45]. Feature 1, relying on the detection of self-intersections in gaze for the trigger events, was introduced as a measure that would be quite restrictive to trigger events, because it does require a certain spatial resolution of the eye-tracking device in order to detect the events. On the other hand, feature 2 was introduced, focusing only on the detection of the changes of movement direction in on the y (vertical) axis. This feature is a lot less restrictive in comparison with the first one, and the changes in the direction of movement can often occur in both dyslexic and control subjects. Having this in mind, a large number of these alterations could be considered a more complex spatial pattern, and the goal was to analyze how these two features would compare in dyslexia detection.
Both features proposed in this paper were shown to be quite effective in separating the dyslexic and control group, and have been shown to be effective (best ACC of 87.8%) when implemented at a sampling frequency of 30 Hz. There is no clear prediction in terms of the impact a lower sampling rate can have on classification in the reported studies, but it is important to note that the sampling frequencies of the reported studies are all above 30 Hz: 50 Hz in [32], 60 Hz in [38], 100 Hz in [33], 200 Hz in [40], and 1000 Hz in [39]. Considering the advancements in computer vision and the possibilities for eye tracking based on web cameras [56,57,58], the results of this paper show promise in accessible dyslexia detection.
The features are designed in such a way that each detected trigger event has a quantifiable contribution to the total VAS score. A higher total VAS score indicates a reader with more difficulties, so, consequently, the trigger events (which can be located in the text) with the higher contribution to the total VAS score can be considered as points of struggle or at least the points in the text that have made the quantification of dyslexic tendency higher. The proposed features show the potential for identifying text locations where reading difficulties are visually manifested with or without compensatory strategies involved. More experimental studies are needed to differentiate between problems in reading and various compensatory strategies [59], which goes beyond the scope of this paper.
Finally, when considering accessible dyslexia detection, the design of screening can be relevant. The results of Figure 3 show the influence of CC selection on the separability of the dyslexic and control group for the SI event feature. The groups have barely any intersections in feature values when they are maximally separated (control subjects evaluated on their best CC and dyslexic subject evaluated on their worst CC), while they are much closer in values when they are brought together (dyslexic subjects evaluated on their best CC and control subject evaluated on their worst CC). These scenarios, of course, represent extreme cases. However, they emphasize the influence of the way text is presented. The pie charts also show a variety of CCs, indicating that no color can be singled out regarding reading facilitation and that the distribution of CCs can vary depending on the sampling frequency.
This is further backed up by the influence of CC shown for the VAS event detection features; Figure 4. One thing that stands out is that, although the VAS event detection feature and the SI event detection feature are conceptually quite similar, they show some differences in the best/worst CC for the dyslexic and control groups. For example, the white background is quite prominent for both features as the best CC for control subjects at 60 Hz. On the other hand, at 30 Hz, white is present as the best CC for controls for the VAS event feature, but is barely present in the distribution for the best CCs of controls for the SI event feature. This would emphasize the importance of repeated evaluations of each CC, as well as using multiple features, as they might represent different reading patterns. Although both features separate the dyslexic and control groups effectively, they might not be completely congruent when it comes to the best way of presenting text.
Figure 3 and Figure 4 visually demonstrate that the color configuration has an impact on the feature values. The appropriate selection of the color configuration could increase or decrease feature values; that is, depending on the choice of colors, dyslexics can be closer or further away from neurotypicals (Figure 3 and Figure 4 present only extreme cases of the color configuration selection). This conclusion about the individual’s affective reaction to colors could be useful not so much for the classification procedure itself, but for the planning of the treatment for dyslexics using color text modifications. Additionally, pie charts in Figure 3 and Figure 4 demonstrate the importance of a personalized approach in the color configuration selection.
The results of this study show that there is no universally facilitating or aggravating CC for presenting text, neither for the control group nor the dyslexic group. However, when given the opportunity to read on their best-/worst-performing CCs, the separability between classes can be influenced. The conclusions of this study stand, therefore, with those from the literature [41,60] in terms of the importance of the text presentation method and emphasize the importance of the evaluation of subjects using various scenarios in order to obtain an accurate diagnosis.

5. Conclusions

In this paper, two novel eye-tracking features were introduced that were used to differentiate between the dyslexic and control groups. The features were designed with interpretability and robustness in mind and were implemented only based on the x and y coordinates of gaze and extracted from the original data (60 Hz) as well as from downsampled data (30 Hz). The features were used as inputs to various ML algorithms for dyslexia classification, and an accuracy of 88.9% for 60 Hz and 87.8% for 30 Hz was obtained. Statistical analysis was performed as well, showing that the features were effective at separating the groups on each CC, but that no CC is universally superior to others. The features’ interpretability, computational efficiency, and robustness regarding sampling frequency give them great potential for accessible dyslexia screening, especially as they focus on the spatial characteristics of gaze, which have been shown to possibly be indicative of dyslexic reading patterns in different languages. Future work would include the implementation of features on a larger dataset, possibly a different experiment scenario involving the reading of a longer text including visual crowding effects, and an analysis of eye movements of different languages, as well as use of the developed features with custom-made eye trackers based on web cameras.

Author Contributions

Conceptualization, I.V., V.K., T.P., A.M.S. and M.M.J.; methodology, I.V., V.K., A.M.S. and M.M.J.; software, I.V.; validation, A.M.S. and M.M.J.; formal analysis, I.V., V.K. and M.M.J.; investigation, I.V., V.K., T.P., A.M.S. and M.M.J.; resources, V.K., T.P. and M.M.J.; data curation, I.V., V.K., T.P., A.M.S. and M.M.J.; writing—original draft, I.V., V.K. and M.M.J.; writing—review and editing, I.V., V.K., T.P., A.M.S. and M.M.J.; visualization, I.V.; supervision, M.M.J.; project administration, M.M.J.; funding acquisition, M.M.J. and T.P. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the Ministry of Education, Science and Technology Development of Serbia, Belgrade, Serbia (contracts 451-03-68/2022-14/200103 and 451-03-68/2022-14/200223).

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki and approved by the ethical committee of the Psychology Department of the University of Niš (a branch of the Serbian Psychology Association), experimental procedure No. 9/2019 (04.09.2019).

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

No new data were created in this study.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Lyon, G.R.; Shaywitz, S.E.; Shaywitz, B.A. A Definition of Dyslexia. Ann. Dyslexia 2003, 53, 1–14. [Google Scholar] [CrossRef]
  2. Peterson, R.L.; Pennington, B.F. Developmental Dyslexia. Lancet 2012, 379, 1997–2007. [Google Scholar] [CrossRef] [Green Version]
  3. Tunmer, W.; Greaney, K. Defining Dyslexia. J. Learn. Disabil. 2009, 43, 229–243. [Google Scholar] [CrossRef]
  4. Singer, E. Coping with Academic Failure, a Study of Dutch Children with Dyslexia. Dyslexia 2008, 14, 314–333. [Google Scholar] [CrossRef]
  5. Stagg, S.D.; Eaton, E.; Sjoblom, A.M. Self-Efficacy in Undergraduate Students with Dyslexia: A Mixed Methods Investigation. Br. J. Spec. Educ. 2018, 45, 26–42. [Google Scholar] [CrossRef]
  6. Glazzard, J. The Impact of Dyslexia on Pupils’ Self-Esteem. Support Learn. 2010, 25, 63–69. [Google Scholar] [CrossRef]
  7. Zou, L.; Zhu, K.; Jiang, Q.; Xiao, P.; Wu, X.; Zhu, B.; Song, R. Quality of Life in Chinese Children with Developmental Dyslexia: A Cross-Sectional Study. BMJ Open 2022, 12, e052278. [Google Scholar] [CrossRef]
  8. Huang, Y.; He, M.; Li, A.; Lin, Y.; Zhang, X.; Wu, K. Personality, Behavior Characteristics, and Life Quality Impact of Children with Dyslexia. Int. J. Environ. Res. Public Health 2020, 17, 1415. [Google Scholar] [CrossRef] [Green Version]
  9. Weiss, B.; Nárai, Á.; Vidnyánszky, Z. Lateralization of Early Orthographic Processing during Natural Reading Is Impaired in Developmental Dyslexia. Neuroimage 2022, 258, 119383. [Google Scholar] [CrossRef]
  10. García Chimeno, Y.; García Zapirain, B.; Saralegui Prieto, I.; Fernandez-Ruanova, B. Automatic Classification of Dyslexic Children by Applying Machine Learning to FMRI Images. Biomed. Mater. Eng. 2014, 24, 2995–3002. [Google Scholar] [CrossRef] [Green Version]
  11. Prado, C.; Dubois, M.; Valdois, S. The Eye Movements of Dyslexic Children during Reading and Visual Search: Impact of the Visual Attention Span. Vision Res. 2007, 47, 2521–2530. [Google Scholar] [CrossRef] [Green Version]
  12. Weiss, B.; Szalma, J.; Vidnyánszky, Z. Data-Driven Detection of Developmental Dyslexia: A Machine Learning Approach Based on Behavioral and Eye-Movement Features. PsyArXiv 2022. [Google Scholar] [CrossRef]
  13. Temelturk, R.D.; Ozer, E. Binocular Coordination of Children with Dyslexia and Typically Developing Children in Linguistic and Non-Linguistic Tasks: Evidence from Eye Movements. Ann. Dyslexia 2022, 72, 426–444. [Google Scholar] [CrossRef]
  14. Alexander-Passe, N. Dyslexia and Depression: The Hidden Sorrow: An Investigation of Cause and Effect; Novinka/Nova Science Publishers: Hauppauge, NY, USA, 2012; pp. vi, 349. [Google Scholar]
  15. Nelson, J.M.; Gregg, N. Depression and Anxiety Among Transitioning Adolescents and College Students With ADHD, Dyslexia, or Comorbid ADHD/Dyslexia. J. Atten. Disord. 2010, 16, 244–254. [Google Scholar] [CrossRef] [Green Version]
  16. Mortimore, T.; Crozier, W.R. Dyslexia and Difficulties with Study Skills in Higher Education. Stud. High. Educ. 2006, 31, 235–251. [Google Scholar] [CrossRef]
  17. Prevett, P.; Bell, S.; Ralph, S. Dyslexia and Education in the 21st Century. J. Res. Spec. Educ. Needs 2013, 13, 1–6. [Google Scholar] [CrossRef]
  18. Macdonald, S.J. Windows of Reflection: Conceptualizing Dyslexia Using the Social Model of Disability. Dyslexia 2009, 15, 347–362. [Google Scholar] [CrossRef]
  19. Macdonald, S.J.; Deacon, L. Twice upon a Time: Examining the Effect Socio-Economic Status Has on the Experience of Dyslexia in the United Kingdom. Dyslexia 2019, 25, 3–19. [Google Scholar] [CrossRef] [Green Version]
  20. Jakovljević, T.; Janković, M.M.; Savić, A.M.; Soldatović, I.; Todorović, P.; Jere Jakulin, T.; Papa, G.; Ković, V. The Sensor Hub for Detecting the Developmental Characteristics in Reading in Children on a White vs. Colored Background/Colored Overlays. Sensors 2021, 21, 406. [Google Scholar] [CrossRef]
  21. Zhou, W.; Fan, Y.; Chang, Y.; Liu, W.; Wang, J.; Wang, Y. Pathogenesis of Comorbid ADHD and Chinese Developmental Dyslexia: Evidence From Eye-Movement Tracking and Rapid Automatized Naming. J. Atten. Disord. 2022, 27, 294–306. [Google Scholar] [CrossRef] [PubMed]
  22. Erbeli, F.; Rice, M.; Paracchini, S. Insights into Dyslexia Genetics Research from the Last Two Decades. Brain Sci. 2022, 12, 27. [Google Scholar] [CrossRef] [PubMed]
  23. Barkus, E.; de Leede-Smith, S.; Roodenrys, S.; Horsley, L.; Matrini, S.; Mison, E. Dyslexia: Links with Schizotypy and Neurological Soft Signs. PsyCh J. 2022, 11, 163–170. [Google Scholar] [CrossRef] [PubMed]
  24. Rauschenberger, M.; Baeza-Yates, R.; Rello, L. Technologies for Dyslexia BT-Web Accessibility: A Foundation for Research; Yesilada, Y., Harper, S., Eds.; Springer London: London, UK, 2019; pp. 603–627. ISBN 978-1-4471-7440-0. [Google Scholar]
  25. Madeira, J.; Silva, C.; Marcelino, L.; Ferreira, P. Assistive Mobile Applications for Dyslexia. Procedia Comput. Sci. 2015, 64, 417–424. [Google Scholar] [CrossRef] [Green Version]
  26. Rello, L.; Kanvinde, G.; Baeza-Yates, R. A Mobile Application for Displaying More Accessible EBooks for People with Dyslexia. Procedia Comput. Sci. 2012, 14, 226–233. [Google Scholar] [CrossRef] [Green Version]
  27. Dyslexia Testing and Evaluation. Available online: https://dyslexiaida.org/testing-and-evaluation/ (accessed on 6 February 2023).
  28. Usman, O.L.; Muniyandi, R.C.; Omar, K.; Mohamad, M. Advance Machine Learning Methods for Dyslexia Biomarker Detection: A Review of Implementation Details and Challenges. IEEE Access 2021, 9, 36879–36897. [Google Scholar] [CrossRef]
  29. Janković, M.M. Biomarker-Based Approaches for Dyslexia Screening: A Review. In Proceedings of the 2022 ZINC, Novi Sad, Serbia, 25–26 May 2022. [Google Scholar]
  30. Adler-Grinberg, D.; Stark, L. Eye Movements, Scanpaths, and Dyslexia. Am. J. Optom. Physiol. Opt. 1978, 55, 557–570. [Google Scholar] [CrossRef]
  31. Eden, G.F.; Stein, J.F.; Wood, H.M.; Wood, F.B. Differences in Eye Movements and Reading Problems in Dyslexic and Normal Children. Vision Res. 1994, 34, 1345–1358. [Google Scholar] [CrossRef]
  32. Rello, L.; Ballesteros, M. Detecting Readers with Dyslexia Using Machine Learning with Eye Tracking Measures. In Proceedings of the 12th International Web for All Conference; Association for Computing Machinery, Florence, Italy, 18–20 May 2015; Association for Computing Machinery: New York, NY, USA, 2015. [Google Scholar]
  33. Nilsson Benfatto, M.; Öqvist Seimyr, G.; Ygge, J.; Pansell, T.; Rydberg, A.; Jacobson, C. Screening for Dyslexia Using Eye Tracking during Reading. PLoS One 2016, 11, e0165508. [Google Scholar] [CrossRef] [Green Version]
  34. Prabha, J.A.; Bhargavi, R.; Harish, B. Predictive Model for Dyslexia from Eye Fixation Events. Int. J. Eng. Adv. Technol. 2019, 9, 235–240. [Google Scholar] [CrossRef]
  35. Prabha, A.J.; Bhargavi, R.; Harish, B. An Efficient Machine Learning Model for Prediction of Dyslexia from Eye Fixation Events. New Approaches Eng. Res. 2021, 10, 171–179. [Google Scholar] [CrossRef]
  36. Appadurai, J.P.; Bhargavi, R. Eye Movement Feature Set and Predictive Model for Dyslexia: Feature Set and Predictive Model for Dyslexia. Int. J. Cogn. Inform. Nat. Intell. 2021, 15, 1–22. [Google Scholar] [CrossRef]
  37. JothiPrabha, A.; Bhargavi, R.; Deepa Rani, B. V Prediction of Dyslexia Severity Levels from Fixation and Saccadic Eye Movement Using Machine Learning. Biomed. Signal Process. Control 2023, 79, 104094. [Google Scholar] [CrossRef]
  38. Asvestopoulou, T.; Manousaki, V.; Psistakis, A.; Smyrnakis, I.; Andreadakis, V.; Aslanides, I.M.; Papadopouli, M. Dyslexml: Screening Tool for Dyslexia Using Machine Learning. arXiv 2019, arXiv:1903.06274. [Google Scholar]
  39. Raatikainen, P.; Hautala, J.; Loberg, O.; Kärkkäinen, T.; Leppänen, P.; Nieminen, P. Detection of Developmental Dyslexia with Machine Learning Using Eye Movement Data. Array 2021, 12, 100087. [Google Scholar] [CrossRef]
  40. El Hmimdi, A.E.; Ward, L.M.; Palpanas, T.; Kapoula, Z. Predicting Dyslexia and Reading Speed in Adolescents from Eye Movements in Reading and Non-Reading Tasks: A Machine Learning Approach. Brain Sci. 2021, 11, 1337. [Google Scholar] [CrossRef] [PubMed]
  41. Vajs, I.; Ković, V.; Papić, T.; Savić, A.M.; Janković, M.M. Spatiotemporal Eye-Tracking Feature Set for Improved Recognition of Dyslexic Reading Patterns in Children. Sensors 2022, 22, 4900. [Google Scholar] [CrossRef]
  42. Nerušil, B.; Polec, J.; Škunda, J.; Kačur, J. Eye Tracking Based Dyslexia Detection Using a Holistic Approach. Sci. Rep. 2021, 11, 15687. [Google Scholar] [CrossRef] [PubMed]
  43. Vajs, I.; Ković, V.; Papić, T.; Savić, A.M.; Janković, M.M. Dyslexia Detection in Children Using Eye Tracking Data Based on VGG16 Network. In Proceedings of the 30th European Signal Processing Conference (EUSIPCO), Belgrade, Serbia, 29 August–2 September 2022. [Google Scholar]
  44. Jakovljević, T.; Janković, M.M.; Savić, A.M.; Soldatović, I.; Čolić, G.; Jakulin, T.J.; Papa, G.; Ković, V. The Relation between Physiological Parameters and Colour Modifications in Text Background and Overlay during Reading in Children with and without Dyslexia. Brain Sci. 2021, 11, 539. [Google Scholar] [CrossRef]
  45. Vajs, I.A.; Kvaščev, G.S.; Papić, T.M.; Janković, M.M. Eye-Tracking Image Encoding: Autoencoders for the Crossing of Language Boundaries in Developmental Dyslexia Detection. IEEE Access 2023, 11, 3024–3033. [Google Scholar] [CrossRef]
  46. Rauschenberger, M.; Baeza-Yates, R.; Rello, L. A Universal Screening Tool for Dyslexia by a Web-Game and Machine Learning. Front. Comput. Sci. 2022, 3, 628634. [Google Scholar] [CrossRef]
  47. Martins, V.F.; Lima, T.; Sampaio, P.N.M.; de Paiva, M. Mobile Application to Support Dyslexia Diagnostic and Reading Practice. In Proceedings of the 2016 IEEE/ACS 13th International Conference of Computer Systems and Applications (AICCSA), Agadir, Morocco, 29 November–2 December 2016; pp. 1–6. [Google Scholar]
  48. Wilkins, A. Reading Through Colour; John Wiley and Sons: West Sussex, UK, 2003. [Google Scholar]
  49. Griffiths, P.G.; Taylor, R.H.; Henderson, L.M.; Barrett, B.T. The Effect of Coloured Overlays and Lenses on Reading: A Systematic Review of the Literature. Ophthalmic Physiol. Opt. J. Br. Coll. Ophthalmic Opt. 2016, 36, 519–544. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  50. Kriss, I.; Evans, B.J.W. The Relationship between Dyslexia and Meares-Irlen Syndrome. J. Res. Read. 2005, 28, 350–364. [Google Scholar] [CrossRef]
  51. Rello, L.; Bigham, J.P. Good Background Colors for Readers: A Study of People with and without Dyslexia. In Proceedings of the 19th International ACM SIGACCESS Conference on Computers and Accessibility; Association for Computing Machinery: New York, NY, USA, 2017; pp. 72–80. [Google Scholar]
  52. Staub, A.; Benatar, A. Individual Differences in Fixation Duration Distributions in Reading. Psychon. Bull. Rev. 2013, 20, 1304–1311. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  53. Van Rossum, G.; Drake, F.L. Python 3 Reference Manual; CreateSpace: Scotts Valley, CA, USA, 2009; ISBN 1441412697. [Google Scholar]
  54. Hunter, J.D. Matplotlib: A 2D Graphics Environment. Comput. Sci. Eng. 2007, 9, 90–95. [Google Scholar] [CrossRef]
  55. Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Dubourg, V.; et al. Scikit-Learn: Machine Learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
  56. Bott, N.T.; Lange, A.; Rentz, D.; Buffalo, E.; Clopton, P.; Zola, S. Web Camera Based Eye Tracking to Assess Visual Memory on a Visual Paired Comparison Task. Front. Neurosci. 2017, 11, 370. [Google Scholar] [CrossRef]
  57. Ferhat, O.; Vilariño, F. Low Cost Eye Tracking: The Current Panorama. Comput. Intell. Neurosci. 2016, 2016, 8680541. [Google Scholar] [CrossRef] [Green Version]
  58. Lin, Z.; Liu, Y.; Wang, H.; Liu, Z.; Cai, S.; Zheng, Z.; Zhou, Y.; Zhang, X. An Eye Tracker Based on Webcam and its Preliminary Application Evaluation in Chinese Reading Tests. Biomed. Signal Process. Control 2022, 74, 103521. [Google Scholar] [CrossRef]
  59. Compensatory Skills and Dyslexia: What Does the Science Say? Available online: https://dyslexiaida.org/compensatory-skills-and-dyslexia-what-does-the-science-say/ (accessed on 7 February 2023).
  60. Rello, L.; Baeza-Yates, R. How to Present More Readable Text for People with Dyslexia. Univers. Access Inf. Soc. 2017, 16, 29–49. [Google Scholar] [CrossRef]
Figure 1. Feature 1—SI event detection and Feature 2—VAS event detection for a dyslexic subject, with the gaze lines corresponding to the trigger events colored red.
Figure 1. Feature 1—SI event detection and Feature 2—VAS event detection for a dyslexic subject, with the gaze lines corresponding to the trigger events colored red.
Brainsci 13 00405 g001
Figure 2. Feature 1—SI event detection and Feature 2—VAS event detection for a control subject, with the gaze lines corresponding to the trigger events colored red.
Figure 2. Feature 1—SI event detection and Feature 2—VAS event detection for a control subject, with the gaze lines corresponding to the trigger events colored red.
Brainsci 13 00405 g002
Figure 3. The boxplots showing SI event feature values for the maximally separated classes (control subjects evaluated on their best CC and the dyslexic subject evaluated on their worst CC) and the classes brought together (dyslexic subjects evaluated on their best CC and the control subject evaluated on their worst CC). Below each boxplot is a pie chart displaying the distribution of CCs that were used for the subjects’ feature values (best- or worst-performing CCs).
Figure 3. The boxplots showing SI event feature values for the maximally separated classes (control subjects evaluated on their best CC and the dyslexic subject evaluated on their worst CC) and the classes brought together (dyslexic subjects evaluated on their best CC and the control subject evaluated on their worst CC). Below each boxplot is a pie chart displaying the distribution of CCs that were used for the subjects’ feature values (best- or worst-performing CCs).
Brainsci 13 00405 g003
Figure 4. The boxplots showing VAS event feature values for the maximally separated classes (control subjects evaluated on their best CC and the dyslexic subject evaluated on their worst CC) and the classes brought together (dyslexic subjects evaluated on their best CC and the control subject evaluated on their worst CC). Below each boxplot is a pie chart displaying the distribution of CCs that were used for the subjects’ feature values (best- or worst-performing CCs).
Figure 4. The boxplots showing VAS event feature values for the maximally separated classes (control subjects evaluated on their best CC and the dyslexic subject evaluated on their worst CC) and the classes brought together (dyslexic subjects evaluated on their best CC and the control subject evaluated on their worst CC). Below each boxplot is a pie chart displaying the distribution of CCs that were used for the subjects’ feature values (best- or worst-performing CCs).
Brainsci 13 00405 g004
Table 1. Evaluation metrics on the test set for each feature and for all observed ML algorithms.
Table 1. Evaluation metrics on the test set for each feature and for all observed ML algorithms.
Feature/FrequencyML Algorithm
LRSVMKNNRF
SI event detection/60 Hz
ACC0.8700.8780.8650.868
Se0.8220.8220.8220.822
Sp0.9170.9330.9070.912
F1 score0.8610.8690.8560.859
AUROC0.8690.8770.8640.867
SI event detection/30 Hz
ACC0.8600.8600.8310.823
Se0.8110.8050.7950.795
Sp0.9070.9120.9650.850
F1 score0.8500.8490.8210.814
AUROC0.8590.8590.8300.822
VAS event detection/60 Hz
ACC0.8890.8890.8810.881
Se0.8490.8540.8590.859
Sp0.9270.9220.9020.902
F1 score0.8820.8830.8760.876
AUROC0.8880.8880.8810.881
VAS event detection/30 Hz
ACC0.8780.8600.8620.860
Se0.8590.8160.8270.827
Sp0.8960.9020.8960.891
F1 score0.8740.8510.8550.852
AUROC0.8780.8590.8620.859
Total reading time
ACC0.7690.7960.7460.738
Se0.6540.6050.6220.589
Sp0.8600.9740.8650.834
F1 score0.7270.7420.7060.674
AUROC0.7900.6750.7580.733
Values in bold represent the best obtained results for the 60 Hz and 30 Hz.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Vajs, I.; Papić, T.; Ković, V.; Savić, A.M.; Janković, M.M. Accessible Dyslexia Detection with Real-Time Reading Feedback through Robust Interpretable Eye-Tracking Features. Brain Sci. 2023, 13, 405. https://doi.org/10.3390/brainsci13030405

AMA Style

Vajs I, Papić T, Ković V, Savić AM, Janković MM. Accessible Dyslexia Detection with Real-Time Reading Feedback through Robust Interpretable Eye-Tracking Features. Brain Sciences. 2023; 13(3):405. https://doi.org/10.3390/brainsci13030405

Chicago/Turabian Style

Vajs, Ivan, Tamara Papić, Vanja Ković, Andrej M. Savić, and Milica M. Janković. 2023. "Accessible Dyslexia Detection with Real-Time Reading Feedback through Robust Interpretable Eye-Tracking Features" Brain Sciences 13, no. 3: 405. https://doi.org/10.3390/brainsci13030405

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop