1 INTRODUCTION

The COVID-19 pandemic has had a significant impact on almost all spheres of everyday life. To reduce the infection rate, many companies and organizations transferred their employees to remote jobs and implemented other restrictive measures. Thereby, they accelerated the progress in digitalization, distance learning, telemedicine, online commerce, etc.

However, the positive technological impetus to the development of the IT industry has inevitably led to the growth of cybercrime. In the year the pandemic began (2020), a sharp increase in the number of cybercrimes, both throughout the world and in Russia, was recorded. According to Data Economy, an autonomous non-profit organization, in 2020, 363 thousand cybercrimes were reported in Russia, which is 77% more than in the previous year [1]. The statistics published by InfoWatch also indicates, in addition to the increase in the number of attacks by hacker groups, the emergence of new risks associated with remote work [2]. In 2021, a sharp increase in the number of intentional information leaks (82% of the total number) and leaks resulted from actions of cybercriminals (up to 63%) was recorded worldwide. Russia’s share in the number of global information leaks is quite impressive and amounts to 16.9%.

According to Ivideon, the number of cyber attacks in 2021, as compared to the previous year, increased by 40% worldwide and by 54% in Russia.

Even though information security has always been one of the highest priorities for any organization, this problem becomes especially important under the current conditions.

The education sector is one of the most vulnerable sectors in terms of cybersecurity, which is due to the mentality of students and circulation of large amounts of confidential information. This makes universities attractive targets for hackers. In September 2021 alone, there were more than 10 incidents involving ransomware that targeted personal data of students and teachers [3].

On the other hand, the pandemic has had a significant impact on the traditional system of higher and school education. According to UNESCO [4], the crisis has affected almost 1.6 billion students in more than 190 countries on all continents. The closure of schools and other educational institutions affected 94% students worldwide. Even though the number of partially or completely closed educational organizations decreased to 7.5% in September 2021 [5], the forms and methods of education have significantly transformed towards distance learning. However, universities are still not ready for full-scale implementation of distance learning. In the context of higher education, online learning is often reduced to webinars, video/audio conferences, and use of e-learning platforms (Moodle, WebTutor, and iSpring Learn).

Online tests and exams are of great importance for e-learning. Tests allow the teacher to receive feedback from each student, estimate the academic progress, and improve the learning process. However, students sometimes resort to a number of academic fraud methods during online tests, including impersonation [6, 7]. Currently, there are no efficient methods for detecting this type of fraud.

2 ACADEMIC INTEGRITY OF ONLINE LEARNING

Academic integrity is a problem not only for the education sector but also for the whole society. Academic integrity violations negatively affect the educational environment, skills acquired by students, and the overall image of the university.

The global transition of higher education to online technologies inevitably activated many traditional forms of academic fraud: plagiarism, falsification of results, unauthorized collaboration, impersonation, etc. In addition to organizational means used to counteract these types of fraud (the university code of conduct for students and administrative sanctions for academic integrity violations), there are also methodological recommendations for teachers. For online classes, it is recommended to

– increase the individuality and specificity of test assignments;

– reduce the factological part of assignments and augment their conceptual part to develop thinking skills;

– increase the number of synchronous checks of answers in online mode;

– increase the number of small and simple assignments and use free forms, e.g., an essay, for large assignments.

The proposed measures [8] partially reduce the number of academic integrity violations; however, they significantly increase the workload on the teacher.

Continuous authentication by hidden monitoring can be a solution to counter unauthorized cooperation and impersonation.

3 KEYSTROKE AUTHENTICATION AND IDENTIFICATION

To protect a computer system from unauthorized access, two-step verification is generally used:

– primary identification: establishing the identity of a user, i.e., confirming the legitimacy of an authorized user;

– dynamic (continuous) authentication: the continuous confirmation of the identity of the legitimate user.

3.1 Authentication Methods

Authentication is a process of comparing the data provided by the user with the data stored in a directory services database. If these data match, then the user is given access to protected resources; otherwise, access is denied [9].

There are several methods for user authentication (see Fig. 1). These methods can be divided into three main groups based on the following paradigms [10]:

Fig. 1.
figure 1

Authentication methods.

– something you know (e.g., password or PIN): knowledge-based authentication;

– something you have (e.g., token or smart card): attribute-based authentication;

– something you are, i.e., physiological and behavioral biometrics.

The “something you are” paradigm uses biometric features: physiological (fingerprint, face, iris, etc.) and/or behavioral (handwriting or keystroke patterns, gait, etc.) [19].

User recognition based on keystroke dynamics is quite attractive to organizations because it has lower implementation cost in comparison to other biometric methods, as it does not require any additional hardware. In addition to a standard keyboard, a highly efficient software application is required. While being rather accurate, this method is also capable of recognizing users in hidden mode, which is comfortable and does not distract users. Individual keystrokes, rhythm, and typing speed determine the keystroke dynamics (KD) of a user. All behavioral characteristics, including the KD, can gradually change with time; however, the probability of stealing or imitating these data is significantly lower than that for physiological data. That is why the KD can be useful for identification and authentication [11, 12]. The advantages, disadvantages, and examples of the authentication methods considered above are summarized in Table 1.

Table 1. Characteristics of authentication methods

3.2 Authentication Modes

Hidden continuous monitoring is the most reasonable (for recognition systems) and comfortable (for users) identification method. According to the global security survey published by IBM (2018), 44% of the respondents considered biometrics to be the most secure authentication method, while 65% found that biometrics facilitates the authentication process [13].

The KD, as a biometric behavioral characteristic, is dynamic in nature. It includes constant and random components. The constant component is due to human physiology, i.e., the user’s ability and skills in working with a keyboard. The random component depends on the psycho-emotional state of the user.

For the recognition, the dynamic behavioral features of the KD are more complex than the physiological ones. However, it is the behavioral features that are difficult to fake [10], which improves the effectiveness of impersonation detection.

Depending on the type of text (fixed or free), two main authentication modes can be distinguished [14]:

– static authentication (primary or by event);

– dynamic authentication (continuous).

It is continuous authentication that makes it possible to organize hidden identity verification of users during the entire session in any application. The recognition system captures keystrokes and compares them with an available template [6, 11, 1719].

Static authentication can supplement the initial login procedure, or it can be activated when suspecting an intrusion [15, 16].

Both static and continuous authentications provide the second level of security when the user is already logged in. Meanwhile, dynamic authentication is aimed at continuous verification of the users or their psycho-emotional states.

This study is devoted to user authentication by continuous monitoring of KD features in online exam applications. More specifically, we focus on the following problems:

– collecting and updating KD samples for domain users;

– extending well-known static identification approaches to the case of dynamic identity recognition based on free and long texts;

– reducing the dimension of the space of extracted keystroke features to improve the selective properties of KD samples and the efficiency of continuous authentication.

3.3 Authentication Lifecycle

KD-based continuous authentication has a registration stage and an authentication stage (see Fig. 2).

Fig. 2.
figure 2

Continuous authentication lifecycle.

At the registration stage, the system receives keystroke data. Then, keystroke features (hold time, flight time, etc.) are extracted, a keystroke profile (template) is formed or modified in the database, and user authentication is carried out.

Thus, the lifecycle of continuous authentication includes four main steps.

I. Collection of KD data is a continuous process when the user works with a keyboard in any application. In Windows, the message interception mechanism is used. The Windows hook makes it possible to detect any keystroke event [20]. The OS captures the ANSI code and timestamps of key presses (Down/Press) and key releases (Up/Release). The accuracy of keystroke measurements is up to milliseconds.

II. Extraction of classification features. The raw keystroke data must be cleaned of outliers and invalid values; in some cases, they need to be normalized. Based on these data, more significant KD features (typing rhythm, typing speed, flight time, etc.) can be obtained, which reflect the unique behavioral characteristics of the user. There are many KD features with the most popular ones among researchers being digrams (digraphs): timestamps of two key states [17, 2123]. The main KD features are as follows:

– hold (dwell) time;

– flight time;

– typing speed;

– number of typing errors;

– degree of arrhythmy when typing;

– specific use of service keys.

Figure 3 shows some of the most commonly used time and frequency features.

Fig. 3.
figure 3

Keystrokes in the down time/up time notation.

– DU: dwell time (DT) is a time interval between a key press (down, press) and a key release (up, release).

– UD: flight time (F-RP) is a time interval between the release of the previous key and the press of the next key.

– UU, F-PP or DD, F-RR are intervals between pressing or releasing one key and pressing or releasing the next key, respectively.

Subsystems for preprocessing timestamp data and extracting KD features form a set of desired features for each keystroke. Using this set, a keystroke profile (template) of the user is generated and placed in the database.

As a result of KD feature extraction, we obtain a bank of profiles, which requires adaptation. Being a biometric behavioral characteristic, the keystroke profile is variable and needs to be updated. For this purpose, the sliding window technique can be employed [49]. The bank of profiles is used to train the classifier and at the stage of identity recognition.

III. User recognition.

To protect a computer system from unauthorized access, two-step verification is generally implemented:

– primary identification: the identification of the user;

– dynamic (continuous) authentication: the continuous confirmation of the identity of the legitimate user.

Authentication is a task of classifying users registered in the system.

The basic methods and algorithms for user classification (recognition) are the same for static and continuous (dynamic) authentication. They can be divided into three groups:

– statistical;

– based on closeness estimates;

– machine learning.

The use of these methods and the main representatives of each group are considered in detail in [20].

Historically, the works on keystroke dynamics recognition dealt with predefined and structured texts, i.e., with static authentication. According to different authors [22, 24], the number of studies devoted to continuous authentication is still small and does not exceed 10–15% of all works on keystroke recognition. The first fruitful investigation of continuous authentication is the work by Gunetti, published in 2005, where recognition accuracy exceeded 90% [25]. For comparison, the accuracy of the very first recognition methods on free texts (Monrose, 1997) was 23%.

The reviews on keystroke recognition published in the last decade made it possible to generalize the data on the effectiveness of continuous authentication (the main results are presented in Table 2). In addition to the references, the table includes the classification parameter, recognition method, and performance indicators. The data were borrowed from our own studies [20, 26] or adapted from the reviews [17, 22, 24, 2733].

Table 2. Dynamic identification methods

Finally, we should note the conditional nature of KD-based approaches to user recognition. These approaches generally include a model and training methods; however, their combinations can be different.

IV. Decision making on user’s legitimacy.

This step fully depends on a particular applied problem and recognition performance indicators.

In the case of dynamic identification, the main goal of continuous monitoring is to provide permanent access to network resources for a user registered in the domain, as well as to prevent access for an unregistered user.

Generally, the probabilities of the corresponding false events are estimated: a false denial of access for a registered user and a false provision of access for an unregistered user. By analogy with estimates in statistical radio engineering, the following errors are most commonly used in keystroke dynamics studies [17, 20, 21, 36, 38].

– False rejection rate (FRR), which is a type I error that estimates the false denial of access to a legitimate (registered) user:

$$FRR = \frac{{FR}}{{TA + FA + TR + FR}}.$$
(1)

– False acceptance rate (FAR), which is a type II error that estimates the false provision of access to an illegitimate user:

$$FRR = \frac{{FA}}{{TA + FA + TR + FR}}.$$
(2)

Here,

– true accept (TA) is the number of valid accepts of legitimate users;

– true reject (TR) is the number of valid rejects of illegitimate users;

– false accept (FA) is the number of invalid accepts of illegitimate users;

– false reject (FR) is the number of invalid rejects of legitimate users.

The denominators of (1) and (2) contain the total number of attempts.

FRR and FAR depend on the adjustable threshold or sensitivity of the algorithm (see Fig. 4).

Fig. 4.
figure 4

Performance indicators for keystroke-based identification.

FRR, FAR, and ERR are sufficient for making a decision on the acceptance or rejection of the user. If the goal of the monitoring system is to provide a high level of security, then low thresholds should be used, which correspond to a large percentage of false rejects (FRR). Large FRR values provide higher security and more difficult access to the system for everyone, both insiders and outsiders. For a high threshold and low sensitivity of recognition algorithms, access becomes easier; however, FAR and the number of falsely accepted users increase. The tradeoff between FRR and FAR has to be determined individually in each case.

Another popular indicator that does not depend on the threshold is the equal error rate (EER). EER values correspond to equal values of FRR and FAR, which makes EER a universal performance indicator for any authentication system.

These three metrics (FRR, FAR, and EER) are most widely used for decision making in monitoring and authentication systems. Studies on keystroke dynamics also frequently employ the receiver operating characteristic (ROC), which is a ratio between TA and FA for different thresholds. ROC reflects the limiting capabilities of algorithms, which is especially important when investigating various classifiers.

4 HIDDEN MONITORING IN ONLINE EXAMINATION SYSTEMS

The COVID-19 pandemic has accelerated the introduction of the blended learning model, and many universities have implemented learning management systems. The main contribution to the final assessment of students' performance in these systems is made by online exams and tests. However, it is these forms of online learning that are most prone to academic fraud because the online environment allows students to work with little or no control. According to some researchers [4345], the success of online exam systems is due to the use of biometric control and continuous monitoring during online exams. Keystroke dynamics makes it possible to organize the hidden monitoring of the identity of a legitimate student (who has passed primary identification) in a mode comfortable for him or her.

The main problems of keystroke dynamics—lifecycle, components, and efficiency—are discussed in Section 3. A specific use of these technologies is considered in Sections 4.1–4.4.

4.1 Authentication Lifecycle

The principles of continuous authentication outlined above are used to confirm the legitimacy of a student in the process of online testing.

The capabilities of continuous authentication in solving this problem are investigated by computational experiment. The experiment is conducted in accordance with the block diagram of the system shown in Fig. 5, which includes three main subsystems:

Fig. 5.
figure 5

Continuous authentication architecture (block diagram).

– registration;

– adaptation or template updating;

– user authentication.

In the case of distance learning, each subsystem operates in continuous mode. As shown in Fig. 5, the common steps of registration and authentication are data collection, data processing, and keystroke feature extraction.

4.1.1. Data collection in the experiment. The development of a continuous authentication system and its performance testing are carried out using some keystroke dataset, statistical or locally collected one.

There are several public datasets based on fixed texts (Buffalo and BiosecurID) and free texts (Clarkson II and Villani-2010) [18, 23, 46, 47]. Some fixed datasets (CMU and WEBGREYC) [46] form templates by using login and password, which is not acceptable for dynamic authentication. In addition, these datasets do not include templates for Russian texts. When recognizing Russian users, English-language templates are of interest only for primary identification and static authentication based on password data.

In this work, local data collection is carried out using a special application developed for Windows. The application has client–server architecture. On the client side, keystroke data are collected locally in the background.

The target audience of this study is the university domain whose users have above-average computer skills. Using the keyboard input interception mechanism implemented in Windows, the so-called hook traps, all keyboard events are recorded. For each event, an array is formed that contains the following data:

– user login in the domain;

– key code;

– event (key pressed/released);

– timestamp.

The collected data are represented in the text format as follows (see Fig. 6).

figure 6

Fig. 6.

The other functions of the application are implemented on the server. This reduces the risk of stealing the keystroke data from less secure client machines.

4.1.2. Data preprocessing. The raw data collected during the user session are preprocessed in the system to eliminate outliers, unreasonable hold times (30–200 ms), unpaired events, and short sessions.

The most important part of the preprocessing is the choice of the duration of the session. Mathematically, this corresponds to the choice of time windows into which the flow of keyboard events is partitioned. The sizes of the windows can differ at the stages of registration and authentication. There are various methods used for this purpose [18]:

– partitioning based on the duration of the user session or on the number of clicks;

– sliding windows.

This study uses a sliding window of 500 characters in continuous authentication mode. The size of the data sample must be large enough to ensure the representativeness of individual characters in the window, as well as guarantee that their estimates are statistically significant and unbiased. However, it is well known that the keystroke pattern tends to change due to the psycho-emotional component. That is why the time window should be sliding. The sliding window is also opposed by the possibility of reverse biometrics: the possibility of compromising long unprotected templates. By combining these two methods, we select the minimum sample size for the consistent estimate of the average values for Russian letters.

4.1.3. Extraction of KD features and generation of keystroke templates. The dimension of the resulting feature space in the continuous authentication problem is quite high. This is due to the size of the Russian alphabet and the large number of KD features.

The analysis of the researches devoted to keystroke features showed that the most popular ones are DU and UD (dwell time and flight time): according to [12, 20], the frequency of using each of them ranges from 30% to 40% in applied research.

In our study, the keystroke data from the last session of each user are sent to the server in the text format.

Then, the statistical characteristics DU and UD are computed for each letter during a user session. The result is an updated keystroke template, adapted after the last session and saved in a JSON file.

4.1.4. Vector parameter generation. The keystroke template stores the processed and reliable data on the keystroke dynamics of users in the last session in terms of the selected KD features (average dwell values for each key).

To improve the informativity and reliability of the generated templates, additional metrics and tools can be used to derive an aggregate KD indicator.

Dimensionality reduction based on extraction of stable features.

Reducing the dimension of the feature space is an important problem for keystroke dynamics, especially in the context of distance education and online testing, where the confirmation of the legitimacy of a student must be carried out as quickly as possible to detect impersonation.

On the one hand, the extraction of stable KD features contributes to the reduction of the feature space. On the other hand, it improves the selective properties of the templates.

When extracting stable features, we can improve the informativity of the KD features themselves [49]. The idea of these methods is to narrow timestamp ranges based on some rule, most often, by using statistical criteria. In this case, the narrowing technique can be applied to individual sessions, individual users, etc.

In addition, to improve the informativity of KD features, heuristic methods can be employed. In this study, this method is associated with the frequency of alphabetic letters in texts. According to the Russian National Corpus (https://ruscorpora.ru), the frequency of Russian letters (expressed in %) decreases from 10.985 (letter О) to 0.9965 (letter Х) and 0.0375 (letter Ъ) (see Fig. 7a). Therefore, to obtain consistent estimates for all Russian letters, we need a very large sliding window, which is unacceptable in online authentication. A way to reduce the size of the window is to use a letter frequency threshold as a stable feature. For the threshold selected in this study (0.55), this means that the user template does not include the rarely used letters Ц, Щ, Э, Ф, Ъ, and Ё (see Fig. 7b).

Fig. 7.
figure 7

Frequencies of Russian letters.

The KD feature vector includes the features (the average DU values of the last session) of 27 Russian letters with the weight coefficients corresponding to the frequency of these letters in texts (see Fig. 7b).

4.1.5. Recognition of legitimate users. Static authentication and dynamic authentication are both aimed at confirming the legitimacy of the user who has already passed the initial registration. The difference is only in the type of the generated text. In the case of dynamic (continuous) authentication, the user creates free text in any OS application at any time; in the case of static authentication, the text and time of its generation are predetermined by the security system.

It should be noted that the confirmation of the user’s legitimacy based on his or her keystroke dynamics is a one-class classification problem. This problem is more complex than multiclass classification because the data of an illegitimate user are not available to the recognition system. The classifier is trained on objects of one class; in the process of testing, the algorithm determines whether a new object belongs to this class. The data of registered users are stored in the system in the form of dynamically updated templates. They can be considered objects of the same class (that of legitimate users).

In online learning systems, one-class classification can be carried out to recognize unregistered users, i.e., malefactors.

In online testing, one-class classification prevents impersonation of students.

Even though dynamic authentication is much more complex than static authentication, the recognition methods are similar. A brief review of the dynamic authentication methods developed over the last decade is presented in Table 3. The classification of recognition methods is quite conventional; however, it is possible to distinguish machine learning methods, statistical methods, and methods based on metric distances [20].

Table 3. Comparative analysis of performance indicators

In this work, the most popular recognition methods are chosen as classifiers in each of the three groups for the one-class classification problem:

– support vector machines (SVM) [17, 21, 34];

– k-nearest neighbors (kNN) algorithm [17, 21, 27, 34, 46];

– Euclidean and Manhattan metrics.

The basic idea of a one-class support vector machine (OCSVM) is to detect the boundaries of one class, rather than to separate objects of several classes as in the multiclass problem [48]. In addition, one-class classification is also known as an anomaly detection problem, which consists in finding patterns in data that do not conform to expected behavior [49], e.g., when the psycho-emotional state of the user changes under the influence of alcohol, etc.

The OCSVM maps feature vectors to a higher-dimensional space by using a kernel function. In the case of a radial basis function (RBF) kernel, there is a hyperplane that separates the majority of objects of a given class (in our case, legitimate users) from the origin. Exceptions are objects that lie closer to the origin than the hyperplane.

A one-class kNN (OCkNN) algorithm estimates the distance between class objects. A new object is considered an exception if the majority of the objects (e.g., their pth part) of the entire class are at a distance exceeding D from the new object. The distance is computed using a selected metric in the feature space (p and D are the parameters of the method).

The OCSVM and OCkNN algorithms are quite easy to implement. Their main requirement is the representativeness of the source dataset.

5 RESULTS

The main processes of the continuous authentication lifecycle in online learning are data registration and student authentication (see Fig. 2). In the process of registration, the continuous collection of keystroke data with subsequent extraction of keystroke features is carried out (Fig. 5).

The templates (profiles) of domain users are dynamic; they are formed by continuous monitoring of keystroke dynamics and do not depend on any particular application. The keystroke data for each domain user are collected in the sliding window and the template is formed using Russian and English letters. The study was carried out in the domain of a national Russian university. The window size was 500 clicks, upon accumulation of which the data were transferred to the server component of the program for preprocessing and KD feature extraction. As the number of clicks grew, the window was shifted and the next set of data was transmitted to the server, thus enabling continuous data registration in the current user session.

The server component of the program computes the average dwell times in the current session for each character and updates the template stored in the bank.

For each user, the bank stores templates for the last 10 sessions, which makes it possible to track changes in keystroke dynamics, which are due to fatigue, psycho-emotional state, etc.

The identification capabilities of templates generated for different users can be estimated visually and statistically. Figure 8 shows templates of six random users of the university domain for Russian letters. The Y-axis represents the dwell time (in ms). The individual lines depict the typing rhythm and determine the keystroke pattern of each user.

Fig. 8.
figure 8

Visualization of domain user templates.

Typing rhythms and speeds differ for different users, which results in different keystroke dynamics. This difference of KD features is also confirmed statistically, e.g., by the type and parameters of the distribution density (see Fig. 9).

Fig. 9.
figure 9

Distribution density for the KD parameters.

For instance, the distribution for User3 is characterized by a small dispersion and average typing speed (expectation), which is in complete agreement with Fig. 8 and is confirmed by small deviations between the mode and median of the series (large and small dashed lines).

Figures 8 and 9 show that the average values for User4 and User1 are approximately equal. However, the distribution for User1 is heavy-tailed. The mode and median for this series differ significantly from one another, as in the case of User06. For a representative sample and high typing speed, this can be a discriminative feature of an altered psycho-emotional state.

User5 also has a high typing speed. In this case, the distribution density is bimodal, which corresponds to good typing skills. For this user, there are typing overlaps when the next key is pressed while the previous one is not released.

The next stage of recognition is the confirmation of the legitimacy of the student. For this purpose, the template of the registered user from the bank is compared with the current template (see Fig. 5). When confirming the legitimacy of the user, there are two possible outcomes:

– the template of the last session matches the template from the bank for the registered user;

– the templates do not match.

The pattern matching is analyzed using SVM, kNN, Euclidean metric, and Manhattan metric. The choice of these methods is substantiated in Section 4.1.5.

The most important characteristic for any method is the decision threshold. The threshold is selected (assigned) by the security system based on priorities of tasks. Small thresholds correspond to small differences between the basic and current templates and complicate access to the corporate network for everyone, including legitimate users. Large thresholds (low sensitivity) correspond to easy access for everyone. Type I and II errors for the Manhattan metric are shown in Fig. 10 depending on threshold values.

Fig. 10.
figure 10

User recognition performance.

In Fig. 10b, FAR, FRR, and EER are plotted taking into account the letter frequency; in Fig. 10a, without taking it into account.

It can be seen that there is a significant decrease in the errors, by an order of magnitude on average at different threshold values. For instance, on the data from Table 3, these values are 10.85 and 0.995 for the Euclidean metric. In Table 3, for all selected methods, the EER values are given for a threshold at which FAR = FRR.

The authentication features based on the Manhattan and Euclidean metrics are almost identical with respect to the frequency of alphabetic letters. However, they have certain differences when the frequency is not taken into account. The kNN method has a slightly better performance, ERR = 0.545, at a slightly higher threshold 15 ms. However, this requires a time-consuming selection of algorithm parameters and does not guarantee that parameter adjustment would not be necessary under real-world conditions. The lack of optimal parameter values explains the low authentication results of the SVM method.

A confusion matrix is also a useful tool for analyzing the events that occur when recognizing legitimate and illegitimate users (see Table 4). The intersections of its rows and columns contain true (T) or false (F) outcomes of recognition: accept (A) or reject (R). The columns correspond to the recognition results, while the rows correspond to the real users.

Table 4. Confusion matrix

The performance indicators shown in Table 3 are computed as follows:

$$Accuracy = \frac{{TA + TR}}{{TA + FA + TR + FR}}.$$
(3)
$$Precision = \frac{{TA}}{{TA + FA}}.$$
(4)
$$Recall = \frac{{TA}}{{TA + FR}}.$$
(5)

These three indicators reflect the quality of authentication of legitimate users from different angles.

Accuracy estimates the correct accepts and rejects for all possible legitimate and illegitimate users.

Precision is the ratio of correctly accepted users to all users accepted by the system.

Recall reflects the number of accepted users out of all legitimate ones. Recall is also referred to as sensitivity of a recognition model.

Accuracy is usually expressed as a percentage, while Precision and Recall vary from 0 to 1.

The last tools for authentication quality assessment are the DET and ROC curves shown in Fig. 11.

Fig. 11.
figure 11

User recognition performance.

These curves confirm the similar performance of the kNN method and the methods that are based on the similarity of templates and take the letter frequency into account.

6 CONCLUSIONS

In this paper, we have considered the approach to confirming the legitimacy of students in the process of distance learning by hidden monitoring of KD features.

It has been found that the hidden identification of students is possible on the basis of continuous monitoring of keystroke data for any software application. As a result, the following conclusions can be made.

1. Correction (adaptation) of KD samples by using a sliding window is required, which allows one to dynamically track changes in the user’s keystroke pattern and psycho-emotional state.

2. Selection of stable keystroke features contributes to the dimensionality reduction of the KD feature space, which improves selective properties of the templates.

3. In this study, as a stable feature, we have used the frequency of alphabetic letters in texts according to the Russian National Corpus (https://ruscorpora.ru). The exclusion of six letters with frequencies below 0.5% from the templates has led to a significant change in all performance indicators. For instance, ERR has decreased by an order of magnitude, on average from 105 to 15, while Accuracy, Precision, and Recall have increased by 6–135 on average, reaching 98% for both the Euclidean and Manhattan metrics.

4. The kNN method has showed slightly better results for optimal parameters: ERR = 0.545 (versus 0.795), Accuracy = 99%, and Precision = 0.98.

5. Combining simple distance estimation algorithms with letter frequency leads to the absence of complex parameter optimization procedures, as in machine learning methods (kNN), while providing similar performance.

With the recognition performance being a key factor in continuous authentication, parameter adaptation significantly reduces the overall efficiency of user validation.