Introduction

Technological and pedagogical advances are redefining education. E-learning is at the center of this conjunction. Along with technology advancements, scalability and reduced costs also made e-learning attractive. Many learning materials, text resources such as basic web pages, and multimedia resources as videos have been uploaded in recent years due to this fast growth of information and communication technology usage in education. The substantial amount of information in the learning systems creates cognitive overload and disorientation. Also, the learner population is highly dynamic and heterogeneous, with differences in their learning preferences, basic knowledge, learning style, and interaction with the learning environment (Chen et al., 2014, Ciloglugil & Inceoglu, 2018, Christudas et al., 2018). Hence, the knowledge delivered in fixed sequences or patterns will create dissatisfaction and disinterest in learners. The “one-size-fits-all” approach may not be satisfactory for the learners (Essalmi et al., 2010). Learners demand personalized knowledge delivery which adapts to their changing needs (Deng et al., 2017; Hwang et al., 2020a). In general, Learning Management Systems, LMS, do not meet the requirements of individual learners depending on their profile. However, taking learners’ profiles into account can improve the learning experiences and course success of students (Imran et al., 2016). In facilitating personalization in LMS, recommender systems can be used to suggest appropriate learning objects to learners in order to enhance their learning. Thus, generating adaptive, customized learning paths is an important research topic in the design of learning environments (George & Lal, 2019; Hwang et al., 2020b; Raj & Renumol, 2021). In education technology, it is advantageous to extract new hidden patterns in learner data for online learning systems. Personalized learning full-path recommendation research is particularly significant for the advancement of E-learning systems (Zhou et al., 2018). Online learners develop massive data with big-data features about their learning habits, which helps discover individual learning patterns (Chen & Zhang, 2014). The data generated from the learning environments can be fed back to the system, contributing to learning evaluation and monitoring (Sachan & Saroha, 2022). Thus, the learning material recommenders can be improved to monitor the student performance and adapt to the changes in the performance and their learning preferences.

According to Chen et al. (2014), the critical challenges to be focused on in the design of the learning resource recommenders are to provide convenient and effective access to the learning resources and boost the learner’s learning experience and satisfaction. So, the recommender system needs to adapt to the learner’s changing performance and preferences. The recommender system should map adaptive and personalized resources reducing the knowledge gap with the learner (Shi et al., 2020). This research work aims to generate personalized learning paths adapted to the changes in learning preferences and performance in real-time. Here, the learner log is constantly monitored, and getting feedback from this log, adjust the values of dynamic factors contributing to personalized content recommendations. A learning object (LO) is needed to learn a topic, and the sequences of the topics form the learning path. Learning objects are internet deliverable and reusable, instructional components that support learning (Wiely, 2002). We plan to find suitable learning objects for each topic and order them based on the knowledge dependencies between the topics.

To reduce the knowledge gap between the learner and the learning resources, the proposed model dynamically calculates the difficulty level of an LO and selects the LO with high similarity with the learner’s ability. The system consists of the learner model, the LO model, learner logs, and the recommender engine (Raj & Renumol, 2018). The learner model represents the learner’s behavioral information (learning preference, learning style), status information (learned LO, basic knowledge), and dynamic information (time for LO completion, score of LOs, number of attempts). The current study uses the Felder-Silverman learning style to analyze the learner’s learning style (Felder & Silverman, 1988). The LO is modeled using IEEE LOM (Risk, 2002), and the fields are used as described in Raj and Renumol (2019). The IEEE LOM fields are Structure, Format, Learning Resource Type, Interactivity Type, Interactivity Level, and Difficulty. An additional field is used to hold the average rating for each LO. To model the learner, learning material, and learner log this study uses the ontology-based method, and the design of the model is explained in Sect. 4. The earlier works show that the ontology-based LO recommender system can perform well in personalized learning environments. The significance of ontology-based models is that they better handle the cold-start issues and data sparsity problems in recommendations. The SPARQL queries extract similar learning objects based on similar learner grouping (Joy et al., 2021). Thus, the proposal works in the following steps:

  • Find all possible sequences of topics forming paths between the starting and final topics.

  • For each topic

    1. o

      Find the Top N LOs based on the learning log of similar learners

    2. o

      Refine the Top N LO list so that the ability of the learner suitably maps with the difficulty of the LO.

  • Form the LO sequences based on the topic sequences

  • Suggest the sequence with the best time duration suitable for the learner’s available time

  • Log the interactions of learner such as score, time taken to learn, number of attempts, the rating is given for each learned material

  • Update the difficulty level of the learning material and ability of the learner according to the logged information

For every course, the instructor annotates the learning resources and develops the knowledge link between the topics. The relationship between the topics is stored as a graph structure called a concept graph or knowledge graph. This work uses a two-layered knowledge graph, with lessons forming the upper layer and LOs in the lower layer. The knowledge graph is traversed to find the concept sequence. The instructor assigns an initial difficulty level for each LOs. Also, the instructor gives an approximate time expected to learn the LO. These values are used to cold start the model. Each time a learner interacts with the LO, the time and score associated with LO are updated. The suggested learning paths are optimized for learner attributes and aim to maximize learner satisfaction and performance.

The major implications of our work are:

  1. (1)

    Design of a learner model and LO model with static and dynamic parameters, where the length of the parameters is comparatively larger than most of the existing works. Most of the recommendation system studies focus on learning object ratings as feedback from the learner. The model in this study considers the dynamic parameters as the time taken for learning, score obtained from learning, number of attempts taken for learning a concept along with the ratings provided by the learners. The model adjusts the difficulty of the learning objects and the ability of a learner to learn that object in real-time so that the adaptivity of the recommendations is enhanced.

  2. (2)

    Real-time learner data analytics is incorporated to improve the accuracy of the next recommendations. Evaluation of the model using real learner data.

  3. (3)

    Implementation of an adaptive learning path recommendation model which works in two steps. Initially, a concept path is constructed by arranging the concept in order following a knowledge graph. Secondly, personalized learning objects are selected for each of the concepts in the path, forming a learning path.

  4. (4)

    The model addresses data sparsity and cold-start issues of recommender systems.

  5. (5)

    A summary of recent studies on the learning path adaptation.

The rest of the paper is organized as follows. The problem statement and research questions are introduced in Sect. 2. Section 3 briefly explains the related works on the current domain from 2018 to 2021. Section 4 elaborates on the design of the knowledge and domain models. Section 5 describes the learning path recommendation model that is proposed in this paper. The subsections of Sect. 5 elaborate on the algorithm used for learning path generation. The experimentation procedure and results are presented in Sect. 6, and Sect. 7 discusses the developments in correlation with the research questions and states the limitations of the work. Section 8 concludes the current work with a discussion on future work.

Problem statement and research questions

The adaptive learning path recommendation model aims to provide learners with the most suitable personalized sequences of learning activities to follow (de Marcos et al., 2008). So, the objective of this work is to find a flexible model for personalized and adaptive recommendations than the “one-size-fits-all” approach. It was stated by Chen et al. (2014) about the importance of providing better access to learning resources to enhance the learner’s performance and satisfaction. Also, suppose the learning model is generating the learning path by optimizing learner performance in terms of learning time and scores obtained. In that case, the paths are more effective (Chen, 2011). So, we decided to focus more on the learning duration, expected score, adaptivity, and acceptance of recommended learning resources. So, the research questions addressed are stated as follows:

RQ1::

Can we accurately predict learning duration and expected score during the learning path recommendation process?

RQ2::

What learner-centric personalization parameters generate adaptive and diverse learning paths in an e-learning environment?

Related works

Many studies address the problem of learning path recommendations. From the literature, we can observe that this area’s works fall under domain-based and heuristic models (Raj & Renumol, 2021). Both studies use a knowledge structure for recommending, considering the similarity between the learning objects. The researchers use various algorithms/methods/tools based on the method chosen for guiding the learning path. They have evaluated the models in online or offline mode, considering how much the students learn through the system. Some studies adopted the system performance evaluation methods, and few studies conducted a user satisfaction study.

Table 1 summarizes the related works study done as a preamble for the current research.

Table 1 Summary of related works from 2018 to 2021

The studies above show that the learning paths are sequences of LOs using different machine learning methods or algorithms. The major personalization parameters used are the learning style, learning time, the score obtained, prior knowledge. The LOs with more similarity with the learner characteristics are recommended. But none of the studies explore the likelihoods that the ‘difficulty’ parameter of LO can be different for different students. Also, the difficulty of an LO differs for the learner to varying points in the timeline as learning progresses. This gap is addressed in the current research paper. It is observed that most studies concentrate on selecting LOs according to learner needs or preferences, only very few works give attention to sequencing the topics. So, here we are using a two-layer model for sequencing the selected LOs based on knowledge relations.

A learning path is the linear list of LOs, organized based on their knowledge relation. So, the problem of recommending a learning path can be reduced to two issues, (1) To know the knowledge relations between topics (2) To suggest appropriate LO for a topic. The knowledge relations are obtained from the knowledge graph of the course. Following the relationships, the knowledge units are organized. And the LOs are selected based on the mapping between the LO and learner characteristics. The model predicts the learning time and score for every LO recommendation. The learner’s performance can be calculated in terms of learning time, scores obtained and the number of attempts needed, and satisfaction can be calculated based on the ratings given for each LO (Nabizadeh et al., 2017; Tarus et al., 2017).

Context: learning material and learner modeling

The learner model represents the learner’s behavioral information (learning preference, learning style), status information (learned LO, basic knowledge), and achievement (time for LO completion, score of LOs, number of attempts).

The learning path adaptation models deal with extracting relationships between the learning materials and the learner to select an appropriate learning path. As it is observed from the earlier studies that modeling the learning materials and learner are very significant (Tarus et al., 2017, Dorça et al., 2016). This section elaborates on the learner and learning material modeling used in this work. Ontologies are used as storage units for saving the learner-related parameters, learning object metadata, and learner activity log (Joy et al., 2021).

Learner model

To achieve learning adaptation, the personalized preferences and differences between the learners should be considered. In this study, we are focused on static and dynamic parameters for modeling the learners. The ontology’s main notion is a learner class, which is represented by various object properties or parameters for each student.

The learner is modeled using static and dynamic information, which are listed as follows:

Static Parameters: Learner identifier (LID), age, gender, stream of study, basic qualification, basic knowledge and learning style < active/reflective, visual/verbal, intuitive/sensitive, global/sequential > . These factors are explicitly fed into the system.

Dynamic parameters which are extracted from the learner activity log: ID of learning materials visited, learning time in minutes, score, rating, count of repeated attempts of the same material. These parameters act as implicit feedback to the recommendation process.

Also, the work considers additional information that is read from the learners as search topic information and the time availability to learn the topic. These values are not stored in the ontology but are stored in variables. The parameters are represented as the type values of the learner ontology (static) and learning log (dynamic) ontology classes.

Learning material model

The learning material metadata forms a significant part in the content recommendation. The learning content forms a hierarchical order (Nabizadeh et al., 2017). The four levels are course, lesson, concept, and learning object. The courses, represented as concept maps/graphs, form the topmost level and are often called subjects. A course can be covered using more than one lesson. The next level of abstraction is concept/topic. They are units of knowledge and learned by a learning object. One concept is mapped to one or more learning objects. These LO selection and recommendations are the major task in an adaptive learning environment (Belacel et al, 2014; Dharani and Geetha, 2013). A sequence LOs forms a learning path. Sequencing the LOs according to the order of the related concepts makes the sequencing more meaningful (Shi et al., 2020). Figure 1 represents the level of learning content abstraction assumed in this work.

Fig. 1
figure 1

Levels of abstraction of the learning content with example

The learning paths form two-layered topics/concepts as the outer layer and the associated LOs in the second layer. The cognitive linkage between the concepts is visualized using a directed graph named a concept graph or knowledge graph. The graph’s vertices form the concepts/topics, and the edges form their relationship (Benmesbah et al., 2021a; Zhu et al., 2018). Sequencing the concepts to develop a concept graph forms one part of the learning path adaptation problem.

The second part of the problem is selecting the appropriate LO for each topic. In Fig. 1, we can see that the Pop() topic is associated with two LOs. So, more than one LOs associated with any topic, and the adaptation of LO selection is significant in personalized learning path recommendation. As stated in the introduction, we have used the IEEE LOM schema for representing the LO metadata.

Out of the nine categories of metadata description, the study adopts the general field and education field as both have significantly contributed to the personalized LO recommendations in our previous works (Joy et al., 2021; Raj & Renumol, 2019). The title is a unique name given to identify the LO. Duration in minutes, the organization of the structure, type or format of the LO, the level of its interactivity and the type (active, expositive, mixed) and learning resource type of the LO forms the static information about the LO. The difficulty level is considered as an integer value in this work, which is dynamically computed for each LO depending upon the learning log. The values of all static metadata and the initial difficulty level of the LO are provided by the instructor. The LearningObject class of the ontology stores LO metadata. This study considers two types of LO: Expository LO, LOex, and Evaluative LOs, LOev (Nabizadeh et al., 2017). For each concept, the model tries to recommend one LOex and one LOev. LOex helps the learner learn the concept, and LOev evaluates the knowledge.

The graphical representation the ontology classes and their relationships are depicted in the Fig. 2. The dotted lines show the common data object shared by two types. The ontology is created completely in Java using a set of JENA APIs. The data is defined using RDF tools. Jena is a Java framework for creating Semantic Web apps. It includes rich Java libraries for writing code that works with different versions of RDF and SPARQL following W3C standards. Jena has a rule-based inference feature, an ontology reasoning engine based on OWL and RDFS ontologies.

Fig. 2
figure 2

Learner and learning object model

Learning path recommendation model

This section discusses the procedure to generate the learning paths as sequences of LO. This study uses a two-layered context/knowledge graph, KG, to represent the relation between the topics/KUs and the LOs, as shown in Fig. 3. We maintain a separate KG with specific starting and ending units for each lesson. The learner is asked to enter both the starting (SU) and target (TU) topics in a particular lesson they want to learn. The model creates a learning path connecting different KUs based on the first layer of KG. Also, from the second layer of the KG, an appropriate LO is chosen. The LOs selected are connected as a graph and are suggested to the students as a learning path (Algorithm 1, Fig. 4). Based on this observation, the learner’s interaction is recorded and later recommends LOs.

Fig. 3
figure 3

An abstract model representing the workflow

Fig. 4
figure 4

Comparison of predicted time for completing LOs and actual time taken

The Depth First Search (DFS) is used to find the sequence of KUs (Algorithm 2) and algorithm 3 proposes the method to rank the LOs associated with each KUs.

Path generation

To generate the learning path, learners choose the starting and target knowledge units/topics SU and TU, respectively. From the graph, the sequence of topics is obtained and stored as a DAG, S = (LT, RE), a subset of KG. S’s starting and ending topics are explicitly obtained from learners’ input. We assume that the learners are familiar with all of the predecessor KUs of the SU. Initially, SU is set as the first and only node of the learning path. A suitable LO is selected for SU, estimate the expected time and score, and attach the LO with SU. Algorithm 3 will help to choose apt LO for the learner. Algorithm 2 is invoked recursively to get all the possible paths from SU to TU and stores the result in the variable PathList[]. The model suggests the shortest path from this set of possible LO paths.

figure a

Get all possible paths

The recursive function AllPathsSelection() will generate all possible paths from SU to TU (Algorithm 2). To nodes in the paths are LOs, and appropriate LOs are selected using Algorithm 3.

figure b

Selecting suitable LOs

A set of LOs represents a topic or knowledge unit. Each LO is a self-sufficient module for learning a particular topic. In this study, we considered two types of LOs: explanatory LO (LOex) and evaluative LO (LOev). Explanatory LOs form the set of descriptive LOs in text, video, audio with different difficulty levels and interactivity. The Evaluative LOs forms the group of LOs that facilitates assessments of each knowledge unit. LOev is also of varying difficulty levels. The initial step in selecting a suitable LO for the required topic is to generate the top N recommendation list of LOs based on the learners’ preferences.

In the case of a new learner, natural learner groups are generated by running SPARQL queries against Learner ontology. The learning history of the existing learners included in the learner group is extracted by running the SPARQL query from the LearnerLog and Learning Material sub-ontologies. Learner similarity with the multivariate clustering method is used in this step, as explained in our previous works (Joy et al., 2019, 2021). For each LO in the top N list, the LO with a difficulty level compatible with the learner’s ability to learn is selected. The LO’s difficulty and the learner’s ability to understand the LO are computed dynamically using the Eqs. (1) and (2) respectively.

In an educational environment, the implicit feedbacks logged by the learners are the test scores, time taken to learn the LO, and the number of attempts made by the learner (Raj & Renumol, 2021). The ability of the learner to learn an LO and the difficulty of an LO is dynamically computed based on these logged parameters (Table 2).

Table 2 Parameters and description
$${A}_{ij}=\alpha \times \left(\frac{{S}_{ij}}{S}\right)+\beta \times \left(\frac{T-{T}_{ij}}{T}\right)+\gamma \times \left(1-{R}_{ij}\right),$$
(1)

Aij in Eq. 1 is the ability of the ith learner to learn the jth LO. Here the weighted sum of the score obtained, the time taken and the number of attempts made by the learner is used for updating the ability of the learner. S and T represent the maximum score obtained and the maximum time taken for the LO. Rij is the number of attempts taken by the ith learner to learn the jth LO. We have assumed that when learner attempts an LO multiple times, they find it more difficult. Hence the value is considered to have an inverse effect in finding the ability of the learners. Hence the Rij is subtracted from the desired maximum number of attempts for any LO, i.e.,1. Similarly, the higher the time taken for learning the LO, the learner is considered to be less able for learning the LO. Here also the maximum time taken recorded for that LO is subtracted from the Tij and weight is applied. The higher the score, the better is the ability, so Sij is taken as a positively correlated parameter. α, β, γ (Table 3) are the weights applied for the mean score, time and number of attempts respectively. The values are obtained by repeated trials for better prediction accuracy, using simulated and previously logged data.

Table 3 Optimal parameter values obtained through trials using simulated and historical data

Suppose the learner is not opting for the provided LO or is not completing the learning process. In that case, the score is adjusted to zero, and the time taken is admitted to the maximum assigned.

$$D_{i} = \delta \times Dj + \xi \times 1 - \left( {\frac{{\hat{s}_{j} }}{S}} \right) + \theta \times \left( {\frac{{\hat{T}_{j} }}{T}} \right).$$
(2)

The Eq. 2 represents the computation of difficulty of the jth learning object, Dj. The previous value of Dj is one parameter in computation. Also, the score obtained and time taken by the learners who have learnt the LOj is also considered for recomputing Dj. We have assumed that as the score obtained by learning the LO increases the difficulty decreases and when the time taken for learning the LO increases the difficulty also increases. The weights applied, δ, ζ, θ are obtained by conducting trials in the simulated and previously logged data.

Each LO is a combination of LOev and LOex. When the LO is initially stored in the LO repository, for each new LO, the instructor assigns a maximum score, maximum learning time, and difficulty level associated with the LO. When the jth LO is processed, each ith learner produces a new set of values as score Sij, time Tij and number of attempts Rij committed to learning the concept. The expectation about efficient learning is a better score in a shorter and fewer attempt (Nabizadeh et al., 2020; Raj & Renumol, 2021). So, we took the weighted sum of the three parameters to quantify the learner’s ability. A lower score and longer learning time make the LO more challenging to learn. Thus, the difficulty of LO is computed as the weighted sum of three parameters, the current difficulty, mean score obtained by all learners who learned the LO and mean of time taken by all learners who learned the LO.

The parameters named as α, β, γ, δ, ζ, θ are weights applied to the factors in the above equation where α + β + γ = 1 and δ + ζ + θ = 1. The optimal values that are observed using simulated data are shown in Table 3.

Consider the score obtained by ith learner for the jth LO is 5 where the maximum score is 10 in their initial attempt, by spending 2 min where the expected time is 3 min, then the ability is calculated as: Aij = 0.4 × (5/10) + 0.3 × (1- (2/3)) + 0.3 × (1–1) = 0.3 If the same happens in three attempts then the ability score will be − 0.3. Similar effects for score and time parameters too.

Suppose, the mean score obtained for the previously considered LO by all of its learners is 8, by spending an average of 3 min, and the LOs previously calculated difficulty is 0.6, then the new value will be: Dj = 0.5 × 0.6 + 0.3 × (1 − (8/10)) + 0.2 × (3/3) = 0.56, since the score decreased the difficulty increased.

The highlight of Algorithm 3 for Selecting LO suitable for a learner is the dynamic computation of ability and difficulty factors. The similarity of ith learner and jth LO, Sij is calculated using Euclidean distance, Eq. (3). The more similar LOs are selected from LOex and LOev lists for a KU, as explained in algorithm 3.

$${S}_{ij}=\sqrt{{\left({A}_{{i}_{j}}-{D}_{j}\right)}^{2}}.$$
(3)

So, if Aij is 0.3 and Dj is 0.56 the Sij = 0.26, which is the difference between the values; square is applied to make the values positive. Higher the Sij more the jth LO compatible for the ith learner, assuming a match between the ability of the learner and difficulty of the LO.

figure c

Experimentation and results

This section explains how the learning path recommendation model is evaluated. Also, it includes details of the experimentation process conducted and the results of the experimentations. We have created concept/knowledge graphs based on a C Programming and Data Structure course comprising 468 learning items and 1065 relationships. Another KG is made for Data Mining Course with 155 LOs and 254 connections. The distribution of LOs that are used in this study is shown in Table 4. Each LO is characterized by IEEE LOM parameters and the table shows the generic nature and count of the LOs used in the study. Both the KGs are fed into the system. The learning objects are crawled from various educational websites and extracted (Joy et al., 2019, 2021). The concepts are mapped based on the CUSAT syllabus for the courses CS201B Computer Programming, CS405 Data Structures and Algorithms, and CS604 Data Mining.

Table 4 Distribution of LOs

The relationships among the LOs are manually established by three instructors based on the course syllabuses and are verified by an expert. The LOs are fed into the system, and the initial fields are annotated by a group of ten instructors who teach the undergraduate courses in Computer Science and Engineering. Each learning object is annotated by at least three instructors. The majority decision is taken into consideration for fixing the initial values for discrete parameters of each LOs. The mean of three values is considered for fixing the initial values for continuous parameters. If there is no common agreement, the LO is passed to the subject expert for final decision. The initial values given to every field describing the characteristics of the LOs are anticipatory values given by the instructors. These values helped in the initiation of recommendation. As elaborated in the previous sections, when the recommendation process progresses, the values for each LO field is updated according to the learner feedback.

According to the learner’s ability to learn and preferences, the current study suggests adaptive and personalized learning paths. The learners should analyze the quality and usability of the output. A total of 96 learners evaluated the model. The participants are enrolled for undergraduate Computer Science and Engineering programs in the two Indian state universities: APJ Abdul Kalam Technological University (KTU) and Cochin University of Science and Technology (CUSAT), Kerala, India. The experiments are conducted between February 2021 and September 2021. The experimentation was not a continuous process for all days. Each batch of experimentation was conducted for a batch of 10 students. Every participant is asked to join the learning process’s three phases: (1) pre-test, (2) learning, and (3) practice. The pre-test and learning style identification is made at the entry into the procedure. The pre-test identifies the knowledge of the learner. In the learning phase they searched for a term, say “stack push”, and the learning path is recommended to them. In the practice phase, they are asked to answer a maximum of three questions based on their study, which evaluates their learning performance. Three instructors monitored the student’s activities and guided them throughout the procedure. All tests are conducted with questions of different difficulty levels based on Bloom’s Taxonomy (Sosniak, 1994; Amstrong, 2016). In the experiment, it is verified that (i) the learning path generation algorithm proposed is better (minimizing the difference of actual and predicted score/time/rating values) than the baseline models (ii) more diverse paths are generated by the proposed approach than the LO-Learner mapping-based algorithm (iii) the learner’s satisfaction is progressing with the recommended learning path. The proposed model is tested against three other models. We have named the existing models considered as SeqSt, SeqDyn, PathSt for better comprehension while discussing the experiments. PathDyn is the name that is given for the proposed model. The models are explained as:

  • SeqSt: The model suggested a sequence of learning materials based on the static values of the parameters used (Tarus et al., 2017). Here, the ontology-based domain modeling method is explored. A personalized sequence of LOs is recommended based on historical library data. Learners previously rate the annotated LOs. This rating is used as the parameter to select LO and user preferences.

  • SeqDyn: This model explores the dynamicity of the learner parameter. The changing values of learner satisfaction are considered here. But the sequencing of LOs is not done considering the knowledge relationship between the learning materials. We have implemented the SeqDyn based on previous works (Joy et al, 2021; Raj & Renumol, 2019). The change in student performance measures is incorporated with the basic model to evaluate the influence of dynamic parameters.

  • PathSt: This model uses a knowledge graph as the structure used to sequence the LOs, thus providing a cognitive linkage between the learning objects (Nabizadeh et al., 2020). The base model tries to optimize suggestions based on the historical or available data of the learner. The difficulty parameter of the LO is considered static here.

  • PathDyn: This proposed model considers a curated knowledge graph connecting the concepts/topics. The LOs are associated with these topics. LOs are selected based on the historical/available data and filtered further optimized by the ability of a learner to learn the LO of a particular difficulty level. Unlike other models here, the learner’s ability and the LO’s difficulty are adjusted based on their academic achievement, time invested to learn, the number of attempts they made, and ratings.

The primary aim of the experimentation is to compare the effectiveness of content sequence recommenders and path recommenders. The sequence recommenders SeqSt and SeqDyn recommend learning object sequences by analyzing the learner log. The learning path recommenders, PathSt and PathDyn suggest a learning path considering the learner log and a pre-designed knowledge graph. Here an inherited cognitive linkage is established between the suggested LO lists. The SeqSt and PathSt uses parameters with static values whereas the dynamicity of the parameters are used by the SeqDyn and PathDyn models. So two different aspects are experimented here, 1. The effectiveness of learning sequence and learning path 2. The effectiveness of static and dynamic parameters. The rest of the section elaborates on the experiment conducted and the result obtained.

A control experiment is conducted to find the model’s accuracy to predict the learning time and score of the learners. The participants are divided into four groups, balancing their prior knowledge and learning style for control experiments. They were asked to conduct multiple topic searches on the available lists. For each iteration in the learning phase, the model predicted the time required to complete the learning path by the learner. Also, the actual time spent by the learner on each learning path is recorded. In the practice phase, both time and score are predicted and observed. Root Mean Square Error (RMSE) as in Eq. (4) and R-Squared (R2) Error as in Eq. (5) are used to derive interpretations from the predicted score and time against the observed values in the practice phase. The number of topic searches by the learners varies from 8 to 11 in the experimentation phase, so the first eight searches are considered here. The results are observed to be stabilizing by the first 8 iterations.

$${\text{RMSE}}=\sqrt{\frac{1}{m}\sum_{i=1}^{m}{\left({x}_{i}-{y}_{i}\right)}^{2}},$$
(4)
$$R2= 1-\frac{\sum_{i-1}^{m}{\left({x}_{i}-{y}_{i}\right)}^{2}}{\sum_{i=1}^{m}{\left(\widehat{y}-{y}_{i}\right)}^{2}},$$
(5)

where m is the number of observations in each iteration, xi is the predicted value, yi is the observed value and \(\widehat{y}\) is the mean of the observations.

RMSE gives the standard deviation of the residuals (prediction error) between the observed, xi and predicted values, yi. Here we have squared the residuals, took the mean and obtained the root of the value.

The results of experimentation are given below in Figs. 4 and 5. Each experimentation are considered for the first 8 iterations and x-axes of the figures represent the times of iteration.

Fig. 5
figure 5

Comparison of the predicted score on completing LOs and actual score obtained

Figures 4 and 5 shows that the proposed model is better in maintaining an accurate prediction of learning duration and expected score. The second-best model is observed as the sequence generating model which is modeled with dynamic parameters. The proposed path recommender shows an average of 30% more accuracy in predicting duration and 27.8% more accuracy in predicting the expected score. The time prediction is crucial as the learner wishes to complete the learning process in an available time depending on their learning goal (Zhu et al., 2018). The predicted score helps select the LO that can help the struggling learners perform better (Jdidou et al., 2021).

The Adaptivity, ADP, measures how much the recommendations suit a learner’s preferences. In the current study, Euclidean distance is used for finding the similarity between a learner and recommended LO is used to measure the adaptivity (Meng et al., 2021). The lesser the value, the more chances for the learner to learn the topic (Plass & Pawar, 2020). The adaptivity is calculated as the mean of similarity measures of all recommendations in that iteration (Eq. 6).

$$AD{P}_{t}=\frac{\sum_{i,j=1}^{N}{Sim}_{ij}}{N},$$
(6)

where N is the total number of learners, LO combinations in the iteration t. Figure 6 shows the adaptivity measure obtained for each model.

Fig. 6
figure 6

Similarity between the recommended LO and learner characteristics

Diversity, DIV of the learning path: Ensures that the learning path suits the learner’s needs is least likely to repeat (Liu et al., 2018, Meng et al., 2021). More diverse paths need to be generated by the model, and different classes of learners should get different learning paths also.

$${\text{DIV}}= \frac{\sum_{i\ne j}^{N}1-\left(\frac{{p}_{i}\cap {p}_{j}}{L}\right)}{N-1},$$
(7)

where DIV represents the diversity of the learning paths recommended and P represents each path. L is the length of a path, and the model suggests n number of total paths. Figure 7 shows the path diversity obtained for each model.

Fig. 7
figure 7

Diversity in recommended LO for different learners

Participants are asked to rank the recommendations on a scale of 1 to 5 (1—poor, 2—average, 3—good, 4—very good, 5—excellent). This feedback from the learners is used to calculate the learner’s satisfaction. The total recommendations made in the entire learning process comprises to 4285 LO ratings. From Fig. 8, the graph’s X-axis denotes the rating share in percentage and the Y axis represents the rating level 1–5. We can observe the rating share of each model. For example, the 27% of LOs rated 1 are recommended by SeqSt, 29.8% by SeqDyn, 27% by PathSt, and 16.2% by PathDyn. Similarly, considering the 5 rated LOs, the rating share is observed as SeqSt -17.39% SeqDyn-24.63% PathSt—26.08% PathDyn-31.88%.

Fig. 8
figure 8

Rating share of each model

Again, suppose the rating levels are greater than two are considered. In that case, the total ratings obtained by each model are shown in Fig. 9. The values are normalized on a scale of 0–100.

Fig. 9
figure 9

Number of Ratings obtained for the satisfactory level > 2, normalized to 0–100 scale

Discussion

This section discusses the results obtained from systematic experimentation to answer RQ1 and RQ2. The RQ1 is a decision problem: Can we accurately predict learning duration and expected score during the learning path recommendation process? From the controlled experimentation performed, we are getting a positive answer for the RQ1. The learners are asked to search and learn various topics available in the repository in the process. RMSE and R2 measures are used to evaluate the regression performance of the model. Figure 4 shows the comparison between the learner’s predicted and actual time taken through 8 different iterations.

Here we can observe that the proposed model, PathDyn, steadily improves prediction accuracy. The average RMSE values are almost decreasing in subsequent iterations for the proposed model. But the baseline models where the static characteristics are alone used for predicting LOs (SeqSt and PathSt) have comparatively higher error rates in predicting the duration of the learning process. The heterogeneous set of learners cannot follow the predictions by the system that uses static learner parameters (Tseng et al., 2008). In contrast, the proposed model uses the time taken log of the learners as one parameter to compute their ability. The time factor is also considered while updating the difficulty of each LO (Meng et al., 2021). So, the model can learn better with an increasing number of iterations and interactions, reducing the error and improving the predictions.

Similarly, a score log is also used while computing the learner’s ability. The score is also considered as a parameter in the computation of the difficulty of each LOs. From Fig. 5, it is evident that the difference in the predicted and observed values of the score is decreasing with an increase in iterations. Also, the R2 measure is showing promising results with the proposed model. The baseline model does not offer any pattern in the relation between the learners’ observed and predicted score measures. We have compared the two models which function on dynamic parameters (SeqDyn and PathDyn), analyzing the real-time implicit learner data. From the result analysis, PathDyn better predicts the score and learning duration. Here, the advantage of the knowledge-based sequencing of LOs forming path helped reduce the error rates (Tarus et al., 2018; Wu et al., 2020).

The RQ2 tries to evaluate the adaptive and dynamic nature of the proposed model, and the question is: What are the factors contributing towards generating adaptive and diverse learning paths in an e-learning environment? The answer to this question is the dynamic parameters. From Fig. 6, it is understood that the adaptivity of the model increases as the interactions increase. With more iterations, the system can log more information about the learners, and based on that, more adaptive LOs can be recommended. The diversity factor is also used to measure recommendations’ adaptive and dynamic behavior. From Figs. 6 and 7, we can observe that the adaptivity and diversity of the recommended learning materials are better with the models designed with the addition of dynamic parameters too. The static parameters help find the learning object that suits the learner’s preferences. But, naturally, the learner’s performance varies as the learning process progresses. The dynamic parameters are needed to be analyzed to recommend LO adaptively to this change in performance. The dynamic parameters considered in this study are the real-time learning duration, score obtained, number of attempts taken to study an LO, and the list of learned materials. These parameters are implicitly logged and analyzed in real-time scenarios. The implicitly collected learner log helps in adapting to the learner’s changing needs, and the integration of implicit and explicit parameters makes the recommendations more adaptive (Gomede et al., 2021; Xie et al., 2019). Thus, we can see that from the results of the experimentations, RQ1 and RQ2 are answered.

The learner satisfaction analysis is also done as discussed in Sect. 6. From the results, it is evident that the learners show more inclination toward the PathSyn generated LOs. The count of LOs recommended by the PathSyn model, which is rated with 3,4 and 5 levels, is more than all other models. Since the ontology model is used for representing the domain knowledge and involved in selecting LOs the cold-start issue is handled by the model (Joy et al., 2019, 2021).

Limitations

Even though the model generates a learning path suitable for the learner’s style and ability, there are few limitations also. The limitations observed in the proposed model for learning path generation are as follows:

  • The linkage between the LOs is made based on the proficiency of the subject experts only. The initial difficulty level and time expected to learn an LO are marked by the experts, so in the cold-start phase, these values are not personalized.

  • The performance of learning path models cannot be compared as the behavior of learners changes dynamically. The data cannot be reused either.

  • We have used simulated data for early training of the system to find the constraint parameters.

  • The number of topics and LOs are limited in this study.

  • The participants are assumed to have learned all the previous topics.

  • Few learners made errors in data entry, and we needed to discard their records completely

Conclusion and future work

This research article presents a model for generating learning paths suitable for learners’ preferences and abilities. The present study tries to solve the learning path adaptation problem by exploring the knowledge relationship between the learning resources. The learner, learning materials, and learner log data are represented as classes in the ontology model. The data objects of these classes are comprised of static and dynamic parameters defining each class. The study focuses on analyzing the dynamic parameters such as the time taken for learning, the obtained score, the number of repeated attempts, and the learning resource rating. Based on the analysis, the ability of the learner to learn a particular learning material is computed in real-time for every recommendation. Also, the difficulty level of the learning material is adjusted based on the learner’s performance and LO rating. The LO is selected in two steps 1. Generate a list of LO using Collaborative filtering exploiting the similarity of learners 2. From this list, select the LOs that best match the learner’s current ability. The LOs are sequenced using a concept graph for the generating the learning path. As the model trust more on the implicit feedback represented as the dynamic parameters, the ability to predict learning duration and expected score is progressing as with the learning process. Thus, recommending more adaptive and diverse learning paths. The comparison results with three existing models show a better performance from the proposed approach with an average accuracy rise of 30% in learning path prediction based on the expected duration of learning 27.8% in expected score prediction with the second-best performing model. The rating levels indicated the enhancement of learner satisfaction and experience with a rise of 25.5% when comparing the rating share with the second-best model. Ninety-six undergraduate Computer Science and Engineering students participated in the study which involved 623 learning materials from C Programming, Data Structures, and Data Mining courses.

Even the proposed model shows progressive results, the progression is slower. More experiments are planned with parameters such as students’ cognitive ability, engagement, and procrastination to recommend learning paths to learners (Agnihotri et al., 2020; Farrell et al, 2019; Raj et al., 2021; Shimada et al., 2018). Also, we have plans to incorporate the model with the existing learning management system for a better e-learning experience. In the current situation of sudden shifting between on-campus and online modes of education caused by COVID-19 pandemic, a more personalized LMS will benefit toward the student engagement, performance and satisfaction (Clark et al., 2021; Patil & Naqvi, 2020).