369 Ajel20249 (1) 1 21
369 Ajel20249 (1) 1 21
Donggil Song
ABSTRACT
The goal of this research is to (1) identify the status and development of AI and ML-based learning
support systems and their impact on human learning, with a specific focus on techniques employed in
previous research, and (2) demonstrate the process of designing a learning support system using AI.
Artificial intelligence (AI) and machine learning (ML) technologies have received attention in
education. The existing research on AI in education is examined, considering the implications of its
application in research. Noteworthy ML techniques from the literature are explained, followed by a
discussion on leveraging AI and ML technologies to enhance learning support. Additionally, with
consideration of both front-end and back-end approaches,a framework for incorporating AI into
education is proposed. Subsequently, a learning design model, Self-regulated Learning with AI
Assistants (SLAA), is suggested for addressing the objectives of AI-based learning support system
design. The categorization of AI and ML techniques in education research reveals nine types,
including supervised learning, mining approaches, and Bayesian techniques. The exploration
illustrates how these techniques can be employed in designing a learning support system. This paper
provides an empirical overview of AI in education, addresses technological and pedagogical
considerations for developing personalized and adaptive learning environments, and outlines the
challenges and potential future research directions.
Keywords: Artificial intelligence, Learning design, Learning system design, Machine learning, Personalized learning, Self-regulated
learning.
DOI: 10.55284/ajel.v9i1.1024
Citation| Song, D. (2024). Artificial intelligence for human learning: A review of machine learning techniques used in education research and
a suggestion of a learning design model. American Journal of Education and Learning, 9(1), 1–21.
Copyright:© 2024 by the author. This article is an open access article distributed under the terms and conditions of the Creative Commons
Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Funding: This study received no specific financial support.
Institutional Review Board Statement: Not applicable.
Transparency: The author confirms that the manuscript is an honest, accurate, and transparent account of the study; that no vital
features of the study have been omitted; and that any discrepancies from the study as planned have been explained. This study followed all
ethical practices during writing.
Competing Interests: The author declares that there are no conflicts of interests regarding the publication of this paper.
History: Received: 22 November 2023/ Revised: 3 January 2024/ Accepted: 15 January 2024/ Published: 12 February 2024
Publisher: Online Science Publishing
1
URL: www.onlinesciencepublishing.com | February, 2024
American Journal of Education and Learning, 2024, 9(1): 1-21
1. INTRODUCTION
Artificial intelligence (AI) technology has garnered attention in education due to the abundance of big data
collected from education-related systems and the capabilities of powerful chatbots. Since the late 1990s, intelligent
tutoring systems have addressed student needs by representing instructional decisions through interactions with
the learner (Beck, Stern, & Haugsjaa, 1996). Other approaches have been explored, including automated essay-
scoring systems (Shermis & Burstein, 2003) and adaptive simulation-based military training (Department of the
Army, 2011). Different procedural interpretations of big data give the system unique meanings, knowledge, and
intelligence, often determined dynamically by the learning context. Recently, generative AI has introduced new
possibilities and simultaneously raised serious concerns in education. As such, a wide variety of AI types,
approaches, and methods can be readily found in education. The characteristics of AI, such as customization and
adaptiveness, offer a potential solution for enhancing current learning designs to better support learners’ diverse
needs.
While numerous research reports on the use of AI in education exist, it can be challenging to discern which
aspects of a learning environment involve AI and how to integrate AI into the development of learning support
systems. Importantly, few studies have explained or suggested specific techniques or a design model that education
researchers can apply. To address these issues, we aim to (1) illustrate specific machine learning (ML) techniques
adopted in previous research, (2) clarify the use of AI in education, and (3) propose a design model for a learning
support system using AI. First, we explain the techniques reported in the literature and suggest a framework for
incorporating AI into education. Second, we present a learning design model to address the objectives of AI-based
learning support system design. Lastly, we discuss the challenges of AI in education and outline future research
agendas.
2. RELATED WORKS
2.1. Machine Learning Techniques for Learning Design
Machine learning (ML) refers to a study or field of techniques or algorithms in which computer programs learn
something (e.g., data patterns) from data to perform a task independently. ML systems can learn primarily from big
data without being explicitly programmed for a specific task. We have reviewed various ML techniques relevant to
educational purposes and summarized them in Table 1.
2
URL: www.onlinesciencepublishing.com | February, 2024
American Journal of Education and Learning, 2024, 9(1): 1-21
the data contains noise (i.e., meaningless, unstructured, and/or corrupted data that the machine cannot understand)
and missing feature values, data preprocessing is necessary. During this phase, irrelevant and redundant features
are removed. To execute the ML process successfully, specific and appropriate ML techniques must be chosen.
Below, we briefly introduce examples of supervised ML techniques.
Naïve Bayes. This is built on Bayes’ theorem for probabilistic classification and functions as a supervised
learning approach. Recognized for its simplicity, it stands out as one of the fastest classification algorithms in terms
of running time; it is particularly suitable for real-time tasks. If a learning designer is dealing with a large dataset
containing a small number of variables, Naïve Bayes proves to be a worthwhile technique.
Decision Tree. As an intuitive technique, Decision Tree visualizes results in a structure reminiscent of a tree.
This tree-shaped representation helps in understanding how the results were calculated. For instance, in a dataset
containing information such as students’ age, gender, GPA, and the successful completion of an exam, each feature
(e.g., age, gender, GPA) can be considered a branch of a tree. The label of the data (the successful completion of the
exam) becomes the conclusion, analogous to a leaf in the tree.
Random Forest. This technique utilizes multiple Decision Trees. Each tree’s vote for class prediction, along
with the prediction results, is used to determine the class that received the most votes. This can generate ensemble
predictions that outperform the individual Decision Tree’s prediction when there is a low correlation among the
trees. In cases where Decision Trees lead to overly tall structures, called deep trees, irregular patterns and
overfitting issues may arise.
Support Vector Machine (SVM). SVM serves both classification and regression analysis purposes. It identifies
decision boundaries (separating hyperplanes) in an n-dimensional space to classify data points. For instance, consider
a 2-dimensional space with math scores on the x-axis and reading scores on the y-axis, each point representing a
student’s scores in both subjects. If two students’ data points are not in the same spot, SVM determines the
hyperplane that separates them. Through training, SVM seeks the optimal line with the maximum distance between
data points. In the case of three input features, the hyperplane becomes a two-dimensional plane, and so forth.
K-Nearest Neighbor. This technique assumes that similar data points are located in their neighborhood
(proximity). It calculates the distance between a test sample and specified training samples on a graph to determine
if the test sample is close to a specific class within the training samples. The parameter K denotes the number of
nearest neighbors. If K is set to 1, a data point is assigned to the class of its single nearest neighbor. The results and
accuracy may vary depending on the chosen value of K.
3
URL: www.onlinesciencepublishing.com | February, 2024
American Journal of Education and Learning, 2024, 9(1): 1-21
ensures that the sum of the squared distance between the data points and the cluster’s centroid is minimized.
Hierarchical Clustering. This is an alternative approach to conventional clustering for identifying groups in the
dataset by building a hierarchy of clusters. Initially, each data point is treated as a separate cluster. The algorithm
then identifies two close clusters and merges them. The distance between two clusters is calculated based on the
length of a straight line between two observations (clusters). This process continues until all clusters are merged,
resulting in a hierarchical relationship between the clusters represented as a dendrogram, a tree-shaped diagram.
Ripley’s K function. This function assesses the structure of underlying patterns in a dataset by analyzing the
spatial point pattern of data on event locations. It determines whether a dataset is dispersed, clustered, or randomly
distributed (i.e., detects deviations from spatial homogeneity). This technique has been employed to describe a set of
data point locations, examine research hypotheses about data point patterns, and estimate parameters in a spatial
point process model.
4
URL: www.onlinesciencepublishing.com | February, 2024
American Journal of Education and Learning, 2024, 9(1): 1-21
5
URL: www.onlinesciencepublishing.com | February, 2024
American Journal of Education and Learning, 2024, 9(1): 1-21
technique accumulates training examples through trial and error to optimize long-term rewards. It is employed in
creating agents (e.g., ChatGPT [Generative Pre-Trained Transformer]) that perform actions in an environment,
receiving rewards based on the agent’s status when it acts.
6
URL: www.onlinesciencepublishing.com | February, 2024
American Journal of Education and Learning, 2024, 9(1): 1-21
likelihood of the next item. This method is particularly useful for text similarity analysis, often using Cosine
Similarity. In this case, N-gram can be used for feature extraction, and Cosine Similarity calculates the cosine of the
angle between the two vectors (e.g., two documents).
7
URL: www.onlinesciencepublishing.com | February, 2024
American Journal of Education and Learning, 2024, 9(1): 1-21
datasets. It analyzes series of repeated actions in a dataset, comparing data with process models to monitor
compliance, detect deviations, and predict discrepancies.
Data Stream Mining. This method extracts data structures from continuous data points and handles a large
amount of real-time data (data changing over time). The term stream refers to continuous, massive sequences of
data items. This technique is useful when the complete dataset is unknown and/or the data is not stored.
8
URL: www.onlinesciencepublishing.com | February, 2024
American Journal of Education and Learning, 2024, 9(1): 1-21
encompass information retrieval systems and personal assistants, such as Internet of Things interactive tools that
communicate directly with the user. The examples given by Oh, Song, and Hong (2020) and Song, Rice, and Oh
(2019), are typical ones for conversational agents. The anticipation is for AI systems to undertake tasks currently
performed by human instructors or teachers, tasks that are both time-consuming and costly.
As illustrated in Figure 1, we present a learning design model that elucidates the adoption of AI techniques
9
URL: www.onlinesciencepublishing.com | February, 2024
American Journal of Education and Learning, 2024, 9(1): 1-21
within the front-end and back-end frameworks. The primary functions are proposed to be twofold: personalization
from a learning analysis viewpoint (the back-end approach) and SRL scaffolding (specifically, metacognition
support) from the interaction design perspective (i.e., the front-end approach).
10
URL: www.onlinesciencepublishing.com | February, 2024
American Journal of Education and Learning, 2024, 9(1): 1-21
systems. Knowledge modeling, along with tracing, could be essential for the monitoring process. Lester et al. (2013)
proposed the use of the Dynamic Bayesian Network for knowledge modeling and tracing during interactive
narrative experiences in an educational game. This model helped dynamically update probabilistic beliefs about a
learner's understanding and knowledge of the learning content, estimating probability values representing students’
knowledge, which were used to predict actual performance on posttests.
Learning Preference. Learning preferences can serve as predictors of learning performance. Dorça et al. (2013)
suggested the use of reinforcement learning for automatically detecting learners’ preferences. Assuming that the
four types of students’ learning preferences (i.e., reflective, intuitive, verbal, and sequential) change over time, their
learning support system collects learners’ performance values and calculates the reinforcement value. For instance,
higher performance updates the reinforcement value, and vice versa. Stronger reinforcement increases the
probability of the current learning preference, while lower reinforcement reduces the probability of the current
preference. The approach successfully detected, monitored, and adjusted students’ learning preferences to support
their learning.
Evaluation. In addition to performance and preference monitoring, learning paths and behaviors require
evaluation.
Learning Paths and Behavior. Chen (2008) delved into personalized learning paths in an e-learning system,
where each course comprised elements like introductions, pretests, topics, summary modules, and posttests.
Learners’ paths (e.g., topic sequence, test types, the difficulty of topics) varied depending on the course sequence.
They represented the course sequence as a chromosome in the Genetic Algorithm, generating personalized learning
paths through reproduction, crossover, and mutation processes. Based on pretest results, the system suggested
appropriate difficulty levels until the student successfully completed their learning. Similarly, mining techniques
can assess learning behavior. Taub and Azevedo (2018) analyzed students’ metacognitive behavior in a game-based
learning environment. Grouping participants based on emotions and game performance, the study identified four
groups: less efficient–low emotions, less efficient–high emotions, more efficient-low emotions, and more efficient-
high emotions. Learning patterns were extracted from log files, indicating whether students’ game performance was
relevant, partially relevant, or irrelevant to their learning. Each student’s sequence of these relevancies within the
gameplay, such as partially relevant-irrelevant to irrelevant-relevant, served as input data for Sequential Pattern
Mining analysis. After identifying noticeable patterns, Differential Sequence Mining was employed to check for
significant sequence differences between groups. It was clear from these methods that the four students’ groups’
hypothesis-testing behavior patterns were not exactly the same. Thiswasn’t clear from their first attempt at
multivariate analysis of variance.
11
URL: www.onlinesciencepublishing.com | February, 2024
American Journal of Education and Learning, 2024, 9(1): 1-21
Schneider and Pea (2015) formulated their N-gram Model by calculating unigram, bigram, and trigram
probabilities, revealing frequently used words in participants’ discussions. The researchers assessed discussion
coherence using Cosine Similarity, identifying how much a student group discussed a topic by building on their
conversation partner’s ideas and arguments. They successfully identified differences in coherence scores between
student groups. Similar approaches can be found in research on online discussions (e.g., (Albatayneh, Ghauth, &
Chua, 2018; Sullivan & Keith, 2019)).
Behavior Data. Harley et al. (2013) explored the existence of different student clusters and the relationship
between learner behavior in an intelligent system and their performance to strengthen adaptive support. Data,
including system-related (e.g., mouse clicks, keyboard entries, facial expressions, diagrams drawn, eye-tracking
information) and survey data, was collected from a system teaching the human circulatory system. Using
Expectation-Maximization, the researchers identified three distinct clusters based on twelve selected features. This
technique was also employed to identify learners’ voices, as seen in learner authentication systems (Kamaraj,
Nidhyananthan, & Sundaram, 2019).
Context and Environment. Regardless of the learner’s behavior and performance, the learning context might
undergo changes. The dynamic interactions between this change and the learner should be analyzed; however, few
studies have been conducted. Further research is required.
Scaffolding. Scaffolding can be offered to learners with the consideration of the learning content, performance,
and motivation.
Learning Content. Morsy and Karypis (2019) devised a course recommendation framework using students’
grades. They defined a good course (grade equal to or higher than GPA) and a bad course. Each student’s course
information was converted to a previous-subsequent co-occurrence frequency matrix for Singular Value
Decomposition (SVD). The system recommends good courses, aiming to help students maintain or improve their
overall grades. SVD demonstrated reasonable performance, particularly in predicting good courses. If learning
resources are well-categorized as a form of formal instructional courses, resource recommendation systems could be
beneficial in supporting learners.
Cetintas et al. (2010) proposed a classification model to estimate the difficulty level of math problems. They
used an SVM classifier for the initial classification, considering individual features of math problem sentences. The
individual features comprised a bag-of-words representation; for instance, “Deep learning is machine learning” can
be represented as {“deep”: 1, “learning”: 2, “is.”: 1, “machine”: 1}. The bag-of-words was used to train the SVM
model to find the best hyperplane classifying problem difficulty levels, a fundamental aspect of learning content
scaffolding.
Dascalu et al. (2015) developed a learning material recommender system based on learning style matching
algorithms. The system analyzes the user profile (e.g., learner interest, education, nationality, and the result of
learning style questionnaires) and provides recommendations for learning material shortcuts and learning tools
depending on learners’ profiles and preferences. Jaccard Similarity was employed to calculate the similarity between
recommendation items. For example, when recommending a learning tool to a learner, the system computes the
similarities (Jaccard index or coefficient) between tools liked by other learners. Learning content recommendation
systems have also been explored in online learning systems contexts, such as LMS integration (De Medio,
Limongelli, Sciarrone, & Temperini, 2020) and MOOCs (Massive Open Online Courses; Xiao, Wang, Jiang, and Li
(2018)).
Performance. Scaffolding is necessary when the learner’s underperformance or misunderstanding is detected.
Pelaez et al. (2019) applied data mining to identify at-risk students using Latent Class Analysis. The researchers
12
URL: www.onlinesciencepublishing.com | February, 2024
American Journal of Education and Learning, 2024, 9(1): 1-21
collected data that included students’ demographics, admission (e.g., SAT [the Scholastic Assessment Test] scores,
high school GPA [grade point average]), and academic records (e.g., final grades). They applied Latent Class
Analysis with the Random Forest technique and identified three clusters of at-risk students, which can be used for
performance-based scaffolding.
Motivation. Motivation scaffolding supports learners’ self-efficacy, self-directed aspiration, and perceptions of
the value of learning activities (Belland, Kim, & Hannafin, 2013). Since motivation scaffolding contributes to learner
retention (Ludwig-Hardman & Dunlap, 2003), the system needs to identify the features impacting learner retention.
Besides, learners’ motivation does not solely stem from learning tasks but also from their self-regulation processes
(Zimmerman, 2002). Still, few studies have explored how to scaffold learners’ motivation through AI techniques.
13
URL: www.onlinesciencepublishing.com | February, 2024
American Journal of Education and Learning, 2024, 9(1): 1-21
Dillenbourg (2019) analyzed learners’ interaction log data (e.g., lecture video access, assessment submission) of a
MOOC in undergraduate engineering. Focusing on the learners’ behavior sequences, their Hierarchical Clustering
analysis yielded distinct clusters. Hao et al. (2015) also used Hierarchical Clustering to identify learning patterns in
a game-based learning environment.
Rasch Model. Waters et al. (2014) investigated learner collaboration patterns using the data of students’ right-
and-wrong responses. In their Rasch Model, a learner is characterized by a learner ability variable (i.e., single latent
ability parameter), and the difficulty of questions is modeled. The model indicated the probability of a learner’s
correct answers. The performance on a large number of consecutive simple tasks not only demonstrated learners’
achievements but also generated learning patterns characterized by a probabilistic model.
Process Mining. Sonnenberg and Bannert (2016) investigated the sequence of students’ learning activities
measured by a think-aloud method. Undergraduate students participated in online learning, which provided
metacognitive prompts to support students’ SRL. They identified both effective and non-effective metacognitive
prompts. Using Process Mining, they found the sequential order of learning activities that provides detailed
information about learning behaviors. Since the learning path follows a sequential order, Process Mining would be
effective for constructing a learner model that incorporates learning path factors.
Network Analysis. Fincham et al. (2018) investigated the student discussion data collected from online courses.
They employed the concept of social ties (i.e., depending on students’ concept definitions, it could involve co-
participation or direct replies in discussion) in their Social Network Analysis. Based on the identified social ties,
they examined discussion nodes (e.g., degree centrality, closeness, betweenness) and network-level structures (e.g.,
density, diameter, path length). The results revealed that different definitions of social ties produced distinctions in
the structural aspects of discussion networks. Similarly, Gruzd et al. (2016) used Social Network Analysis to
explore learner discussions in social media, revealing interaction patterns that could be utilized for learning pattern
profiling.
Learners’ chatter information can reveal their learning paths and patterns. Siebert-Evenstone et al. (2017)
collected student chat log data in a collaborative online learning environment, segmenting it by utterance. Using
Epistemic Network Analysis to analyze the co-occurrence of concepts within discourse data, they identified the
structure of connections in the coded student chat data. Another type of network analysis method examines
learning behaviors. Hecking et al. (2014) investigated patterns of learners’ resource usage (e.g., videos, scientific
literature) collected from the event log in online courses. They used Clique Percolation to identify subgroups and
found different resource usage patterns between different types of online courses. This method is useful for
identifying patterns of learner-system interaction, a crucial component of learning profiling.
Advanced Statistical Techniques. Pattern analysis is not just for learners but can also be applied to instructors. Xu
and Recker (2011) investigated clusters of teacher behaviors (e.g., browsing library resources, creating materials,
and sharing activities) when using a digital library service. The researchers used Latent Class Analysis, providing
different types of probability statistics, including the variables’ significance. They identified multiple clusters of
teachers, such as isolated users, goal-oriented teachers, and classroom practitioners.
When building a learner model, emotional aspects can be considered. Althoff et al. (2016) investigated
psychotherapy counseling conversations about crisis intervention (e.g., anxiety, depression, suicidal thoughts)
through text messages. Their dataset includes over 3.2 million messages. They used HMM to identify progression
through counseling stages, specifically to capture the dialogue structure. The results revealed five stages of
conversation progress: introduction, main issue and clarification request, problem discussion, actionable strategies,
and wrap-up. Additionally, for time-varying data, such as learners’ decisions in online learning environments,
14
URL: www.onlinesciencepublishing.com | February, 2024
American Journal of Education and Learning, 2024, 9(1): 1-21
researchers have used Stochastic Gradient Descent for data analysis (e.g., Kassak et al. (2016)). Some studies have
also employed temporal and sequential data processing methods, such as analyzing learner behaviors of body
motions, gestures, and gaze (Andrade et al., 2017; Andrade, Delandshere, & Danish, 2016), student behavior in
MOOCs (Geigle & Zhai, 2017), and learner interactions in intelligent systems (Shen et al., 2018).
Cognitive Strategy Profiling. Ripley’s K Function. Mallavarapu et al. (2015) analyzed students’ learning
activities for spatial reasoning problems in a game. Ripley’s K Function was employed to handle the spatial metric,
quantifying the density of points at varying scales of distance. They examined students’ exploration of two-
dimensional spatial patterns (i.e., spatial strategies) in the game, with each solution represented by Ripley’s metric.
Learners’ cognitive strategies were reflected in their problem-solving behavior in reasoning problems.
Bayesian Knowledge Tracing. Cui et al. (2019) utilized Bayesian Knowledge Tracing to analyze the student
response process in a game-based assessment environment. They identified skills that can be learned in the game
and created a model that updates the probability of mastering a skill using student answers. The technique revealed
changes in a learner’s status of skill or knowledge mastery over time during gameplay.
SRL Profiling. In learning environments, a significant portion of learning depends on learners’ decisions, which
are related to their SRL skills.
Decision Tree. Sabourin et al. (2013) analyzed middle school students’ SRL behaviors in a game-based learning
environment and classified the students into SRL categories. They used multiple features, including demographics,
pretest scores, personality surveys, goal orientations, emotion surveys, the number of gameplays, and off-task
behaviors. Each student’s SRL levels were categorized into low, medium, and high levels. The data was trained with
supervised ML techniques such as Naïve Bayes, Decision Tree, SVM, Logistic Regression, and Neural Network. As
a result of cross-validation, Decision Tree showed the highest predictive accuracy. Adopting this technique in the
learning support system could automatically identify students with low SRL levels.
15
URL: www.onlinesciencepublishing.com | February, 2024
American Journal of Education and Learning, 2024, 9(1): 1-21
scores, GPA, gender, ethnicity, and e-portfolio logins, submissions, and hits. When using the data set of academic
and engagement information, Naïve Bayes performed the best among classification methods (e.g., Decision Tree,
Logistic Regression, Random Forests). Specifically, their Naïve Bayes model correctly predicted 42 of the 48
students who were kept in school.
Random Forest. Spoon et al. (2016) examined students’ final exam scores, course completion, LMS data, and
institutional data. They utilized Random Forest to identify students who are at risk of failing a course. In their
model, passing and non-passing grades were classified into trees, with final exam scores serving as the y-values in
regression trees. If a student’s successful course completion is predicted by less than 50% of the trees in their forest,
the student can be considered at-risk. The researchers argue that their model’s ROC (Receiver Operating
Characteristic) curve, a graphical plot that reveals the classification power by representing the true positive rate
against the false positive rate, shows better performance than the institution’s traditional approach that uses quiz or
exam scores.
K-Nearest Neighbor. Gray et al. (2016) investigated how to predict at-risk undergraduate students using data
collected from the registration process, a learner profiling survey of first-year students, and their first-year
academic performance. They looked at how accurate different modeling methods were (e.g., Naïve Bayes, Decision
Tree, Logistic Regression, Back Propagation Neural Network, Support Vector Machine). K-Nearest Neighbor had
the best accuracy for a certain dataset when predicting students who might be at risk.
4. DISCUSSION
Different types of AI techniques used in education research were identified, revealing the front- and back-end
frameworks, and a learning design model, SLAA, was proposed. The explanation of the AI or ML techniques can be
useful for a researcher who intends to adopt a specific technique for their studies. With the front- and back-end
framework, researchers can differentiate their approaches when using AI in education.
Researchers can adopt SLAA to design their learning support system with specific AI techniques. Still, there
are some issues we were not able to address in our model. First, for the learner-AI interaction, the interface’s role
involves noticing, interpreting, and responding; thus, AI-infused learning support systems need to understand
natural human languages. Voice recognition and text-to-speech technologies have evolved rapidly in recent history.
It seems that the learner interface could adopt these technologies to support conversational interactions between
the learner and systems. Modeling learners’ conversation content, patterns, and intention could be crucial for
scaling natural language understanding of agent-based approaches, which recognize goals and tasks and interact
with the environment and learners. Second, along with the cognitive and metacognitive scaffolds, the learner’s
affective aspects (e.g., emotion, sensation, and feeling) should be considered when designing an AI-infused learning
support system. Affective aspects play a significant role in learning (Immordino‐Yang & Damasio, 2007) and impact
learning, such as learning judgment, learning interest and attention, motivation, knowledge retention, problem-
solving skills, and decision-making (Blanchard, Volfson, Hong, & Lajoie, 2009). The learner’s affective aspects can
be taken into consideration in the prediction modeling to maximize the probability of learning success. Lastly,
learning support systems need to support a group of learners’ goal settings, analyze collaborative performances, and
encourage the active cooperation of the learning community. Along with individual learner modeling, group
learning or collaborative learning modeling needs to be considered. Future systems need to structure the flow of
collaborative learners’ knowledge. Learning environments can also adopt social learning networks, which support
knowledge sharing as a type of learning community (Spoelstra et al., 2014), as human-AI collaboration will likely
increase.
16
URL: www.onlinesciencepublishing.com | February, 2024
American Journal of Education and Learning, 2024, 9(1): 1-21
It is anticipated that AI will tackle a wider range of learning problems and offer effective scaffolding. However,
it should also be noted that concerns and possible pessimism exist about embracing the future of AI and
undertaking AI-led education initiatives.
REFERENCES
Aguiar, E., Ambrose, G. A. A., Chawla, N. V., Goodrich, V., & Brockman, J. (2014). Engagement vs performance: Using
electronic portfolios to predict first semester engineering student persistence. Journal of Learning Analytics, 1(3), 7-33.
https://doi.org/10.18608/jla.2014.13.3
Albatayneh, N. A., Ghauth, K. I., & Chua, F.-F. (2018). Utilizing learners’ negative ratings in semantic content-based
recommender system for e-learning forum. Journal of Educational Technology & Society, 21(1), 112-125.
Althoff, T., Clark, K., & Leskovec, J. (2016). Large-scale analysis of counseling conversations: An application of natural language
processing to mental health. Transactions of the Association for Computational Linguistics, 4, 463-476.
https://doi.org/10.1162/tacl_a_00111
Andrade, A., Danish, J. A., & Maltese, A. V. (2017). A measurement model of gestures in an embodied learning environment:
Accounting for temporal dependencies. Journal of Learning Analytics, 4(3), 18–46.
https://doi.org/10.18608/jla.2017.43.3
Andrade, A., Delandshere, G., & Danish, J. A. (2016). Using multimodal learning analytics to model student behavior: A
systematic analysis of epistemological framing. Journal of Learning Analytics, 3(2), 282-306.
https://doi.org/10.18608/jla.2016.32.14
Beck, J., Stern, M., & Haugsjaa, E. (1996). Applications of AI in education. Crossroads, 3(1), 11-15.
Belland, B. R., Kim, C., & Hannafin, M. J. (2013). A framework for designing scaffolds that improve motivation and cognition.
Educational Psychologist, 48(4), 243-270.
Blanchard, E. G., Volfson, B., Hong, Y.-J., & Lajoie, S. P. (2009). Affective artificial intelligence in education: From detection to
adaptation. In V. Dimitrova, R. Mizoguchi, B. du Boulay & A. Grasser (Eds.), Artificial intelligence in education.
Building learning systems that care: From knowledge representation to affective modelling. In (pp. 81-88): IOS Press.
https://doi.org/10.3233/978-1-60750-028-5-81.
Boroujeni, M. S., & Dillenbourg, P. (2019). Discovery and temporal analysis of MOOC study patterns. Journal of Learning
Analytics, 6(1), 16—33. https://doi.org/10.18608/jla.2019.61.2
Cetintas, S., Si, L., Xin, Y. P., Zhang, D., Park, J. Y., & Tzur, R. (2010). A joint probabilistic classification model of relevant and
irrelevant sentences in mathematical word problems. Journal of Educational Data Mining, 2(1), 83-101.
https://doi.org/10.5281/zenodo.3554741
Chen, C.-M. (2008). Intelligent web-based learning system with personalized learning path guidance. Computers & Education,
51(2), 787-814. https://doi.org/10.1016/j.compedu.2007.08.004
Cui, Y., Chu, M.-W., & Chen, F. (2019). Analyzing student process data in game-based assessments with Bayesian knowledge
tracing and dynamic Bayesian networks. Journal of Educational Data Mining, 11(1), 80-100.
https://doi.org/10.5281/zenodo.3554751
Dascalu, M.-I., Bodea, C.-N., Moldoveanu, A., Mohora, A., Lytras, M., & de Pablos, P. O. (2015). A recommender agent based on
learning styles for better virtual collaborative learning experiences. Computers in Human Behavior, 45, 243-253.
https://doi.org/10.1016/j.chb.2014.12.027
De Medio, C., Limongelli, C., Sciarrone, F., & Temperini, M. (2020). MoodleREC: A recommendation system for creating
courses using the moodle e-learning platform. Computers in Human Behavior, 104, 106168.
https://doi.org/10.1016/j.chb.2019.106168
17
URL: www.onlinesciencepublishing.com | February, 2024
American Journal of Education and Learning, 2024, 9(1): 1-21
Department of the Army. (2011). The U.S. Army learning concept for 2015. Fort eustis: Training and doctrine command. Retrieved
from https://adminpubs.tradoc.army.mil/pamphlets/TP525-8-2.pdf
Desmarais, M. C., & Baker, R. S. d. (2012). A review of recent advances in learner and skill modeling in intelligent learning
environments. User Modeling and User-Adapted Interaction, 22(1-2), 9-38. https://doi.org/10.1007/s11257-011-9106-8
Dorça, F. A., Lima, L. V., Fernandes, M. A., & Lopes, C. R. (2013). Comparing strategies for modeling students learning styles
through reinforcement learning in adaptive and intelligent educational systems: An experimental analysis. Expert
Systems with Applications, 40(6), 2092-2101. https://doi.org/10.1016/j.eswa.2012.10.014
Ezen-Can, A., & Boyer, K. E. (2015). Understanding student language: An unsupervised dialogue act classification approach.
Journal of Educational Data Mining, 7(1), 51-78. https://doi.org/10.5281/zenodo.3554707
Ezen-Can, A., Grafsgaard, J. F., Lester, J. C., & Boyer, K. E. (2015). Classifying student dialogue acts with multimodal learning
analytics. Paper presented at the Proceedings of the Fifth International Conference on Learning Analytics and
Knowledge, New York, NY: ACM.
Ferguson, R., & Clow, D. (2015). Consistent commitment: Patterns of engagement across time in massive open online courses
(MOOCs). Journal of Learning Analytics, 2(3), 55-80. https://doi.org/10.18608/jla.2015.23.5
Fincham, E., Gašević, D., & Pardo, A. (2018). From social ties to network processes: Do tie definitions matter? Journal of
Learning Analytics, 5(2), 9–28. https://doi.org/10.18608/jla.2018.52.2
Gardner, J., & Brooks, C. (2018). Evaluating predictive models of student success: Closing the methodological gap. Journal of
Learning Analytics, 5(2), 105-125. https://doi.org/10.18608/jla.2018.52.7
Geigle, C., & Zhai, C. (2017). Modeling student behavior with two-layer hidden markov models. Journal of Educational Data
Mining, 9(1), 1-24.
Gray, G., McGuinness, C., Owende, P., & Hofmann, M. (2016). Learning factor models of students at risk of failing in the early
stage of tertiary education. Journal of Learning Analytics, 3(2), 330-372. https://doi.org/10.18608/jla.2016.32.20
Gruzd, A., Paulin, D., & Haythornthwaite, C. (2016). Analyzing social media and learning through content and social network
analysis: A faceted methodological approach. Journal of Learning Analytics, 3(3), 46-71.
https://doi.org/10.18608/jla.2016.33.4
Hao, J., Shu, Z., & von Davier, A. (2015). Analyzing process data from game/scenario-based tasks: An edit distance approach.
Journal of Educational Data Mining, 7(1), 33-50. https://doi.org/10.5281/zenodo.3554705
Harley, J. M., Bouchet, F., Trevors, G. J., & Azevedo, R. (2013). Clustering and profiling students according to their interactions
with an intelligent tutoring system fostering self-regulated learning. Journal of Educational Data Mining, 5(1), 104-146.
https://doi.org/10.5281/zenodo.3554613
Hecking, T., Ziebarth, S., & Hoppe, H. U. (2014). Analysis of dynamic resource access patterns in online courses. Journal of
Learning Analytics, 1(3), 34-60. https://doi.org/10.18608/jla.2014.13.4
Immordino‐Yang, M. H., & Damasio, A. (2007). We feel, therefore we learn: The relevance of affective and social neuroscience to
education. Mind, Brain, and Education, 1(1), 3-10. https://doi.org/10.1111/j.1751-228X.2007.00004.x
Kai, S., Almeda, M. V., Baker, R. S., Heffernan, C., & Heffernan, N. (2018). Decision tree modeling of wheel-spinning and
productive persistence in skill builders. Journal of Educational Data Mining, 10(1), 36-71.
https://doi.org/10.5281/zenodo.3344810
Kamaraj, A., Nidhyananthan, S., & Sundaram, K. (2019). Voice biometric for learner authentication: Biometric authentication. In
Biometric Authentication in Online Learning Environments. In (pp. 150-181): IGI Global.
https://doi.org/10.4018/978-1-5225-7724-9.ch007.
Kassak, O., Kompan, M., & Bielikova, M. (2016). Student behavior in a web-based educational system: Exit intent prediction.
Engineering Applications of Artificial Intelligence, 51, 136-149. https://doi.org/10.1016/j.engappai.2016.01.018
18
URL: www.onlinesciencepublishing.com | February, 2024
American Journal of Education and Learning, 2024, 9(1): 1-21
Knight, S., Buckingham Shum, S., Ryan, P., Sándor, Á., & Wang, X. (2018). Designing academic writing analytics for civil law
student self-assessment. International Journal of Artificial Intelligence in Education, 28, 1-28.
https://doi.org/10.1007/s40593-016-0121-0
Kotsiantis, S. B., Zaharakis, I. D., & Pintelas, P. E. (2006). Machine learning: A review of classification and combining
techniques. Artificial Intelligence Review, 26, 159-190. https://doi.org/10.1007/s10462-007-9052-3
Kowalski, J., Zhang, Y., & Gordon, G. J. (2014). Statistical modeling of student performance to improve Chinese dictation skills
with an intelligent tutor. Journal of Educational Data Mining, 6(1), 3-27. https://doi.org/10.5281/zenodo.3554679
Lee, A. V. Y., & Tan, S. C. (2017). Promising ideas for collective advancement of communal knowledge using temporal analytics
and cluster analysis. Journal of Learning Analytics, 4(3), 76-101. https://doi.org/10.18608/jla.2017.43.5
Lester, J. C., Ha, E. Y., Lee, S. Y., Mott, B. W., Rowe, J. P., & Sabourin, J. L. (2013). Serious games get smart: Intelligent game-
based learning environments. AI Magazine, 34(4), 31-45. https://doi.org/10.1609/aimag.v34i4.2488
Ludwig-Hardman, S., & Dunlap, J. C. (2003). Learner support services for online students: Scaffolding for success. International
Review of Research in Open and Distributed Learning, 4(1), 1-15. https://doi.org/10.19173/irrodl.v4i1.131
Mahzoon, M. J., Maher, M. L., Eltayeby, O., Dou, W., & Grace, K. (2018). A sequence data model for analyzing temporal
patterns of student data. Journal of Learning Analytics, 5(1), 55–74. https://doi.org/10.18608/jla.2018.51.5
Mallavarapu, A., Lyons, L., Shelley, T., Minor, E., Slattery, B., & Zellner, M. (2015). Developing computational methods to
measure and track learners’ spatial reasoning in an open-ended simulation. Journal of Educational Data Mining, 7(2), 49-
82. https://doi.org/10.5281/zenodo.3554669
Mao, Y. (2018). Deep learning vs. Bayesian knowledge tracing: Student models for interventions. Journal of Educational Data
Mining, 10(2), 28-54. https://doi.org/10.5281/zenodo.3554691
Miller, L. D., Soh, L.-K., Samal, A., Kupzyk, K., & Nugent, G. (2015). A comparison of educational statistics and data mining
approaches to identify characteristics that impact online learning. Journal of Educational Data Mining, 7(3), 117-150.
https://doi.org/10.5281/zenodo.3554731
Mirriahi, N., Liaqat, D., Dawson, S., & Gašević, D. (2016). Uncovering student learning profiles with a video annotation tool:
Reflective learning with and without instructional norms. Educational Technology Research and Development, 64, 1083-
1106. https://doi.org/10.1007/s11423-016-9449-2
Morsy, S., & Karypis, G. (2019). Will this course increase or decrease your GPA? towards grade-aware course recommendation.
Journal of Educational Data Mining, 11(2), 20-46.
Muis, K. R., Psaradellis, C., Chevrier, M., Di Leo, I., & Lajoie, S. P. (2016). Learning by preparing to teach: Fostering self-
regulatory processes and achievement during complex mathematics problem solving. Journal of Educational Psychology,
108(4), 474-492. https://doi.org/10.1037/edu0000071
Oh, Y. E., Song, D., & Hong, H. (2020). Interactive computing technology in anti-bullying education: The effects of
conversation-bot’s role on K-12 students’ attitude change toward bullying problems. Journal of Educational Computing
Research, 58(1), 200-219. https://doi.org/10.1177/0735633119839177
Pelaez, K., Levine, R., Fan, J., Guarcello, M., & Laumakis, M. (2019). Using a latent class forest to identify at-risk students in
higher education. Journal of Educational Data Mining, 11(1), 18-46. https://doi.org/10.5281/zenodo.3554747
Pintrich, P. R. (1999). The role of motivation in promoting and sustaining self-regulated learning. International Journal of
Educational Research, 31(6), 459-470. https://doi.org/10.1016/S0883-0355(99)00015-4
Russell, S., & Norvig, P. (2010). Artificial intelligence: A modern approach (3rd ed.). Upper Saddle River: Prentice-Hall.
Sabourin, J. L., Shores, L. R., Mott, B. W., & Lester, J. C. (2013). Understanding and predicting student self-regulated learning
strategies in game-based learning environments. International Journal of Artificial Intelligence in Education, 23, 94-114.
https://doi.org/10.1007/s40593-013-0004-6
19
URL: www.onlinesciencepublishing.com | February, 2024
American Journal of Education and Learning, 2024, 9(1): 1-21
Schneider, B., & Pea, R. (2015). Does seeing one another’s gaze affect group dialogue? A computational approach. Journal of
Learning Analytics, 2(2), 107-133. https://doi.org/https://doi.org/10.18608/jla.2015.22.9
Searle, J. R., Kiefer, F., & Bierwisch, M. (1980). Speech act theory and pragmatics. In (Vol. 10, pp. 205-220). Dordrecht: D.
Reidel.
Serban, I. V., Lowe, R., Henderson, P., Charlin, L., & Pineau, J. (2018). A survey of available corpora for building data-driven
dialogue systems: The journal version. Dialogue & Discourse, 9(1), 1-49. https://doi.org/10.5087/dad.2018.101
Shen, S., Mostafavi, B., Barnes, T., & Chi, M. (2018). Exploring induced pedagogical strategies through a Markov decision
process framework: Lessons learned. Journal of Educational Data Mining, 10(3), 27-68.
https://doi.org/10.5281/zenodo.3554713
Shermis, M. D., & Burstein, J. C. (2003). Automated essay scoring: A cross-disciplinary perspective. Hillsdale, NJ: Lawrence Erlbaum.
Siebert-Evenstone, A. L., Irgens, G. A., Collier, W., Swiecki, Z., Ruis, A. R., & Shaffer, D. W. (2017). In search of conversational
grain size: Modeling semantic structure using moving stanza windows. Journal of Learning Analytics, 4(3), 123–139.
https://doi.org/10.18608/jla.2017.43.7
Song, D. (2018). Learning analytics as an educational research approach. International Journal of Multiple Research Approaches,
10(1), 102-111. https://doi.org/10.29034/ijmra.v10n1a6
Song, D., Rice, M., & Oh, E. Y. (2019). Participation in online courses and interaction with a virtual agent. International Review of
Research in Open and Distributed Learning, 20(1), 43-62. https://doi.org/10.19173/irrodl.v20i1.3998
Sonnenberg, C., & Bannert, M. (2016). Evaluating the impact of instructional support using data mining and process mining: A
micro-level analysis of the effectiveness of metacognitive prompts. Journal of Educational Data Mining, 8(2), 51-83.
https://doi.org/10.5281/zenodo.3554597
Spoelstra, H., Van Rosmalen, P., & Sloep, P. (2014). Toward project-based learning and team formation in open learning
environments. Journal of Universal Computer Science, 20(1), 57-76. https://doi.org/10.3217/jucs-020-01-0057
Spoon, K., Beemer, J., Whitmer, J. C., Fan, J., Frazee, J. P., Stronach, J., . . . Levine, R. A. (2016). Random forests for evaluating
pedagogy and informing personalized learning. Journal of Educational Data Mining, 8(2), 20-50.
https://doi.org/10.5281/zenodo.3554595
Sullivan, F. R., & Keith, P. K. (2019). Exploring the potential of natural language processing to support microgenetic analysis of
collaborative learning discussions. British Journal of Educational Technology, 50(6), 3047-3063.
https://doi.org/10.1111/bjet.12875
Taherkhani, A., & Malmi, L. (2013). Beacon-and schema-based method for recognizing algorithms from students’ source code.
Journal of Educational Data Mining, 5(2), 69-101. https://doi.org/10.5281/zenodo.3554635
Taub, M., & Azevedo, R. (2018). Using sequence mining to analyze metacognitive monitoring and scientific inquiry based on
levels of efficiency and emotions during game-based learning. Journal of Educational Data Mining, 10(3), 1-26.
https://doi.org/10.5281/zenodo.3554711
Vaessen, B. E., Prins, F. J., & Jeuring, J. (2014). University students' achievement goals and help-seeking strategies in an
intelligent tutoring system. Computers & Education, 72, 196-208. https://doi.org/10.1016/j.compedu.2013.11.001
Waters, A. E., Studer, C., & Baraniuk, R. G. (2014). Collaboration-type identification in educational datasets. Journal of
Educational Data Mining, 6(1), 28-52. https://doi.org/10.5281/zenodo.3554681
Wolters, C., & Taylor, D. (2012). A self-regulated learning perspective on student engagement. In S. L. Christenson, A. Reschly,
& C. Wylie (Eds.), Handbook of research on student engagement. In (pp. 635–651): Springer.
https://doi.org/10.1007/978-1-4614-2018-7_30.
20
URL: www.onlinesciencepublishing.com | February, 2024
American Journal of Education and Learning, 2024, 9(1): 1-21
Xiao, J., Wang, M., Jiang, B., & Li, J. (2018). A personalized recommendation system with combinational algorithm for online
learning. Journal of Ambient Intelligence and Humanized Computing, 9, 667-677. https://doi.org/10.1007/s12652-017-
0466-8
Xu, B., & Recker, M. (2011). Understanding teacher users of a digital library service: A clustering approach. Journal of
Educational Data Mining, 3(1), 1-28. https://doi.org/10.5281/zenodo.3554701
Yoo, J., & Kim, J. (2014). Can online discussion participation predict group project performance? Investigating the roles of
linguistic features and participation patterns. International Journal of Artificial Intelligence in Education, 24, 8-32.
https://doi.org/10.1007/s40593-013-0010-8
Zimmerman, B. J. (1990). Self-regulated learning and academic achievement: An overview. Educational Psychologist, 25(1), 3-17.
https://doi.org/10.1207/s15326985ep2501_2
Zimmerman, B. J. (2002). Becoming a self-regulated learner: An overview. Theory into Practice, 41(2), 64-70.
https://doi.org/10.1207/s15430421tip4102_2
Online Science Publishing is not responsible or answerable for any loss, damage or liability, etc. caused in relation to/arising out of the
use of the content. Any queries should be directed to the corresponding author of the article.
21
URL: www.onlinesciencepublishing.com | February, 2024