Conference PaperPDF Available

Learning Analytics in MOOCs: Can Data Improve Students Retention and Learning?

Authors:

Abstract and Figures

In order to study learners’ behaviors and activities in online learning environments such as MOOCs, the demanding for a framework of practices and procedures to collect, analyze and optimize their data emerged in the educational learning horizon. Learning Analytics is the field that arose to comply with such needs and was denominated as a “technological fix to the long-standing problems” of online learning platforms (Knox, 2014). This paper discusses the significance of applying Learning Analytics in MOOCs to overcome some of its issues. We will mainly focus on improving students’ retention and learning using an algorithm prototype based on divergent MOOC indicators, and propose a scheme to reflect the results on MOOC students.
Learning Analytics – MOOCs Scheme The second stage is where Learning Analytics server processes log files. It picks up keywords that help in determining students' interactions inside the bulk text of log files. In addition, it filters unstructured and duplicated data in order to be handled properly. In the third stage, the main activities of each student are parsed separately in order to support the algorithm procedures as discussed in the previous section. Upon calculating the weight of each interaction using mathematical operations, the results are formulated into a sequence of partitions for each student in furtherance of being dispatched to the next stage. The fourth stage is where the collected, organized, operated interactions of learners' data are interpreted for the visualization part. The adopted method to show the results is the user dashboards. According to Verbert et al. (2014), dashboards support awareness, reflection, sense-making and ease learners to track their progress. The user interface should support feedback on activities and predict performance of students. The ideal of gamification could be presented in this proposed scheme using instruments such as: a progress bar or a colorful gauge. The aim is to boost students' motivation for learning and to sustain their interest in MOOCs. At this stage, the dashboard is intended to show a student's progress compared to other students. An indication of being behind the others, commensurate, or overhead will lead learners to react accordingly. The last stage is the feedback section. As a consequence to the informative notification from the user dashboard, the awaited reaction is associated to a fruitful feedback. Timely, individual and empowering have been attributed as the needed qualities of feedbacks to get the awareness of students (Race, 2001). Thus, a student will be able to get updates about his/her performance on a weekly basis to ignite the learning competition and ambition. As a matter of fact, closing the feedback loop and enabling learners to react effectively is what pursued in our final scheme.
… 
Content may be subject to copyright.
Learning Analytics in MOOCs: Can Data Improve
Students Retention and Learning?
Mohammad Khalil
Educational Technology
Graz University of Technology
Graz, Austria
mohammad.khalil@tugraz.at
Martin Ebner
Educational Technology
Graz University of Technology
Graz, Austria
martin.ebner@tugraz.at
Abstract: In order to study learners’ behaviors and activities in online learning environments
such as MOOCs, the demanding for a framework of practices and procedures to collect,
analyze and optimize their data emerged in the educational learning horizon. Learning
Analytics is the field that arose to comply with such needs and was denominated as a
“technological fix to the long-standing problems” of online learning platforms (Knox, 2014).
This paper discusses the significance of applying Learning Analytics in MOOCs to overcome
some of its issues. We will mainly focus on improving students’ retention and learning using
an algorithm prototype based on divergent MOOC indicators, and propose a scheme to reflect
the results on MOOC students.
Introduction
Massive Open Online Courses (MOOCs) provide massive amounts of data about learners and how
they interact with an online learning environment. Since Siemens and Downes launched their first open online
course back in 2008, MOOCs have been steadily spreading across the Internet (McAuley et al., 2010). Due to
its openness, MOOC students vary in their heterogeneity such as age, gender, educational background and
location. With an internet access, a student from anywhere in the world can access high quality courses such as
the ones provided by Harvard University or Massachusetts Institute of Technology through the well-known
MOOC provider, edX1. In addition to that, learners are not only limited to a single type path learning
specialization. For instance, a student of social science major can attend a computer programming course, and
this does not stop here. (S)he can get a certificate after successfully achieving all the required exams of that
course.
According to the high number of MOOCs enrollees, a rich content of information can be stored in the
databases of MOOCs servers. The accumulation of such data leads to what is so-called “Big Data”. This term
has become familiar in the recent years and it refers to “datasets whose size is beyond the ability of typical
database software tools to capture, store, manage, and analyze.” (Manyika et al., 2011). However, when noisy,
unstructured and steep heterogeneous data are filtered, the examination and analysis becomes conceivable. In
2011, a special field called Learning Analytics has emerged after the growing needs to understand behaviour
and attitudes of learners in online learning platforms and the needed advice in learning (Siemens, 2010). It is
presumed that Learning Analytics is firmly related to other fields such as web analytics, educational data
mining, academic analytics and business intelligence (Elias, 2011). The rapid growth of analytics on
educational data is related to the advancements of computer technology and tools as well as the ease of
acquiring educational data that are produced by online learning environments (Schön et al, 2011).
As a related topic, Knox (2014) as well as Khalil and Ebner (2015a) discussed the high potential of
Learning Analytics when it is applied to educational datasets of MOOCs. Logging mouse clicks, tracking
forums activity, time spent on tasks, quiz performance as well as login frequency of learners generate
1 http://www.edx.org (Last access, 04.12.2015)
Preview version of this paper. Content and pagination may change prior to final publication.
EdMedia 2016 - Vancouver, BC, Canada, June 28-30, 2016
consequential data which acts as a rich source of valuable knowledge for Learning Analytics researchers in the
MOOCs environment. The promising goals of combining both fields can be linked with frameworks such as the
one provided by (Khalil & Ebner 2015b) in which Learning Analytics can: a) predict grade, knowledge or
performance of learners in MOOCs; b) intervene to control drop-out rate and detect students at-risk; c)
recommend and personalize platforms to suit the needs of learners, teachers and institutions; d) reflect the
outcome knowledge to improve future experience such as students experience; e) benchmark the weak points of
learning environment systems, which is MOOC platforms in our case. Accordingly, this research study will aim
attention at the implementation of Learning Analytics in the leading Austrian MOOC platform, iMooX
(http://www.imoox.at).
This research work is a continuation of previous studies on this platform (Khalil & Ebner, 2015a;
Khalil, Kastl & Ebner, 2016), and an edge to get into a deeper interval of analysis of learners data to examine
the potential of Learning Analytics in order to: a) improve learning in MOOCs in general; b) enhance the
completion rate in MOOCs; and c) study learner patterns and predict at-risk students. The research study will
present the used methodology to establish the first algorithm prototype and introduce its results of testing
students’ information. Moreover, we will propose a Learning Analytics MOOCs scheme to employ the
feedback principle which will lead to accomplish our goals.
This publication is organized as follows: the next section lists the related work. Section 3 covers the
research methodology. Section 4 gives an overview about the MOOC platform and the course structure. Section
5 shows our data analysis and the algorithm extraction. Finally, section 6 proposes the Learning Analytics
MOOC scheme.
Related Work
Predicting behaviour of students and recognizing who are at-risk is not a new topic. For instance, in
1985, Noel and Levitz described retention, linked it with students’ success and stated that the course topic is a
major factor to increase a successful completion rate (Noel & Levitz, 1985). It is noticed that subjects about
attrition rate and completion rate are highly debated in the old time education and educational technology.
In relation to MOOCs, psychological and motivational factors have been proposed as phenomena to
clarify the dropout rate and at-risk students (Khalil & Ebner, 2014). Santos et al. (2014) defined a threshold to
detect dropout rate criteria. They found that forum activity is a reliable indicator to predict students who might
drop in MOOCs. Further, Lackner, Ebner and Khalil (2015), adduced to revise MOOCs duration. They found
that splitting them into two periods will increase the retention rate. Additionally, some studies pointed out to the
number of assignments attempts as a factor to understand the completion rate (Coffrin et al., 2014). Similar to
this research study, Balakrishnan and Coetzee (2013) compared behaviour of students who dropped out and
who completed MOOCs by using the mathematical Markov chain. They found out that the students who do not
login and check their course progress, usually dropout dramatically. At the end, the proposed Learning
Analytics MOOCs scheme was influenced by Course Signals (Arnold & Pistilli, 2012). The Application is
made to provide a direct feedback to students through an intervention which is based on data analysis.
Moreover, it works as a traffic light in which a student gets a red light color if (s)he is at-risk to fail and gets a
green light color when (s)he is close to succeed in the course, while a yellow color indicates a potential problem
of succeeding.
Research Methodology
This research study employed a procedure of collecting the data from two MOOCs presented in 2014
on the iMooX platform. The first course called Gratis Online Lernen, and abbreviated in this paper as GOL2014
(Ebner et al, 2015). The second course called Lernen im Netz, and is abbreviated as LIN2014 (Lackner & Kopp,
2014) The Learning Analytics application parses log files, which contains records of the students’ activities on
the learning platform, and filter them. After that, the filtered data are exported into a readable file and the results
are clustered into two main categories: completed students, and those who successfully completed the course
and did all the tasks. The second category is the students who dropped out during the course. The activities of
each group are then assorted to categories and the prevailing behaviours are analyzed. At the end, we observed
the differences in order to introduce an algorithm prototype based on weights to predict students who will drop
during the course and proposed a scheme with a view to implement our results.
Preview version of this paper. Content and pagination may change prior to final publication.
EdMedia 2016 - Vancouver, BC, Canada, June 28-30, 2016
MOOC-Platform & Courses Overview
The Platform
iMooX is an online learning platform that follows the xMOOC family and was first introduced in
2013. It is the first MOOC platform in Austria and was founded by the collaboration of Graz University of
Technology and University of Graz. The platform is enriched by Open Educational Resource courses and
adheres to open education and lifelong learning paths (Neuböck, Kopp, & Ebner, 2015). iMooX supports the
online pedagogy of offering courses to students on a weekly basis. Concurring with MOOCs forms of learning,
it offers videos, multiple attempts quizzes, online discussions through forums and interactive learning objects.
In addition, certificates are offered to the students who successfully pass all the required tasks at no cost.
Courses Description
Two courses were analyzed in this research study: Gratis Online Lernen and Lernen im Netz.
Gratis Online Lernen (GOL2014)
GOL2014 was an eight-week course that started in October 2014, offered by Graz University of
Technology and provided in German language. The course focused on educating people through the internet
and instructed them on how to do it. The workload was set to be 2 hours/week. There were 1012 participants in
the course. 217 students successfully completed the course; therefore 21.5% was the completion rate in this
MOOC.
Lernen im Netz (LIN2014)
LIN2014 was also an eight-week course that started in October 2014 till the mid of December 2014.
The MOOC was offered by the University of Graz and taught in German language. The workload was set to be
5 hours/week and the topics were about using social media in the Open Educational Resources. 519 was the
number of students who registered for the course, and the completion rate was 25%.
Analysis and Algorithm Extraction
In order to generate the prediction algorithm, an examination of GOL2014 and LIN2014 were
scrutinized. Both of these two courses attracted the largest sample of students in iMooX platform in 2014.
Basically, using the Learning Analytics approach, MOOC interactions have been filtered into indicators. These
indicators are: a) quiz attempts; b) discussion forum readings; c) discussion forum writings; and d) login
frequency. Each student profile was then dedicated with these indicators separately. In addition, all the collected
data were distributed based on a weekly scale. Respectively, we calculated the total interactions for the
completed students, who successfully finish MOOCs, and for the students who dropped and calculated the
average.
Figure 1a, shows the behaviour of students who completed the course and the ones who dropped out in
GOL2014. The left figure displays the average of all interactions in the whole course period, while the right
figure shows the average of interactions from week3 to week6. The reason of analyzing that period is due to the
stability of the dropout rate during that duration. Usually, students register in the first weeks to discover the
material and the course, therefore the dropout rate in that period is quite high (Lackner, Ebner & Khalil, 2015).
Moreover, the research study found that students, who stayed till week 5, have the potential to complete courses
more often than in the earlier weeks. In the same fashion, figure 1b demonstrates the behaviour of both student
types in the LIN2014 course. By observation of the line graph in both of the figures, students behaved nearly
identical.
Preview version of this paper. Content and pagination may change prior to final publication.
EdMedia 2016 - Vancouver, BC, Canada, June 28-30, 2016
The figures show that the difference between quiz attempts is obvious between both of student types.
Under that circumstance, quiz attempts factor has been pulled to be the second critical criteria in the algorithm.
Figure 1: The average of MOOC interactions for completed and dropout students. Top (a) GOL2014
course; bottom (b) LIN2014 course
Preview version of this paper. Content and pagination may change prior to final publication.
EdMedia 2016 - Vancouver, BC, Canada, June 28-30, 2016
Figure 2: Results of applying the algorithm on random samples from other courses. (a) Dropout
students; (b) Completed students
However, the greatest difference gap among MOOC indicators is the forums reading activity. This
explains several studies such as the one by Ezen-Can et al (2015), which have drawn attention to the significant
effect of MOOCs forum interaction to improve learning and attrition.
Furthermore, the third remarkable difference is the login frequency which is noted as a decisive player
in determining at-risk student, and this was highlighted in several studies like the one by Balakrishnan and
Coetzee (2013). In contrast, the writings were not as efficient as readings; hence, the allocated weight for
writing is the lowest.
Generally after all, each MOOC indicator was weighted and calculated based on the difference
between the activity performance of students who dropped and the ones who completed in both of the two cases
MOOCs. Subsequently, the algorithm predicts the retention percentage of each student and notify him/her at a
peak point when (s)he is at-risk. We defined the weights according to their adequate significance to (W1, W2,
W3, and W4), in which W1>W2>W3>W4. The following equation articulates the algorithm expression:
Success Rate (SR) = W1.Readings + W2.Quiz_Attempts + W3.Login_Frequency + W4.Writings
The first round testing showed promising results. In figure 2a, we see that students are dropping during
the weeks when the algorithm generates low numbers. For instance, some students dropped when the score was
below 40. However, some other students kept on track even when their score was 20. In such scenario, the
student will be notified that (s)he is at-risk, and a proper reaction is required. Further, figure 2b displays a
sample of students who completed the course. The orange tabs predict an at-risk situation, whereas the green
records mean the student has the potential to complete the MOOC.
Learning Analytics – MOOCs Scheme
In this part, we are looking to attain the findings and achieve the feedback goal, so that learners can get
the maximum benefits and be notified when their performance is in the danger area. Figure 3 shows the
proposed Learning Analytics MOOCs scheme. It is divided into five stages. The first stage starts when the
data is generated through the activities of the learners. The log system records their quiz performance and
attempts, discussion forum activity, their post and read count as well as committing login frequency to the
database.
Preview version of this paper. Content and pagination may change prior to final publication.
EdMedia 2016 - Vancouver, BC, Canada, June 28-30, 2016
Figure 3: Learning Analytics – MOOCs Scheme
The second stage is where Learning Analytics server processes log files. It picks up keywords that help
in determining students’ interactions inside the bulk text of log files. In addition, it filters unstructured and
duplicated data in order to be handled properly. In the third stage, the main activities of each student are parsed
separately in order to support the algorithm procedures as discussed in the previous section. Upon calculating
the weight of each interaction using mathematical operations, the results are formulated into a sequence of
partitions for each student in furtherance of being dispatched to the next stage.
The fourth stage is where the collected, organized, operated interactions of learners’ data are
interpreted for the visualization part. The adopted method to show the results is the user dashboards. According
to Verbert et al. (2014), dashboards support awareness, reflection, sense-making and ease learners to track their
progress. The user interface should support feedback on activities and predict performance of students. The
ideal of gamification could be presented in this proposed scheme using instruments such as: a progress bar or a
colorful gauge. The aim is to boost students’ motivation for learning and to sustain their interest in MOOCs. At
this stage, the dashboard is intended to show a student’s progress compared to other students. An indication of
being behind the others, commensurate, or overhead will lead learners to react accordingly.
The last stage is the feedback section. As a consequence to the informative notification from the user
dashboard, the awaited reaction is associated to a fruitful feedback. Timely, individual and empowering have
been attributed as the needed qualities of feedbacks to get the awareness of students (Race, 2001). Thus, a
student will be able to get updates about his/her performance on a weekly basis to ignite the learning
competition and ambition. As a matter of fact, closing the feedback loop and enabling learners to react
effectively is what pursued in our final scheme.
Conclusion
Learning Analytics is a field that promises to provide various solutions for online learning
environments such as MOOCs. Despite that these courses are free to join, and attract a large volume of the
community; the high ratio of dropout and the contradiction about its ability to create a true learning
environment hinder the advancement of such a domain. Therefore, in this paper, we discussed the application of
Learning Analytics on two MOOCs and proposed a brief literature review as well as described the walkthrough
of tracking students’ traces. Additionally, and in order to surpass MOOCs dilemmas, we implied our solution
Preview version of this paper. Content and pagination may change prior to final publication.
EdMedia 2016 - Vancouver, BC, Canada, June 28-30, 2016
for the purposes of predicting student at-risk and notify them beforehand by using an algorithm in order to
increase retention rate, improve learning and study their behaviour. Moreover, we explained how we extracted
the algorithm in details and proposed a Learning Analytics MOOCs scheme that employ principles such as
awareness and feedback.
In the meantime, the research team is working on testing more samples before the implementation
stage and is developing compact visualizations that accommodate students and the institution needs.
Additionally, we consider and respect users’ privacy, therefore an implementation of a de-identification
procedure (Khalil & Ebner, 2016), is undergoing to cover the Learning Analytics application and the Learning
Analytics – MOOCs Scheme.
References
Arnold, K. E., & Pistilli, M. D. (2012). Course signals at Purdue: using learning analytics to increase student success.
In Proceedings of the 2nd International Conference on Learning Analytics and Knowledge (pp. 267-270). ACM.
Balakrishnan, G., & Coetzee, D. (2013). Predicting student retention in massive open online courses using hidden Markov
models. Electrical Engineering and Computer Sciences University of California at Berkeley.
Coffrin, C., Corrin, L., de Barba, P., & Kennedy, G. (2014). Visualizing patterns of student engagement and performance in
MOOCs. InProceedings of the fourth international conference on learning analytics and knowledge (pp. 83-92). ACM.
Ebner, M., Schön, S. & Käfmüller, K. (2015). Inverse Blended Learning bei „Gratis Online Lernen“ über den Versuch,
einen Online-Kurs für viele in die Lebenswelt von EinsteigerInnen zu integrieren. In: Digitale Medien und
Interdisziplinarität. Nistor, N. & Schirlitz, S. (Hrsg). Waxmann, Medien in der Wissenschaft Bd 68. pp. 197-206
Elias, T. (2011). Learning analytics: Definitions, processes and potentials. Retrieved November 2015 from
http://learninganalytics.net/LearningAnalyticsDefinitionsProcessesPotential.pdf.
Ezen-Can, A., Boyer, K. E., Kellogg, S., & Booth, S. (2015). Unsupervised modeling for understanding MOOC discussion
forums: a learning analytics approach. In Proceedings of the Fifth International Conference on Learning Analytics And
Knowledge (pp. 146-150). ACM.
Khalil, H. & Ebner, M. (2014). MOOCs Completion Rates and Possible Methods to Improve Retention A Literature
Review. In Proceedings of World Conference on Educational Multimedia, Hypermedia and Telecommunications 2014 (pp.
1236-1244). Chesapeake, VA: AACE.
Khalil, M., & Ebner, M. (2015a). A STEM MOOC for school children—What does learning analytics tell us?. In Interactive
Collaborative Learning (ICL), 2015 International Conference on (pp. 1217-1221). IEEE.
Khalil, M., & Ebner, M. (2015b). Learning analytics: principles and constraints. In Proceedings of World Conference on
Educational Multimedia, Hypermedia and Telecommunications (pp. 1326-1336).
Khalil, M., Kastl, C., & Ebner, M. (2016). Portraying MOOCs Learners: a Clustering Experience Using Learning Analytics.
In Proceedings of the European Stakeholder Summit on experiences and best practices in and around MOOCs (EMOOCS
2016), Graz, Austria, pp.265-278.
Khalil, M. & Ebner, M. (2016). “De-Identification in Learning Analytics”. Journal of Learning Analytics, 3 (1), pp. 129-
138.
Knox, J. (2014). From MOOCs to Learning Analytics: Scratching the surface of the 'visual'. eLearn, 2014(11), 3.
Lackner, E. & Kopp, M. (2014) Do MOOCs need a Special Instructional Design?. In: Proceedings of EDULEARN14
Conference 7th-9th July 2014, Barcelona, Spain, pp. 7138-7147
Lackner, E., Ebner, M., & Khalil, M. (2015). MOOCs as granular systems: design patterns to foster participant
activity. eLearning Papers, 42, 28-37.
Manyika, J., Chui, M., Brown, B., Bughin, J., Dobbs, R., Roxburgh, C., & Byers, A. H. (2011). Big data: The next frontier
for innovation, competition, and productivity.
McAuley, A., Stewart, B., Siemens, G., & Cormier, D. (2010). The MOOC model for digital practice. Retrieved December
Preview version of this paper. Content and pagination may change prior to final publication.
EdMedia 2016 - Vancouver, BC, Canada, June 28-30, 2016
2015 from http://www.davecormier.com/edblog/wp-content/uploads/MOOC_Final.pdf.
Neuböck, K., Kopp, M., Ebner, M. (2015). What do we know about typical MOOC participants? First insights from the
field, In: Proceedings of eMOOCs 2015 conference, Lebrun, M., de Waard, I., Ebner, M., Gaebel, M., Mons, Belgium, pp.
183-190.
Noel, L., & Levitz, R. (1985). Increasing student retention: New challenges and potential. In L. Noel, R. Levitz, & D. Saluri
(Eds.), Increasing student retention (pp. 1-27). San Francisco, CA: Jossey-Bass.
Race, P. (2001). Using feedback to help students learn. The Higher Education Academy, York. Retrieved December 2015
from http://wap.rdg.ac.uk/web/FILES/EngageinFeedback/Race_using_feedback_to_help_students_learn.pdf.
Santos, J. L., Klerkx, J., Duval, E., Gago, D., & Rodríguez, L. (2014). Success, activity and drop-outs in MOOCs an
exploratory study on the UNED COMA courses. In Proceedings of the Fourth International Conference on Learning
Analytics and Knowledge (pp. 98-102). ACM.
Schön, M., Ebner, M. & Kothmeier, G. (2012). It's Just About Learning the Multiplication Table. In Proceedings of the 2nd
International Conference on Learning Analytics and Knowledge (LAK '12), Simon Buckingham Shum, Dragan Gasevic,
and Rebecca Ferguson (Eds.). ACM, New York, NY, USA, 73-81
Siemens, G. (2010). What are Learning Analytics? Retrieved December 2015 from
http://www.elearnspace.org/blog/2010/08/25/what-are-learning-analytics/.
Verbert, K., Govaerts, S., Duval, E., Santos, J. L., Van Assche, F., Parra, G., & Klerkx, J. (2014). Learning dashboards: an
overview and future research opportunities. Personal and Ubiquitous Computing, 18(6), 1499-1514.
Preview version of this paper. Content and pagination may change prior to final publication.
EdMedia 2016 - Vancouver, BC, Canada, June 28-30, 2016
... Therefore, several researchers suggested the use of these elements to motivate students and also to provide teachers with feedback about their students' performance. This can further help them predict at-risk students [16,6]. For example, the number of the collected badges from the submitted activities and students' rank on the leaderboard, which is based on their collected number of points from their interactions with the learning environment, are indicators of students' performance in the course, hence they can be used to help the system predict the students with low performance (at-risk of failing) or dropping a class. ...
... It refers to the number of activities realized from the total of activities requested in a course. This factor has been recommended by Khalil and Ebner [16] to help in predicting at-risk students who have not completed the requested activities; and, (5) Forum and chat interactions which refers to students' participation in online discussions, such as the number of posts read, posts created and replies. This factor has been often used by Liu et al. [36] and Khalil and Ebner [16]. ...
... This factor has been recommended by Khalil and Ebner [16] to help in predicting at-risk students who have not completed the requested activities; and, (5) Forum and chat interactions which refers to students' participation in online discussions, such as the number of posts read, posts created and replies. This factor has been often used by Liu et al. [36] and Khalil and Ebner [16]. ...
Chapter
Online learning is gaining increasing attention by researchers and educators since it makes students learn without being limited in time or space like traditional classrooms. Particularly, several researchers have also focused on gamifying the provided online courses to motivate and engage students. However, this type of learning still faces several challenges, including the difficulties for teachers to control the learning process and keep track of their students’ learning progress. Therefore, this study presents an ongoing project which is a gamified intelligent Moodle (iMoodle) that uses learning analytics to provide dashboard for teachers to control the learning process. It also aims to increase the students’ success rate with an early warning system for predicting at-risk students, as well as providing real-time interventions of supportive learning content as notifications. The beta version of iMoodle was tested for technical reliability in a public Tunisian university for three months and few bugs were reported by the teacher and had been fixed. The post-fact technique was also used to evaluate the accuracy of predicting at-risk students. The obtained result highlighted that iMoodle has a high accuracy rate which is almost 90%.
... Analytics dashboards for learners and instructorsoften already integrated into the general design of MOOC platforms such as Coursera, FutureLearn and edXhave been proposed as viable means to enable personalized, immediate, and accurate feedback at the large scale typically found in MOOCs (Pérez-Sanagustin et al. 2021;Vigentini, Clayphan, and Chitsaz 2017). That is, analytics dashboards for learners on these platforms generally present data about personal performances (e.g., proportions of correct quiz answers) based on the assumption that this constitutes personalized 'real time' feedback that inspires learners to reflect on and regulate their own learning (ibid.; Khalil and Ebner 2016). Alternatively, analytics dashboards for instructors generally present data about (sub)groups of learners (e.g., completion rates, or average grades on quizzes among clusters of (in)active learners), which is framed as 'pedagogically neutral' data to inform decisions about feedback (e.g., targeted messages to learners) or other interventions (e.g., adjustments in learning materials) (Eradze and Tammets 2017;cf. ...
... Knox 2018). A popular way to render the operations of these analytics dashboards is through the model of the feedback loop that displays how dashboards 'feed back' data to instructors and/or learners after a smooth and singular flow of data processing through algorithms (Khalil and Ebner 2016;Vigentini, Clayphan, and Chitsaz 2017). These models assume that learners and instructors 'close the loop' by acting on feedback displayed on the analytics dashboard. ...
Article
This paper examines the enactment of feedback in Massive Open Online Courses (MOOCs), focusing on analytics dashboards. Building on scholarship that recognizes data practices as entangled and ‘messy’, the paper problematizes the model of the feedback loop that assumes that analytics dashboards ‘feed back’ data to instructors and/or learners through a singular flow of data processing. By setting out an empirical study that focuses on four MOOCs, in two universities and on two platforms, the paper maps where, when, and how design teams, instructors, and learners are involved in the enactment of feedback through and beyond analytics dashboards. The findings draw on visualizations that highlight complex relations among people and technologies, which include multiple ‘loops’. The paper concludes with questioning the need to capture feedback in a singular loop and suggests prioritizing continued attention to the roles and responsibilities of people – educational designers, instructors, and learners – in MOOCs.
... Learning analytics is widely applied on traditional-based learning environments and also computer-based education such as adaptive educational hypertext systems and learning management systems [15][16][17][18][19][20]. Apart from the analysis of data on test scores, students demographic and educational psychological questionnaires, learning analytics also uses finegrained data from learning browsing patterns, number of attempts, clicks discussion forum participation, computer mouse, facial reactions, and online chats, textual and numerical reviews on the quality of courses. ...
... These metrics are provided in Eqs. (13)- (16). We compared our results with the other classification techniques, PCA-ANN, K-Nearest Neighbor (K-NN), Decision Trees (DT), Support Vector Machines (SVM) and Naïve Bayes (NB). ...
Article
Full-text available
Computational intelligence approaches have proven to be effective in enhancing online learning systems. Although many studies have been conducted to reveal the learners’ satisfaction in online learning platforms, the use of machine learning in the analysis of big datasets for this aim has rarely been explored. In addition, although the analysis of online reviews on courses has been carried out in other fields, there are very few contributions in the area of online learning platforms. This study, therefore, aims to perform learner satisfaction analysis through the use of machine learning. We develop a new method using text mining and supervised learning techniques with the aid of the ensemble learning approach. A boosting approach, AdaBoost, is used in ANN for ensemble learning to improve its performance. We employ Artificial Neural Network (ANN) approach, dimensionality reduction and Latent Dirichlet Allocation (LDA) for textual data analysis. Principal Component Analysis (PCA) is used for data dimensionality reduction. We perform several experimental evaluations on the big datasets obtained from the online learning platforms. The accuracy and computation time of the proposed method are assessed on the obtained dataset. The method is compared with several machine learning approaches to show its effectiveness in big datasets analysis. The results showed that the method is effective in predicting learners’ satisfaction from online reviews. In addition, the proposed method outperform other classifiers, K-Nearest Neighbor (K-NN), Decision Trees (DT), Support Vector Machines (SVM) and Naïve Bayes (NB), in case of accuracy. The results are discussed and research implications from different perspectives are provided for future developments of educational decision support systems.
... Чередниченко [16] ("социализации человека") и др. Проблемы влияния IT и сетевых технологий на качество обучения, социализации личности и развитие его личной культуры, организации межличностной коммуникации рассмотрены в работах [17][18][19][20][21] и др. ...
... Learning Analytics is concerned with techniques to process and interpret the rich content of information stored in databases as a result of digital and digitized systems, MOOCs for example [4]. Central to Learning Analytics is the collection, analysis and use of student data. ...
Conference Paper
Full-text available
Since its emergence in 2011, the field of Learning Analytics demands tools that deal with the exhausts of digital learning systems. This paper presents our first prototype 'OXALIC' in an attempt to introduce a standalone Learning Analytics tool for the Open edX MOOC platform. Open edX is largely used by thousands of organizations around the world. Nonetheless, one of the most challenging issues of employing Learning Analytics in Open edX platforms is having the ability to analyze "in-depth" log files. Open edX platform is deficient in providing the same features as the 'edX' system where the latter offers data packages and the prior struggles to explore advanced analytics. The paper reports on the architecture of OXALIC, functionalities, and the user interface. We foresee promising results for future directions of OXALIC as a rigid contribution to Learning Analytics in MOOCs.
Article
Full-text available
Im Beitrag werden die Verbreitung und Nutzung von „Massive Open Online Courses“ (MOOCs) an österreichischen Hochschulen aus der Perspektive von Studierenden und Universitäten skizziert. Dazu wurden zwei Analysen durchgeführt. (1) Zum einen wurden Studienanfänger:innen der TU Graz zu ihren Erfahrungen mit Online-Kursen befragt (N=1.207). Die Ergebnisse zeigen einen deutlichen Zuwachs der Online-Lernvorerfahrungen, insbesondere im Vergleich vor und nach COVID-19. Zum anderen erfolgte (2) eine Analyse der aktuellen Leistungsvereinbarungen aller öffentlichen österreichischen Universitäten. Die Nennung der Begriffe MOOCs bzw. iMooX.at in mehr als der Hälfte der Leistungsvereinbarungen verdeutlicht den zunehmenden Stellenwert von MOOCs.
Chapter
Online learning has proved its effectiveness in the last few years among a wide range of learners. Massive Open Online Courses (MOOCs) have revolutionized the shape of learning because they are considered to be a substitutional tool to the conventional educational system for many reasons, such as flexibility in timing and eliminating the economic and geographical constraints to the learners. MOOCs also enable learners from different cultures to communicate and share their knowledge through forums. Nevertheless, MOOCs are encountering several challenges that are required to be addressed, such as the higher dropout rates among learners at different phases of the course, and reduction in participation level of learners. In this chapter, we aim to address the most familiar four challenges and enhance the MOOCs experience through providing a framework of integrating a Learning Analytics technique and Intelligent Conversational Agent (LAICA) to improve the MOOCs experience for learners and educators.
Research
Full-text available
Introduction: Learning analytics(LA) is an evolving field which utilises analytic tools like BI, social media data analytics, etc., to improve learning and education. It is basically the measurement, collection, analysis and reporting of data about learners and their contexts, for purposes of understanding and optimizing learning and the environments in which it occurs. This process generates lots of data that need to analyzed and put to fruitful use so that learning can be enhanced. Based on literature review, this paper presents the need for learning estimation on MOOC Courses using LA. Methodology: A four stage process is followed to formulate this study as a literature review and gap identifications that concerns Learning Analytics in MOOC. The stages includes extensive search of related literature from reputed national and international academic works, studying the literature and selecting the principal studies, examining and tabulation of various expects of studies, like objectives, analysis tools used, major findings, etc. and reporting the review in form of gap analysis. Findings and discussion: Findings suggests that apart from LA approaches, its implementation in education and learning is more predominant objective. Sophisticated predictive and regression models are applied on gathered data for obtaining better meaningful insights. But no significant study focuses on Management education or Indian Context or both. Findings shows that all online courses are doing traditional evaluation using online assignments, tests etc. to issue certificates. It is found that there are still numerous unexplored areas related to LA and MOOC method to estimate/calculate/ proposes the quantification of learning of students/learners in MOOC courses.
Book
Game-based learning environments and learning analytics are attracting increasing attention from researchers and educators, since they both can enhance learning outcomes. This book focuses on the application of data analytics approaches and research on human behaviour analysis in game-based learning environments, namely educational games and gamification systems, to provide smart learning. Specifically, it discusses the purposes, advantages and limitations of applying such approaches in these environments. Additionally, the various smart game-based learning environments presented help readers integrate learning analytics in their educational games and gamification systems to, for instance, assess and model students (e.g. their computational thinking) or enhance the learning process for better outcomes. Moreover, the book presents general guidelines on various aspects, such as collecting data for analysis, game-based learning environment design, system architecture and applied algorithms, which facilitate incorporating learning analytics into educational games and gamification systems. After a general introduction to help readers become familiar with the subject area, the individual chapters each discuss a different aim of applying data analytics approaches in educational games and gamification systems. Lastly, the conclusion provides a summary and presents general guidelines and frameworks to consider when designing smart game-based learning environments with learning analytics.
Chapter
Full-text available
Over the last ten years learning analytics (LA) has grown from a hypothetical future into a concrete field of inquiry and a global community of researchers and practitioners. Although the LA space may appear sprawling and complex, there are some clear through-lines that the new student or interested practitioner can use as entry points. Four of these are presented in this chapter, 1. LA as a concern or problem to be solved, 2. LA as an opportunity, 3. LA as field of inquiry and 4. the researchers and practitioners that make up the LA community. These four ways of understanding LA and its associated constructs, technologies, domains and history can hopefully provide a launch pad not only for the other chapters in this handbook but the world of LA in general. A world that, although large, is open to all who hold an interest in data and learning and the complexities that follow from the combination of the two.
Article
Full-text available
Learning analytics has reserved its position as an important field in the educational sector. However, the large-scale collection, processing, and analyzing of data has steered the wheel beyond the borders to face an abundance of ethical breaches and constraints. Revealing learners’ personal information and attitudes, as well as their activities, are major aspects that lead to identifying individuals personally. Yet, de-identification can keep the process of learning analytics in progress while reducing the risk of inadvertent disclosure of learners’ identities. In this paper, the authors discuss de-identification methods in the context of the learning environment and propose a first prototype conceptual approach that describes the combination of anonymization strategies and learning analytics techniques.
Conference Paper
Full-text available
Massive Open Online Courses are remote courses that excel in their students' heterogeneity and quantity. Due to the peculiarity of being massiveness, the large datasets generated by MOOCs platforms require advance tools to reveal hidden patterns for enhancing learning and educational environments. This paper offers an interesting study on using one of these tools, clustering, to portray learners' engagement in MOOCs. The research study analyse a university mandatory MOOC, and also opened to the public, in order to classify students into appropriate profiles based on their engagement. We compared the clustering results across MOOC variables and finally, we evaluated our results with an eighties students' motivation scheme to examine the contrast between classical classes and MOOCs classes. Our research pointed out that MOOC participants are strongly following the Cryer's scheme of Elton (1996).
Conference Paper
Full-text available
Massive Open Online Courses (MOOCs) have been tremendously spreading among Science, Technology, Engineering and Mathematics (STEM) academic disciplines. These MOOCs have served an agglomeration of various learner groups across the world. The leading MOOCs platform in Austria, the iMooX, offers such courses. This paper highlights authors’ experience of applying Learning Analytics to examine the participation of secondary school pupils in one of its courses called “Mechanics in everyday life”. We sighted different patterns and observations and on the contrary of the expected jubilant results of any educational MOOC, we will show, that pupils seemingly decided to consider it not as a real motivating learning route, but rather as an optional homework.
Article
Recently, interest in how this data can be used to improve teaching and learning has also seen unprecedented growth and the emergence of the field of learning analytics. In other fields, analytics tools already enable the statistical evaluation of rich data sources and the identification of patterns within the data. These patterns are then used to better predict future events and make informed decisions aimed at improving outcomes (Educause, 2010). This paper reviews the literature related to this emerging field and seeks to define learning analytics, its processes, and its potential to advance teaching and learning in online education.
Conference Paper
Massively Open Online Courses (MOOCs) have gained attention recently because of their great potential to reach learners. Substantial empirical study has focused on student persistence and their interactions with the course materials. However, most MOOCs include a rich textual dialogue forum, and these textual interactions are largely unexplored. Automatically understanding the nature of discussion forum posts holds great promise for providing adaptive support to individual students and to collaborative groups. This paper presents a study that applies unsupervised student understanding models originally developed for synchronous tutorial dialogue to MOOC forums. We use a clustering approach to group similar posts, compare the clusters with manual annotations by MOOC researchers, and further investigate clusters qualitatively. This paper constitutes a step toward applying unsupervised models to asynchronous communication, which can enable massive-scale automated discourse analysis and mining to better support students' learning.
Article
The visualization of big MOOC data enables us to see trends in student behaviors and activities around the globe, but what is it that we are not seeing?