ArticlePDF Available

Abstract

In this paper, an early intervention solution for collegiate faculty called Course Signals is discussed. Course Signals was developed to allow instructors the opportunity to employ the power of learner analytics to provide real-time feedback to a student. Course Signals relies not only on grades to predict students' performance, but also demographic characteristics, past academic history, and students' effort as measured by interaction with Blackboard Vista, Purdue's learning management system. The outcome is delivered to the students via a personalized email from the faculty member to each student, as well as a specific color on a stoplight -- traffic signal -- to indicate how each student is doing. The system itself is explained in detail, along with retention and performance outcomes realized since its implementation. In addition, faculty and student perceptions will be shared.
Course Signals at Purdue: Using Learning Analytics to
Increase Student Success
Kimberly E. Arnold
Purdue University
519 Young Hall, 155 S. Grant Street
West Lafayette, IN 47907 USA
kimarnold@purdue.edu
Matthew D. Pistilli
Purdue University
517 Young Hall, 155 S. Grant Street
West Lafayette, IN 47907 USA
mdpistilli@purdue.edu
ABSTRACT
In this paper, an early intervention solution for collegiate faculty
called Course Signals is discussed. Course Signals was
developed to allow instructors the opportunity to employ the
power of learner analytics to provide real-time feedback to a
student. Course Signals relies not only on grades to predict
students performance, but also demographic characteristics,
past academic history, and students’ effort as measured by
interaction with Blackboard Vista, Purdue’s learning
management system. The outcome is delivered to the students
via a personalized email from the faculty member to each
student, as well as a specific color on a stoplight traffic signal
to indicate how each student is doing. The system itself is
explained in detail, along with retention and performance
outcomes realized since its implementation. In addition, faculty
and student perceptions will be shared.
Categories and Subject Descriptors
J.1 [Administrative Data Processing]: Education
General Terms
Measurement, Performance
Keywords
Learning Analytics, College Student Success, Early
Intervention, Retention
1. INTRODUCTION
The first year of college is arguably the most critical with regard
to the retention of students into subsequent years of study [2, 3,
8, 9]. Noel and Levitz indicate that retention, or the lack of
attrition from college, is a by-product of student success [4].
Tinto has spent much of his career investigating the necessary
conditions for student success, and notes that academic support
is among the primary pieces necessary to ensure success. In his
1993 book, Leaving College, Tinto proposed three necessary
conditions for student persistence. First, an institution needed to
put programs into place that placed the welfare of the students
higher than that of the university. Second, programs and
solutions should be focused on all students at an institution, not
just a specific subpopulation. Finally, solutions implemented to
enhance student success, and therefore persistence, needed to
help integrate a student academically into the institution [6].
Helping a student become academically integrated to the
institution is key, as Course Signals helps to promote integration
in several ways. First, it allows faculty members to send
personalized emails to students that contain information about
their current performance in a given course. Second, faculty
members can encourage students to visit various help resources
on campus or office hours activities that contribute to a
student becoming more fully integrated into the institution.
Third, it employs learner analytics to allow for the integration of
real-time data on student performance and interaction with the
LMS with demographic and past academic history information.
This combination creates an intentionally created environment
for the students that does ―not leave learning to chance,‖
something Tinto noted was necessary to ensure that a solution
would be broadly effective in helping students persist to
graduation [7]. The remainder of this paper will describe Course
Signals in detail, including its development and outcomes
realized as a result of its implementation. In addition, faculty
and student perceptions will be shared.
2. COURSE SIGNALS OVERVIEW
2.1 Description of Course Signals
Course Signals (CS) is a student success system that allows
faculty to provide meaningful feedback to student based on
predictive models. The premise behind CS is fairly simple:
utilize the wealth of data found at an educational institution,
including the data collected by instructional tools, to determine
in real time which students might be at risk, partially indicated
by their effort within a course. Through analytics, large data sets
are mined and statistical techniques are applied to predict which
students might be falling behind. The goal is to produce
―actionable intelligence‖ in this case, guiding students to
appropriate help resources and explaining how to use them. [1]
A predictive student success algorithm (SSA) is run on-demand
by instructors. CS works by mining data from multiple
university sources and subsequently transforming the data into a
generated risk level with supporting information for each
student [1]. The algorithm that predicts students’ risk statuses
has four components: performance, measured by percentage of
points earned in course to date; effort, as defined by interaction
with Blackboard Vista, Purdue’s LMS, as compared to students’
peers; prior academic history, including academic preparation,
high school GPA, and standardized test scores; and, student
Permission to make digital or hard copies of all or part of this work for
personal or classroom use is granted without fee provided that copies are
not made or distributed for profit or commercial advantage and that
copies bear this notice and the full citation on the first page. To copy
otherwise, or republish, to post on servers or to redistribute to lists,
requires prior specific permission and/or a fee.
LAK’12, 29 April – 2 May 2012, Vancouver, BC, Canada.
Copyright 2012 ACM 978-1-4503-1111-3/12/04…$10.00.
characteristics, such as residency, age, or credits attempted.
Each component is weighted and pulled into the proprietary
algorithm, which then calculates a result for each student. Based
on results of the SSA, a red, yellow or green signal is displayed
on a student’s course homepage. A red light indicates a high
likelihood of being unsuccessful; yellow indicates a potential
problem of succeeding; and a green signal demonstrates a high
likelihood of succeeding in the course. Instructors then
implement an intervention schedule they create, possibly
consisting of:
Posting of a traffic signal indicator on a student’s
LMS home page;
E-mail messages or reminders;
Text messages;
Referral to academic advisor or academic resource
centers; or,
Face to face meetings with the instructor. [1]
With Course Signals, students are not placed at risk due to one
single factor; risk is determined by a contextualized landscape
that varies from student to student based on the data comprising
the four components of the SSA. The SSA transforms both static
and dynamic data points into a single score, improving the
reliability of the prediction. Since a course-specific risk
indicator is created for each student based on performance, peer-
based behavior, and educational preparation data, instructors
can intervene early and give students a realistic opportunity to
adapt their behavior to be more specific in a given course.
2.2 History of Course Signals
Facing challenges of under prepared students, budget crises,
decreasing retention and longer graduation periods, higher
education is working to provide solutions to these challenges
while at the same time balancing the demands of providing
exceptional student service to foster student success. In an
attempt to ease these mounting pressures, CS was developed to
help identify students potentially at risk of not reaching their full
potential in a course. Once identified, instructors have the ability
to deliver meaningful interventions suggesting behaviors a
student may wish to change in order to improve her chances of
success. [1]
In 2007, Purdue University piloted Course Signals. According
to Pistilli and Arnold, the system ―was built from the ground up
using empirical data at every stage to ensure the most predictive
student success algorithm [5]. Course Signals became
automated in spring 2009 and partnered with SunGard Higher
education in October 2010 in order to help other institutions
harness the power of learning analytics. Today, nearly 24,000
students have been impacted by the CS project, and more than
145 instructors have used CS in at least one of their courses.
3. ACADMIC PERFORMANCE AND
RETENTION OUTCOMES
3.1 Impact on Academic Performance
Undeniably, one performance measure of student success is final
course grade. Research indicates that courses that implement CS
realize a strong increase in satisfactory grades, and a decrease in
unsatisfactory grades and withdrawals. Individual courses see
variable success with: an increase in As and Bs ranging from
2.23 to 13.84 percentage points; a decrease in Cs ranging from
1.84 to 9.38 percentage points; and a decrease in Ds and Fs
ranging from 0.59 to 9.40 percentage. Combining the results
of all courses using CS in a given semester, there is a 10.37
percentage point increase in As and Bs awarded between CS
users and previous semesters of the same courses not using CS.
Along the same lines, there is a 6.41 percentage point decrease
in Ds, Fs, and withdrawals awarded to CS users as compared to
previous semesters of the same courses not using CS.
3.2 Impact on Student Retention
With increased student success in individual courses comes an
expected increase in retention to the University as well, and the
data indicate this nicely. Course Signals has been employed at
Purdue since 2007, and its use for each beginning cohort at the
institution is described in detail below.
3.2.1 Methodology
The fall 2007, 2008, and 2009 beginner cohorts were compared
to a master list of all Course Signals participants to determine
who from those entering cohorts took courses utilizing Course
Signals or not, and, if applicable, the number of times students
took courses with Course Signals. From there, students were put
through the retention module in the University’s data reporting
system. They were analyzed based on the number of times they
had a course with CS a number ranging from zero to five.
The beginner cohort for each year consists of the students who
are both first-time in college and carrying full-time credit loads.
The students who comprise each cohort is determined the
second week of each fall semester, and this data set is frozen;
once created, students do not leave the cohort for any reason.
Each semester, every student in each cohort has some form of
retention or leaving behavior from simply remaining enrolled,
to graduating, to being academically dropped or withdrawing
voluntarily. The retention rate is calculated by adding those who
are still enrolled and who have graduated and dividing that sum
by the total number of first-time full-time students in the
original cohort.
3.2.2 Results
As indicated below, the students who began at Purdue in fall
2007 (Table 1), 2008 (Table 2), or 2009 (Table 3) and
participated in at least one Course Signals course are retained at
rates significantly higher than their peers who had no Course
Signals classes but who started at Purdue during the same
semester. Further, students who have two or more courses with
CS are consistently retained at rates higher than those who had
only one or no courses with Signals. The analysis detailed in
Tables 1, 2, and 3 does not account for when a student had a
course with CS, only that at some point during their academic
career they did. Tables 4, 5, and 6 examine that aspect.
For the 2007, 2008 and 2009 cohorts, instances of CS use across
successive semesters was compared to students’ retention
behavior for the following semester. This analysis asked if,
within a set of semesters, if a student had at least one course
with Course Signals. So, for example, for the 2007 cohort, the
first row looks at whether or not a student had CS in a course
during either the Fall 2007 or Spring 2008 semester, then
determines if they were retained into the Fall 2008 semester.
The comparison is against students who did not have a course
with CS during the same time period. In short, there is a
noticeable impact on students having a course with CS early in
their academic career; basically, the earlier a student encounters
CS the better. Combined with the first analysis, the earlier and
the more occurrences, the greater the likelihood students will be
retained.
Table 1. Retention Rate for the 2007 Entering Cohort
Number
of CS
Courses
Cohort
Size
Year of Retention
1 Year
2 Year
3 Year
4 Year
No CS
5,134
83.44%
73.14%
70.47%
69.40%
At least 1
1,518
96.71%
94.73%
90.65%
87.42%
1 instance
1,311
96.57%
94.13%
89.70%
86.50%
2 or more
207
97.58%
98.55%
96.62%
93.24%
Number
of CS
Courses
Cohort
Size
Year of Retention
1 Year
2 Year
3 Year
No CS
4,221
81.69%
75.08%
73.21%
At least 1
2,690
96.25%
89.55%
85.17%
1 instance
2,125
95.62%
88.00%
83.58%
2 or more
565
98.58%
95.40%
91.15%
Table 3. Retention Rate for the 2009 Entering Cohort
Number of
CS
Courses
Cohort
Size
Year of Retention
1 Year
2 Year
No CS
3,164
87.67%
81.89%
At least 1
2,962
90.34%
83.22%
1 instance
2,296
87.72%
80.87%
2 or more
666
99.40%
91.44%
Table 4. Analysis of Retention by Semester of Course Signals
Use for the 2007 Entering Cohort
Comparison to Students without CS in
Same Time Period
χ
2
value
P-value
CS in First Two Terms Retained to Third
18.57
1.64E-05
CS in First Three Terms Retained to Fourth
35.10
3.13E-09
CS in First Four Terms Retained to Fifth
131.95
< 2.2e-16
CS in First Five Terms Retained to Sixth
1073.18
< 2.2e-16
CS in First Six Terms Retained to Seventh
2.32
0.1278*
CS in First Seven Terms Retained to Eighth
725.57
< 2.2e-16
* Not significant
Table 5. Analysis of Retention by Semester of Course Signals
Use for the 2008 Entering Cohort
Comparison to Students without CS in
Same Time Period
χ
2
value
P-value
CS in First Two Terms Retained to Third
1.23
0.267*
CS in First Three Terms Retained to Fourth
2234.7
< 2.2e-16
CS in First Four Terms Retained to Fifth
131.95
0.5348*
CS in First Five Terms Retained to Sixth
1611.42
< 2.2e-16
* Not significant
Table 6. Analysis of Retention by Semester of Course Signals
Use for the 2009 Entering Cohort
Comparison to Students without CS in
Same Time Period
χ
2
value
P-value
CS in First Two Terms Retained to Third
309.67
< 2.2e-16
CS in First Three Terms Retained to Fourth
362.31
< 2.2e-16
In addition to the previous analyses, it should be noted that in
every case for the students from the 2007, 2008, and 2009
cohorts, students in courses with CS have a lower average
standardized test scores than those in non-CS courses. While
this aspect needs to be further investigated, early indications
show that lesser-prepared students, with the addition of CS to
difficult courses, are faring better with academic success and
retention to Purdue than their better-prepared peers in courses
not utilizing Course Signals.
4. FEEDBACK FROM STUDENTS AND
INSTRUCTORS
While the quantitative data provide evidence that there is an
impact on students’ grades and retention behavior, there exist
additional data that support the use of CS. The instructors who
have employed CS, as well as students who have benefited from
the system, have provided information via surveys, focus
groups, and interviews that continue to warrant the usage of the
system.
4.1 Student Perception and Feedback
One of the major objectives of academic analytics is to identify
underperforming students and intervene early enough to allow
them the opportunity to change their behavior. For this reason
the Course Signals development team has closely tracked the
student experience with Signals since the pilot stage. At the end
of each semester, a user survey gathers anonymous feedback
from students, with more than 1,500 students surveyed across
five semesters. In addition, several focus groups have been held.
Students report positive experiences with Course Signals overall
(89% of respondents stated CS provided a positive experience
and 58% said they would like to use CS in every course). Most
students perceive the computer-generated e-mails and warnings
as personal communication between themselves and their
instructor. The e-mails seem to minimize their feelings of ―being
just a number,‖ which is particularly common among first-
semester students. Students also find the visual indicator of the
traffic signal, combined with instructor communication, to be
informative (they learn where to go to get help) and motivating
(74% said their motivation was positively affected by CS) in
changing their behavior.
Of the roughly 1,500 student responses, only two wrote of
becoming demoralized by the ―constant barrage‖ of negative
messages from their instructor. While this perception should not
be downplayed, negative feedback from instructors, especially
for students who might not be prepared for the rigors of higher
education, can be difficult to receive. Aside from these two
instances, however, the remainder of the negative feedback
concerned faculty use of the tool. For example, many students
spoke of over penetration (e-mails, text messages, and LMS
messages all delivering the same message), stale traffic signals
on their home pages (an intervention was run but not updated,
giving a false impression of a student’s status), and a desire for
even more specific information. This information
notwithstanding, the overwhelming response from the students
is that Course Signals is a helpful and important tool that aids in
their overall academic success at Purdue. Faculty believe this as
well, as indicated in the following section.
4.2 Faculty Perception and Feedback
While the student success algorithm predicts which students
might be in jeopardy of not doing as well as they could in a
course, it is the faculty and instructors who use the information
provided by Course Signals to intervene. It could certainly be
argued that these instructors, armed with the data provided by
learner analytics, are the most important weapons against
student under performance in the classroom. To wit, one
instructor asserted that ―I want my students to perform well, and
knowing which ones need help, and where they need help,
benefits me as a teacher.‖
Faculty have easy access to CS data via the faculty dashboard.
Using learner analytics, faculty can provide action-oriented and
helpful feedback much earlier in the semester, which students
appreciate. This particularly benefits students early in their
academic careers, as they often are not fully aware of the
behaviors they must exhibit or actions they must take in order to
be successful.
Faculty also say that students tend to be more proactive as a
result of the Course Signals interventions. While students still
tended to procrastinate, they began thinking of big projects and
assignments earlier. Instructors and TAs also noticed that
students posted more questions about assignment requirements
well before the due dates. Because of the ability of academic
analytics to assess risk early and in real time, the instructors
consistently indicate that students are benefiting from knowing
how they are really doing in a course and, moreover, understand
the importance of completing assignments, and performing well
on quizzes and tests.
In general, faculty and instructors have a positive response to
CS but many approached the system with caution. Before using
Course Signals, faculty initially expressed concern about floods
of students seeking help; however, few actually reported
problems after they began using the system. The most
commonly reported issue being an excess of e-mails from
concerned students. In addition, faculty reported concerns about
creating a dependency in newly arrived students instead of the
desired independent learning traits. A final faculty concern was
the lack of best practices for using CS, demonstrating that
instructors and students share the same concern about the lack of
best practices.
There is little that can be done to mitigate the first concern,
since one of the goals of CS is to have students take a more
active role in their success. The second concern is mitigated by
the strong retention results discussed in this paper. The final
issue was addressed by creating and posting best practice tips at
http://www.itap.purdue.edu/learning/tools/signals.
5. CONCLUSION
The use of learner analytics through the application of Course
Signals to difficult courses has shown great promise with regard
to the success of first and second year students, as well as their
overall retention to the University. To date, over 23,000 students
across 100 courses have been impacted by Course Signals and
over 140 instructors have utilized the system. Plans call for the
expansion of CS to include as many as 20,000 students a
semester within the next 18 months, and the upper
administration the institution is in strong support of this goal.
While this analysis is not without its limitations or areas for
continued improvement of the algorithm behind CS or the use of
CS, the outcomes are such that continued use of Course Signals
as a means of helping instructors provide detailed feedback to
their students, and to ultimately assist students in their academic
endeavors is highly warranted.
6. ACKNOWLEDGMENTS
Our thanks to Kyungmin ―Mike‖ Ahn for his work analyzing the
data associated with the retention numbers.
7. REFERENCES
[1] Arnold, K. E. 2010. Signals: Applying academic analytics.
EDUCAUSE Quarterly, 33, 1.
www.educause.edu/library/EQM10110
[2] Barefoot, B. O., Gardner, J. N., Cutright, M., Morris, L. V.,
Schroeder, C. C., Schwartz, S. W., Siegel, M. J., Swing, R. L.
2005. Achieving and sustaining institutional excellence for the
first year of college. Jossey-Bass, San Francisco, CA.
[3] Kuh, G. D., Kinzie, J., Schuh, J. H., Whitt, E. J., and
Associates. 2005. Student success in college: Creating
conditions that matter. Jossey-Bass, San Francisco.
[4] Noel, L., Levitz, R., Saluri, D., and Associates. 1985.
Increasing student retention: Effective programs and practices
for reducing the dropout rate. Jossey-Bass, San Francisco.
[5] Pistilli, M.D., Arnold, K. E. 2010. Purdue Signals: Mining
real-time academic data to enhance student success. About
campus: Enriching the student learning experience, 15, 3, 22-
24.
[6] Tinto, V. 1993. Leaving college: Rethinking the causes and
cures of student attrition (2nd Ed.). University of Chicago Press,
Chicago.
[7] Tinto, V. 2005. College student retention: Formula for
success. In: A. Seidman, (ed.): College student retention:
Formula for student success, American Council on Education
and Praeger, Westport, CT.
[8] Upcraft, M. L., Gardner, J. N., and Associates. 1989. The
freshman-year experience: Helping students survive and
succeed in college. Jossey-Bass, San Francisco.
[9] Upcraft, M. L., Gardner, J. N., Barefoot, B. O., and
Associates. 2004. Challenging and supporting the first-year
student: A handbook for improving the first year of college.
Jossey-Bass, San Francisco.
... Timely completion of a degree program not only reflects the effectiveness of an educational institution in running its process but also significantly impacts students' academic and career trajectories [1], [2], [3]. On-time graduation has been linked to better employment prospects, higher earning potential, and better overall life satisfaction [4], [5]. Conversely, delays in graduation can lead to increased financial burdens, extended time in education, and more likely potential disengagement from academic goals [6], [7]. ...
Article
Full-text available
Graduating on schedule is a critical milestone for students in higher education, serving as a key indicator of both institutional effectiveness and student success. This study uses machine learning techniques to predict on-time graduation in Indonesian higher education. A dataset comprising 133 students from an engineering department over four academic years (2019–2023) was analyzed using the CRISP-DM framework. The research employed nine machine learning models, including Random Forest, Logistic Regression, Neural Networks, etc., to identify key predictors of on-time graduation. The result showed that Random Forest outperformed other models by achieving an accuracy of 85% and an AUC of 0.875. Additionally, the study developed a learning analytics dashboard to visualize predictive insights, offering actionable data for educators and administrators. The system’s performance was evaluated based on functionality, usability, efficiency, and reliability as the key intersecting factors from ISO/IEC 25010 and WebQEM frameworks, validating its quality and relevance for practical educational use. The result demonstrated high functionality, efficiency, and reliability, and positive usability feedback was received from both students and educators. The findings highlight the top ten important factors, such as cumulative GPA (CGPA), extracurricular involvement, programming, and social science courses, that predict on-time graduation, providing valuable insights for enhancing student outcomes in Indonesian higher education.
... Ensuring transparency and obtaining informed consent is crucial to maintain trust and respect individual privacy (Antoniou et al., 2012). Data analytics and business intelligence are not limited to number-crunching (Arnold & Pistilli, 2012). Data visualization plays a crucial role in presenting data in a comprehensible manner (Reich, 2019). ...
Book
Digital Transformation in Higher Education is a pivotal reference through the transformative power of emerging technologies in academia. Addressing the dual nature of technology as both a challenge and an opportunity, this book presents a rich overview of strategies for integrating digital technology-driven advancements. From chatbots enhancing adult learning to AI ethics and micro-credentials, it illuminates a future where digital transformation is key for the unique value proposition of higher education. Each chapter offers actionable insights and case studies, setting the stage for educators and leaders to harness digital technology's potential and lead with innovation.
Chapter
The development of new technologies in the 21st century has given researchers, instructors, and institutions the opportunity to track and store learners’ data in virtual learning environments. Large quantities of data are stored in learning management systems (e.g., Moodle). Thus, recent studies and researches are interested in collecting and analysing those data for improving the quality of learning. This process is what is known as ‘learning analytics,’ which is an emerging field in education. Learning analytics (LA) helps in studying the patterns of interactions among learners themselves as well as interactions with instructors and course content. It helps also in measuring the degree of learners’ presence in the platform and the frequency of interaction. Such information helps to identify areas of improvement and to enhancing learners’ engagement levels. The current study investigates Moroccan students’ perceptions and expectations with regard to the implementation of learning analytics in higher education and the related privacy issues. It is an emerging field in higher education and little is known about students’ perceptions and expectations, mainly in the Moroccan context. The obtained findings revealed that university students have distinct levels of awareness about the concept of LA. They have a strong desire that it will help them personalize their learning experience as well as supporting their learning journey through feedback and suitable interventions. The findings have also revealed the students’ concern about their data privacy and the necessity for asking for their consent and being transparent about data handling and data accuracy. The findings aim to inform university stakeholders about students’ perspectives which will help tailoring instructional strategies and guiding the development of best practices for the implementation of LA in higher education. Globally, international learners benefit from LA implementation for a full integration within the host universities and for easy intervention when facing challenges. Informing university stakeholders offers numerous insights and the implementation of LA requires stakeholders’ awareness.
Article
Pocos estudios con Analítica del Aprendizaje han intentado predecir los resultados de conjunto de un año académico. Esta investigación desarrolló un modelo predictivo del riesgo de suspender el primer año en un Grado de Negocios (i.e., obtener menos créditos de los necesarios para aprobar), utilizando Regresión Logística con datos de dos cohortes de estudiantes (n=1046). El modelo utiliza la tasa de asistencia, calificaciones de evaluación continua de tres asignaturas y los clics en la LMS, del primer semestre del año. Se probó con 74 estudiantes de una cohorte diferente, obteniendo una precisión alta, pero una sensibilidad baja. Los resultados sugieren que la asistencia a clases, las habilidades de comunicación y la competencia numérica son transversales al éxito académico. El artículo revela una relación no lineal entre la actividad en la LMS y la media académica y propone un método para tratarlo.
Chapter
In the wake of the COVID-19 pandemic, educational institutions globally face unprecedented challenges, with data analytics emerging as a vital tool to bridge learning gaps across K-12 and higher education. This chapter explores how data analytics can transform education by personalizing learning, optimizing instructional strategies, and assessing academic progress. With access to vast amounts of data, institutions can make informed, data-driven decisions in curriculum design and resource allocation. By utilizing data sources such as demographics, engagement metrics, and performance data, educators gain insights into learning trends and outcomes. The chapter discusses practical applications, including adaptive learning tools and predictive analytics for early interventions to improve student retention. It also addresses challenges related to data privacy, equity, and ethical use. By embracing emerging technologies like AI, blockchain, and VR, educational institutions can foster a culture of continuous improvement, enhancing resilience and educational outcomes on a global scale.
Chapter
Bu bölümde, öğrenme analitiklerinde idiografik yaklaşımın önemini ve potansiyeli incelenmektedir. Artan çevrimiçi eğitim uygulamaları ve dijital veri bolluğu göz önüne alındığında, öğrenme analitiği uygulamalarının eğitim süreçlerini optimize etmedeki rolü giderek artmaktadır. Ancak, geleneksel değişken merkezli yaklaşımlar bireysel deneyimlerin karmaşık ve dinamik doğasını açıklamakta sınırlı kalabilmektedirler. Kişi merkezli yaklaşımlar, değişken merkezli yaklaşımlara göre biraz daha bu konuda derinlemesine araştırma imkanı sunabilsede, kişileri benzer özelliklerine göre gruplaması nedeniyle de daha az özelleştirilmiş öneriler getirebilmektedir. Bu doğrultuda bu sınırlılıklardan hareketle, idiografik (kişiye özgü) yaklaşımın öğrenme analitiğindeki rolünü ele alınmaya çalışılmıştır. Her öğrenciyi benzersiz bir birey olarak değerlendiren bu yaklaşım, kişiselleştirilmiş öneriler sunma, zamansal değişimleri analiz etme ve çok düzeyli öğrenme süreçlerini inceleme açısından önemli avantajlar sağladığı söylenebilir.
Article
In this article the argument is made that student success should be the foundation for college student retention efforts. Utilizing a “black box” approach, together with a theoretical framework derived from expert systems thinking and the application of qualitative research methods, it is shown that the knowledge base and behavioral repertoire of successful students can be identified systematically and used to help students succeed in college. The information is locally derived and locally relevant, thus insuring that successful students can be identified for a given campus. To illustrate the approach, results are presented from a study involving Chicano students in the Southwest.
Article
Academic analytics helps address the public's desire for institutional accountability with regard to student success, given the widespread concern over the cost of higher education and the difficult economic and budgetary conditions prevailing worldwide. Purdue University's Signals project applies the principles of analytics widely used in business intelligence circles to the problem of improving student success within a course and, hence, improving the institution's retention and graduation rates over time. Through its early stages, the Signals project's success has demonstrated the potential of academic analytics. Those early efforts have led to additional projects to develop: (1) Student success algorithms (SSAs) customized by course; (2) Intervention messages sent to students; and (3) New strategies for identifying students at risk. The premise behind Signals is fairly simple--utilize the data collected by instructional tools to determine in real time which students might be at risk, partially indicated by their effort within a course. Through analytics, the institution mines large data sets continually collected by these tools and applies statistical techniques to predict which students might be falling behind. The goal is to produce "actionable intelligence"--in this case, guiding students to appropriate help resources and explaining how to use them. Early reviews by administrators, faculty, and students have been positive, as has empirical data on the system's impact. The Signals system is based on a Purdue-developed SSA designed to provide students early warning--as early as the second week of the semester--of potential problems in a course by providing near real-time status updates of performance and effort in a course. Each update provides the student with detailed, positive steps to take in averting trouble. By no means is Purdue unique in its interest in academic analytics. Institutions across the world, large and small, public and private, research and teaching, have begun forays into various data source modeling strategies in an effort to find actionable data to support their goals. This article offers a snapshot of the experience at Purdue. (Contains 4 figures and 5 endnotes.)
Article
The dimensions and consequences of college student attrition and features of institutional action to deal with attrition are discussed. Patterns of student departure from individual colleges as opposed to permanent college withdrawal are addressed. After synthesizing the research on multiple causes of student leaving, a theory of student departure from college is presented based on the work of Emile Durkheim and Arnold Van Gennep. The theory proposes that student departure may serve as a barometer of the social and intellectual health of college life as much as of the students' experiences at the college. The quality of faculty-student interaction and the student's integration into the school are central factors in student attrition. Attention is directed to features of retention programs, including the time of college actions and variations in policy necessary for different types of students and colleges. It is suggested that effective retention lies in the college's commitment to students. The content, structure, and evaluation methods for assessment of student retention and departure are considered, along with the use of assessment information for developing effective retention programs. (SW)
Article
How Purdue University is changing the academic behavior of struggling students.
Student success in college: Creating conditions that matter
  • G D Kuh
  • J Kinzie
  • J H Schuh
  • E J Whitt
Kuh, G. D., Kinzie, J., Schuh, J. H., Whitt, E. J., and Associates. 2005. Student success in college: Creating conditions that matter. Jossey-Bass, San Francisco.
Challenging and supporting the first-year student: A handbook for improving the first year of college
  • M L Upcraft
  • J N Gardner
  • B O Barefoot
Upcraft, M. L., Gardner, J. N., Barefoot, B. O., and Associates. 2004. Challenging and supporting the first-year student: A handbook for improving the first year of college. Jossey-Bass, San Francisco.
Achieving and sustaining institutional excellence for the first year of college
  • B O Barefoot
  • J N Gardner
  • M Cutright
  • L V Morris
  • C C Schroeder
  • S W Schwartz
  • M J Siegel
  • R L Swing
Barefoot, B. O., Gardner, J. N., Cutright, M., Morris, L. V., Schroeder, C. C., Schwartz, S. W., Siegel, M. J., Swing, R. L. 2005. Achieving and sustaining institutional excellence for the first year of college. Jossey-Bass, San Francisco, CA.