Key Takeaways
- Using existing student data and current grade information, Purdue created Signals, a system that gives students early and frequent performance notifications and helps faculty members steer students toward additional campus resources as needed.
- The student success algorithm considers not only demographic and performance data but also data on student effort and engagement.
- Results thus far show that students who have engaged with Course Signals have higher average grades and seek out help resources at a higher rate than other students.
Like other institutions across the country, Purdue University collects huge amounts of data on students and stores it in some warehouse, possibly never to be touched again. Also like other universities, Purdue offers many “gateway” courses — large, usually entry-level courses in which students must demonstrate a level of proficiency (a grade of C or better) to move on to upper-level or later-sequenced courses. Students in these large-enrollment courses rarely know where they stand or what they need to do to succeed in the course.
Our question, then, was how to use the data collected about students, along with their current grade data, in a way that would give them early and frequent notifications about their performance in a class. In so doing, we could point them to various campus resources that could help them successfully complete a course.
Building on a History of Success
Purdue is one of the top 30 largest public universities in the United States, serving more than 72,000 students. Enrollment on its main campus in West Lafayette is approximately 40,000, including students from more than 120 countries and all 50 states. Recognized by the Association of American Universities as one of the nation’s leading research institutions, Purdue continues to focus efforts on increasing student success and excellence in the education it provides. This drive to achieve goals in both areas impelled development of Signals.
The Course Signals System
In planning our system, we considered both the predictive power of data and the linkages between student success and effort and engagement. Then, under the guidance of John Campbell, we developed Signals, a system that takes course management system (CMS; in our case, Blackboard) data into account in predicting whether a student would succeed in a given course.
Course Signals is a student success system developed on a predictive analytic model (see figure 1) that contains elements from the academic technologies and the student information system (see figure 2). The model (1:52 minutes) is behaviorally based and considers student performance, effort, and characteristics. The algorithm runs on real-time data and provides a risk indicator for each student (see figure 3). Using this risk indicator (a red, yellow, or green traffic signal) as a formative guide, faculty members can give students in their courses meaningful feedback, suggesting behaviors that students might wish to change to improve their chances of success, thus placing the emphasis squarely on action.

Figure 1. The analytic model

Figure 2. Elements affecting risk

Figure 3. Familiar stoplight risk indicator
Faculty members Larry Deboer, Agriculture; Frank Dooley, Agricultural Economics; and Jenna Seabold, General Studies, explain when they run Signals interventions (2:34 minutes):
Faculty members Jenna Seabold, General Studies; Frank Dooley, Agricultural Economics; and Larry Deboer, Agriculture, explain how long it takes to run an intervention (0:58 minute):
Acquiring Data — and Faculty Support
Initially, our biggest difficulty was the acquisition of data. As a siloed university, different data was owned by different stewards, and compiling all the data needed for the student success algorithm was an uphill battle that took well over a year. However, now that we have achieved this, our most pressing concern is Course Signals implementation.
Over the semesters, we have developed data-derived best practices and effective messaging solutions. The research is sound and the practices are simple; however, getting all faculty members on board with those practices is a challenge. For example, we have found that one of the most effective uses of Course Signals is to intervene early and frequently. The average number of interventions sent within successful courses ranges from 4.8 to 6.2, depending on semester. However, faculty are already carrying intense workloads, and intervening more than two times is not as common as we would like (the overall average is 2.3).
A final challenge we faced was in finding a way to explain to students how their risk indicator was calculated. For this, we created a short video detailing the components of the student success algorithm.
Video explaining algorithm success components (2:16 minutes):
Impact on Grades and Help-Seeking Behavior
Course Signals has produced numerous positive results.1 For example, when grade results from courses using Course Signals were compared to those of a previous semester (same course, same term, same professor), significant differences were found. As Table 1 shows, collectively, students have earned more A’s and B’s and fewer D’s and F’s.
Table 1. Grade Improvements Using Course Signals
| Course | Technology | Grades (Percentage) | ||||
|---|---|---|---|---|---|---|
|
|
| N = | A/B | C | D/F | Withdrew |
| Agronomy | Signals | 78 | 80.77 | 14.10 | 2.56 | 3.56 |
|
| Comparison Cohort | 78 | 69.86 | 23.29 | 4.11 | 2.74 |
| Statistics | Signals | 297 | 56.9 | 27.61 | 13.8 | 1.68 |
|
| Comparison Cohort | 486 | 30.45 | 33.13 | 31.07 | 5.14 |
| Psychology | Signals | 2,263 | 65.16 | 18.22 | 13.69 | 2.93 |
|
| Comparison Cohort | 2,445 | 56.31 | 22.49 | 19.42 | 1.78 |
In the case of trackable help-seeking behavior, students who experienced Course Signals sought help earlier and more frequently than their peers (see figure 4). Anecdotal evidence indicates that students continue this behavior in future semesters. The increase in help-seeking behavior suggests that Course Signals empowers students to become better independent learners. Pattern matrix data suggests that when students do “graduate” (move up) a level — from red to yellow or green or from yellow to green — they rarely fall back into higher risk categories.

Figure 4. Biology Resource Center usage
Analysis also demonstrates that using Course Signals positively affects student retention (see table 2). Students experiencing Signals at least once in their academic career at Purdue are retained at a rate of 87.42 percent, while their peers who did not are retained at 69.40 percent — a difference of nearly 20 percentage points.
Table 2. Retention Rate after Four Years
| No. of Signals Courses | Cohort Size (N =) | 1 Year (%) | 2 Years (%) | 3 Years (%) | 4 Years (%) |
|---|---|---|---|---|---|
| No courses | 5,134 | 83.44 | 73.14 | 70.47 | 69.40 |
| 1 instance | 1,311 | 96.57 | 94.13 | 89.70 | 86.50 |
| 2 or more | 207 | 97.58 | 98.55 | 96.62 | 93.24 |
The four-year graduation rate has also seen improvement. Of those students never enrolled in a course employing Course Signals, 41.20 percent graduated after four years at the institution. In contrast, 45.27 percent of students enrolled in at least one Signals course graduated in four years, a four percentage point difference. Current research shows that students enrolled in two or more courses using Signals have a four-year graduation rate of 38.65 percent. While the number of students graduating after four years dips for students who have had Signals in two or more courses, we are not alarmed by these results. It is expected that the five-year graduation rate for students with two or more courses with Signals will surpass the other two groups when rates are released in fall 2012 for three key reasons:
- First, Signals is used largely (but not only) in gateway courses. This is pertinent because many times students will bring Advanced Placement credit for these courses or test out of them altogether. Most of the students enrolled in two or more Signals courses have to take these courses because they did not test out of them. This extends their time to degree. It is imperative to note that a larger percentage of the students taking two or more Signals courses are still enrolled for a fifth year of study than the percentage of students taking no Signals courses.
- Second, students often change majors at Purdue; this usually results in an increased time to degree, which is the case for the students here.
- Finally, many of our majors are, by design, five-year programs, which also explains part of the lower four-year graduation rate for students in two or more Signals courses. Overall, preliminary results suggest that Course Signals students graduate sooner, allowing them to enter the work world before many of their classmates (see table 3).
Table 3. Four-Year Graduation Rates
| No. of Signals Courses | Cohort Size (N =) | 4 Year (%) |
|---|---|---|
| No courses | 5,134 | 41.20 |
| At least 1 | 1,518 | 45.27 |
| 2 or more | 207 | 38.65 |
More than 100 courses taught by 162 instructors have used Course Signals to help their students become more successful, and more than 25,000 students have experienced intervention via this student success system. Additionally, faculty, students, and administration have responded favorably to Course Signals.2
Jenna Seabold, General Studies, and Frank Dooley, Department of Agricultural Economics, explain how Course Signals helps instructors (1:00 minute):
Jenna Seabold, General Studies, explains how her use of Course Signals affects her students (1:19 minutes):
Implementing and Scaling Signals
Course Signals proves that learning analytics can be successfully implemented in a scaled capacity. Further, the student success algorithm goes beyond the traditional characteristics of an early warning system and looks at effort and engagement data along with demographic and academic preparation and performance data. Course Signals is commercially available through Ellucian, making it possible for other campuses to replicate our results.
- Kimberly E. Arnold and Matthew D. Pistilli, “Course Signals at Purdue: Using Learning Analytics to Increase Student Success,” Proceedings of the 2nd International Conference on Learning Analytics, ACM Press, 2012 (in press).
- Angela van Barneveld, Kimberly E. Arnold, and John P. Campbell, “Analytics in Higher Education: Establishing a Common Language,” ELI Paper 1, EDUCAUSE Learning Initiative, Jan. 2012.
© 2012 Matthew D. Pistilli, Kimberly E. Arnold, and Matthew D. Bethune. The text of this EDUCAUSE Review Online article (July 2012) is licensed under the Creative Commons Attribution-Noncommercial-No derivative works 3.0 license.

















1 Comment
Several questions
There are a number of assumptions in this effort that inspire questions:
It appears that the measurement point in the Signals system is when the student misses a deadline, submits sub-standard work, fails to interact, or some other indicator. If we look at the measurement point as a point along a cascade of events, the causal factors - the factors we are really interested to know about - are further upstream. In other words, the actual problem that led to the event(s) of interest is unknown until an intervention is undertaken and elicited from the student. This makes the Signals system effective, but at best, a blind reaction.
The concept of "student sucess", in this context, is similar in some ways to teen pregnancy prevention. There is a combination of factors leading to the desired result, such as learning self-regulation (meta-cognitive skills), and becoming an informed, mature agent of one's sexuality (college-readiness). One strategy that has been employed is to open a dialog between at-risk individuals and those who have had experience in the area of interest. This cycle of sharing experiences to the points where others may also encounter them is the basis of a user-based learning analytics system I proposed at the Emerging Learning Design conference in Montclair, NJ this past June, 2012. It draws from Dr. Brenda Dervin's research on Sense-making Methodology and Dr. Michael Nilan's research in User-based Design.
The basis of the proposal differs from the Signals system in that it proposes that we try to close the gap between the point at which the student encounters a stopping point, cognitive shortfall, or impediment and when the information is made available to the student (when they most need help to overcome a gap in their cognitive movement). This is accomplished by collecting the experiences of those who have completed a course and mapping their experiences to the points within the course so that the meta-cognitive "lessons learned" of a large sample of prior students is made available to current students in the same course.
This proposal oversimplfies the range of causal factors, but it aspires to account for those associated with learning. More can be read on my presentation summary here: http://bit.ly/M5LTli
Steve Covello
Rich Media Specialist
Granite State College
Concord, NH
Posted by: weevie833 on August 3, 2012
Log in to comment