MOOC Dropout Prediction
Massive Online Open Courses (MOOCs)
For the tabular track, we trained machine learning models to predict if a student will drop out based on their participation in a MOOC (Massive Open Online Courses). Because of MOOCS low cost and easily accessible nature compared to traditional college education, there has been a drastic rise in interest in MOOCs recently. Despite their popularity, dropout rates remain extremely high, often exceeding 90%. When there are more than sixty-thousand students in a course, additional tools to help MOOC teachers allocate resources are extremely valuable. A machine learning model that can predict which students will complete the course and why could be just such a tool.
The Data
The MOOC dropout dataset contains a set of 42 features and a Boolean value describing whether the student completed or dropped out of the course. In accordance with Vignesh Muthukar's implementation, we narrowed the data down to 10 pertinent features:
viewed
: whether the student has ever opened the course (Average: 0.61)gender
: female = 0, male = 1 (no third option given) (Average: 0.9)grade
: student’s grade at the time of gathering (Average: 3.4%)nevents
: how much they interacted with the course (Average: 535)ndays_active
: how many days they logged in (Average: 6)nplay_video
: how many times they watched a video (Average: 58)nchapters
: how lessons they completed (Average: 2)age
: age of student (Average: 26)votes
: how many times they voted on the course forum (Average: 0.67)num_words
: how many total words they wrote in the course forum (Average: 79)
Here are a few examples from the dataset. The students from the first two rows completed the course, while the last two dropped out.
viewed | gender | grade | nevents | ndays_act | nplay_video | nchapters | age | votes | num_words |
---|---|---|---|---|---|---|---|---|---|
1 | 1 | 0.95 | 14005 | 65 | 361 | 16 | 24 | 0 | 17 |
1 | 1 | 0.85 | 4717 | 43 | 770 | 18 | 27 | 2 | 358 |
1 | 1 | 0.03 | 1529 | 19 | 18 | 6 | 20 | 0 | 15 |
1 | 1 | 0.04 | 4051 | 30 | 460 | 8 | 17 | 0 | 23 |
Our Models
While a human brain may take seconds to minutes pouring over every aspect of a student to predict their success or failure in a course, for a machine learning model this task is trivialized to a handful of milliseconds. As such, the MOOC dropout dataset has been a popular dataset for model/architecture testing. While there have been numerous models constructed to make these predictions, we have chosen to use a multi-layer perceptron (a neural network architecture) and a support-vector machine, as demonstrated by Vignesh Muthukumar (Muthukumar, 2019). Initialization of each is simple, as shown below:
# Support-Vector Machine:
from sklearn import svm
clf = svm.SVC(C=0.5, gamma='scale', probability=True)
# Multi-Layer Perceptron:
from sklearn.neural_network import MLPClassifier
mlp = MLPClassifier(hidden_layer_sizes=(3, 15, 10), max_iter=1000)
''' Alternate initialization for the MLPClassifier: '''
import keras
model = keras.models.Sequential([
keras.layers.Dense(3, input_dim=10, activation='relu'),
keras.layers.Dense(15, activation='relu'),
keras.layers.Dense(10, activation='relu'),
keras.layers.Dense(2, activation='softmax')
])
Both networks achieve good results, but they can be messy, and skews in the data can make predictions unbalanced. As such, these models make for perfect candidates to explain.
Want to get a feel for the black box model? Try querying the Multi-Layer Perceptron here:
viewed | gender | grade | nevents | ndays_act | nplay_video | nchapters | age | votes | num_words |
Why Explain These AIs?
These models are strong candidates for explainability for several reasons. First, given the notorious “black box” nature of neural networks and high-dimension SVMs, there are few options (apart from explainability methods) that offer insight into why the model may classify certain students as likely or unlikely to drop out of the MOOC. Additionally, there can be potentially significant ramifications for these students if they are predicted as likely to drop out, or not – a student flagged as likely to drop out may lose motivation, and the prediction may become a self-fulfilling prophecy in cases where the student would have otherwise stayed in the course. On the other hand, a student who is truly at risk of dropping out but not identified as such may fail to be noticed by the instructional staff, due to the large and relatively autonomous nature of MOOCs. These significant stakes call for added scrutiny and insight into models making these predictions. Finally, explainability methods applied to MOOC dropout prediction can offer insight into what features successful and unsuccessful students exhibit.