Predicting student grades from online, collaborative social learning metrics using K-NN

Matthew Yee-King, Andreu Grimalt-Reynes, Mark d’Inverno

Research output: Chapter in book/report/conference proceedingConference proceedingpeer-review

5 Citations (Scopus)

Abstract

We describe a collaborative video annotation system that aims to engage learners in a focused, collaborative process of content sharing and discussion, and explain how it was used in an online creative programming MOOC on Coursera. We explore the use of K-NN (K nearest neighbour) to predict which of a variable number of evenly spaced, final grade bands students will fall into based solely on a feature vector consisting of the total number of UI click and mouseover events they generated during the course. We were able to classify students into pass/fail bands with 88% precision; with 3 grade bands, precision was 77%, going down to 31% with 10 grade bands. Typically, a feature subset containing less than half of the available features provided the best performance.

Original languageEnglish
Title of host publicationProceedings of the 9th International Conference on Educational Data Mining, EDM 2016
EditorsTiffany Barnes, Min Chi, Mingyu Feng
PublisherInternational Conference on Educational Data Mining
Pages654-655
Number of pages2
Publication statusPublished - 29 Jun 2016
Event9th International Conference on Educational Data Mining, EDM 2016 - Raleigh, United States
Duration: 29 Jun 20162 Jul 2016
https://educationaldatamining.org/EDM2016/
https://www.educationaldatamining.org/EDM2016/proceedings/edm2016_proceedings.pdf

Publication series

NameProceedings of the International Conference on Educational Data Mining

Conference

Conference9th International Conference on Educational Data Mining, EDM 2016
Country/TerritoryUnited States
CityRaleigh
Period29/06/162/07/16
Internet address

Fingerprint

Dive into the research topics of 'Predicting student grades from online, collaborative social learning metrics using K-NN'. Together they form a unique fingerprint.

Cite this