What Should I Learn First: Introducing LectureBank for NLP Education and Prerequisite Chain Learning

@inproceedings{Li2019WhatSI,
  title={What Should I Learn First: Introducing LectureBank for NLP Education and Prerequisite Chain Learning},
  author={Irene Li and Alexander R. Fabbri and R. Tung and Dragomir R. Radev},
  booktitle={AAAI},
  year={2019}
}
Recent years have witnessed the rising popularity of Natural Language Processing (NLP) and related fields such as Artificial Intelligence (AI) and Machine Learning (ML). Many online courses and resources are available even for those without a strong background in the field. Often the student is curious about a specific topic but does not quite know where to begin studying. To answer the question of "what should one learn first," we apply an embedding-based method to learn prerequisite relations… Expand
MOOCRep: A Unified Pre-trained Embedding of MOOC Entities
TLDR
MOOCRep is developed, a novel method based on Transformer language model trained with two pre-training objectives: graph-based objective to capture the powerful signal of entities and relations that exist in the graph, and domain-oriented objective to effectively incorporate the complexity level of concepts. Expand
Heterogeneous Graph Neural Networks for Concept Prerequisite Relation Learning in Educational Data
TLDR
This paper proposes a novel concept prerequisite relation learning approach, named CPRL, which combines both concept representation learned from a heterogeneous graph and concept pairwise features and extends CPRL under weakly supervised settings to make the method more practical. Expand
MOOCCube: A Large-scale Data Repository for NLP Applications in MOOCs
TLDR
A large-scale data repository of over 700 MOOC courses, 100k concepts, 8 million student behaviors with an external resource is presented and a prerequisite discovery task is conducted as an example application to show the potential of MOOCCube in facilitating relevant research. Expand
Improved Concept Embeddings for Learning Prerequisite Chains
Word embedding is one the most common techniques for feature learning in Natural Language Processing (NLP), allowing us to represent words and concepts as vectors in an ndimensional space and toExpand
R-VGAE: Relational-variational Graph Autoencoder for Unsupervised Prerequisite Chain Learning
TLDR
This paper proposes a model called the Relational-Variational Graph AutoEncoder (R-VGAE) to predict concept relations within a graph consisting of concept and resource nodes that is notably the first graph-based model that attempts to make use of deep learning representations for the task of unsupervised prerequisite learning. Expand
Linguistically-Driven Strategy for Concept Prerequisites Learning on Italian
TLDR
The paper introduces ITA-PREREQ, to the best of the authors' knowledge the first Italian dataset annotated with prerequisite relations between pairs of educational concepts, and describes the automatic strategy devised to build it. Expand
Unsupervised Cross-Domain Prerequisite Chain Learning using Variational Graph Autoencoders
TLDR
This paper proposes unsupervised cross-domain concept prerequisite chain learning using an optimized variational graph autoencoder and expands an existing dataset by introducing two new domains––CV and Bioinformatics (BIO). Expand
Heterogeneous Graph Neural Networks for Multi-label Text Classification
TLDR
A heterogeneousgraph convolutional network model is proposed to solve the MLTC problem by modeling tokens and labels as nodes in a heterogeneous graph to take into account multiple relationships including token-level relationships. Expand
Digging Into Prerequisite Annotation
TLDR
This paper investigates possible reasons for PR disagreement and the nature of PR itself, with the aim of contributing to the development of shared strategies for PR annotation, analysis and modelling in textbooks. Expand
AAN: Developing Educational Tools for Work Force Training
Millions of computing jobs remain unfilled, largely due to an insufficiently trained workforce which cannot keep up with the latest technological advances. We aim to address this problem byExpand
...
1
2
...

References

SHOWING 1-10 OF 30 REFERENCES
TutorialBank: A Manually-Collected Corpus for Prerequisite Chains, Survey Extraction and Resource Recommendation
TLDR
This work introduces TutorialBank, a new, publicly available dataset which aims to facilitate NLP education and research and is notably the largest manually-picked corpus of resources intended for NLPEducation which does not include only academic papers. Expand
Learning Concept Graphs from Online Educational Data
TLDR
This paper addresses an open challenge in educational data mining, i.e., the problem of automatically mapping online courses from different providers onto a universal space of concepts, and predicting latent prerequisite dependencies among both concepts and courses, with a novel approach for inference within and across course-level and concept-level directed graphs. Expand
Prerequisite Relation Learning for Concepts in MOOCs
TLDR
This work proposes a representation learning-based method for learning latent representations of course concepts, and investigates how different features capture the prerequisite relations between concepts. Expand
Modeling Relational Data with Graph Convolutional Networks
TLDR
It is shown that factorization models for link prediction such as DistMult can be significantly improved through the use of an R-GCN encoder model to accumulate evidence over multiple inference steps in the graph, demonstrating a large improvement of 29.8% on FB15k-237 over a decoder-only baseline. Expand
Course Concept Extraction in MOOCs via Embedding-Based Graph Propagation
TLDR
This paper proposes to learn latent representations for candidate concepts via an embedding-based method and develops a graph-based propagation algorithm to rank the candidate concepts based on the learned representations. Expand
Modeling Concept Dependencies in a Scientific Corpus
TLDR
An information-theoretic view of concept dependency is formulated and methods to construct a “concept graph” automatically from a text corpus are presented to support search capabilities that may be tuned to help users learn a subject rather than retrieve documents based on a single query. Expand
Graph Convolutional Encoders for Syntax-aware Neural Machine Translation
We present a simple and effective approach to incorporating syntactic structure into neural attention-based encoder-decoder models for machine translation. We rely on graph-convolutional networksExpand
Distributed Representations of Sentences and Documents
TLDR
Paragraph Vector is an unsupervised algorithm that learns fixed-length feature representations from variable-length pieces of texts, such as sentences, paragraphs, and documents, and its construction gives the algorithm the potential to overcome the weaknesses of bag-of-words models. Expand
Structured Generation of Technical Reading Lists
TLDR
An open-source implementation of the method of automatically generating reading lists, selecting documents based on their pedagogical value to the learner and ordering them using the structure of concepts in the domain is provided. Expand
Graph-based Neural Multi-Document Summarization
TLDR
This model improves upon other traditional graph-based extractive approaches and the vanilla GRU sequence model with no graph, and it achieves competitive results against other state-of-the-art multi-document summarization systems. Expand
...
1
2
3
...