Contextual Regression: An Accurate and Conveniently Interpretable Nonlinear Model for Mining Discovery from Scientific Data

  title={Contextual Regression: An Accurate and Conveniently Interpretable Nonlinear Model for Mining Discovery from Scientific Data},
  author={Chengyu Liu and Wei Wang},
Machine learning algorithms such as linear regression, SVM and neural network have played an increasingly important role in the process of scientific discovery. However, none of them is both interpretable and accurate on nonlinear datasets. Here we present contextual regression, a method that joins these two desirable properties together using a hybrid architecture of neural network embedding and dot product layer. We demonstrate its high prediction accuracy and sensitivity through the task of… 
Engineering a Histone Reader Protein by Combining Directed Evolution, Sequencing, and Neural Network Based Ordinal Regression
The ordinal regression model trained on only two sorts successfully predicts chromodomain CBX1 mutants that would have stronger binding affinity with the H3K9me3 peptide and can be extracted using contextual regression, a method to interpret nonlinear models, which successfully guides identification of strong binders not even present in the original library.
Biogenesis mechanisms of circular RNA can be categorized through feature extraction of a machine learning model
A contextual regression model trained to predict the formation of circular RNA from a random genomic locus on human genome solidify the idea that multiple biogenesis mechanisms co-exist for different subset of human circRNAs and uncover a potential new links between circRNA biogenesis and flanking CpG island.


"Why Should I Trust You?": Explaining the Predictions of Any Classifier
LIME is proposed, a novel explanation technique that explains the predictions of any classifier in an interpretable and faithful manner, by learning aninterpretable model locally varound the prediction.
Applied Logistic Regression
Applied Logistic Regression, Third Edition provides an easily accessible introduction to the logistic regression model and highlights the power of this model by examining the relationship between a dichotomous outcome and a set of covariables.
Predicting the sequence specificities of DNA- and RNA-binding proteins by deep learning
This work shows that sequence specificities can be ascertained from experimental data with 'deep learning' techniques, which offer a scalable, flexible and unified computational approach for pattern discovery.
Large-scale imputation of epigenomic datasets for systematic annotation of diverse human tissues
The imputed datasets provide the most comprehensive human regulatory region annotation to date, and the approach and the ChromImpute software constitute a useful complement to large-scale experimental mapping of epigenomic information.
Adam: A Method for Stochastic Optimization
This work introduces Adam, an algorithm for first-order gradient-based optimization of stochastic objective functions, based on adaptive estimates of lower-order moments, and provides a regret bound on the convergence rate that is comparable to the best known results under the online convex optimization framework.
Visualizing and Understanding Convolutional Networks
A novel visualization technique is introduced that gives insight into the function of intermediate feature layers and the operation of the classifier in large Convolutional Network models, used in a diagnostic role to find model architectures that outperform Krizhevsky et al on the ImageNet classification benchmark.
Data Mining with Decision Trees - Theory and Applications
  • L. Rokach, O. Maimon
  • Computer Science
    Series in Machine Perception and Artificial Intelligence
  • 2007
This 2nd Edition is dedicated entirely to the field of decision trees in data mining; to cover all aspects of this important technique, as well as improved or new methods and techniques developed after the publication of the first edition.
Not Just a Black Box: Learning Important Features Through Propagating Activation Differences
DeepLIFT (Learning Important FeaTures), an efficient and effective method for computing importance scores in a neural network that compares the activation of each neuron to its 'reference activation' and assigns contribution scores according to the difference.
Distributed Representations of Words and Phrases and their Compositionality
This paper presents a simple method for finding phrases in text, and shows that learning good vector representations for millions of phrases is possible and describes a simple alternative to the hierarchical softmax called negative sampling.
Long Short-Term Memory
A novel, efficient, gradient based method called long short-term memory (LSTM) is introduced, which can learn to bridge minimal time lags in excess of 1000 discrete-time steps by enforcing constant error flow through constant error carousels within special units.