#### Filter Results:

- Full text PDF available (49)

#### Publication Year

1993

2017

- This year (6)
- Last 5 years (48)
- Last 10 years (52)

#### Publication Type

#### Co-author

#### Journals and Conferences

#### Key Phrases

Learn More

- Leila Wehbe, Brian Murphy, Partha Talukdar, Alona Fyshe, Aaditya Ramdas, Tom Mitchell
- PloS one
- 2014

Story understanding involves many perceptual and cognitive subprocesses, from perceiving individual words, to parsing sentences, to understanding the relationships among the story characters. We present an integrated computational model of reading that incorporates these and additional subprocesses, simultaneously discovering their fMRI signatures. Our… (More)

We deal with two independent but related problems, those of graph similarity and subgraph matching, which are both important practical problems useful in several fields of science, engineering and data analysis. For the problem of graph similarity, we develop and test a new framework for solving the problem using belief propagation and related ideas. For… (More)

- Aaditya Ramdas, Ryan J. Tibshirani
- ArXiv
- 2014

This paper presents a fast and robust algorithm for trend filtering, a recently developed nonparametric regression tool. It has been shown that, for estimating functions whose derivatives are of bounded variation, trend filtering achieves the minimax optimal error rate, while other popular methods like smoothing splines and kernels do not. Standing in the… (More)

- Anna Ma, Deanna Needell, Aaditya Ramdas
- SIAM J. Matrix Analysis Applications
- 2015

The Kaczmarz and Gauss-Seidel methods both solve a linear system Xβ = y by iteratively refining the solution estimate. Recent interest in these methods has been sparked by a proof of Strohmer and Vershynin which shows the randomized Kaczmarz method converges linearly in expectation to the solution. Lewis and Leventhal then proved a similar result for the… (More)

- Akshay Balsubramani, Aaditya Ramdas
- UAI
- 2016

We propose a new algorithmic framework for sequential hypothesis testing with i.i.d. data, which includes A/B testing, nonparametric two-sample testing, and independence testing as special cases. It is novel in several ways: (a) it takes linear time and constant space to compute on the fly, (b) it has the same power guarantee (up to a small factor) as a… (More)

- Aaditya Ramdas, Nicolás García Trillos, Marco Cuturi
- Entropy
- 2017

Nonparametric two-sample or homogeneity testing is a decision theoretic problem that involves identifying differences between two random variables without making parametric assumptions about their underlying distributions. The literature is old and rich, with a wide variety of statistics having being designed and analyzed, both for the unidimensional and… (More)

Given a weighted graph with N vertices, consider a real-valued regression problem in a semisupervised setting, where one observes n labeled vertices, and the task is to label the remaining ones. We present a theoretical study of `p-based Laplacian regularization under a d-dimensional geometric random graph model. We provide a variational characterization of… (More)

We propose a class of nonparametric two-sample tests with a cost linear in the sample size. Two tests are given, both based on an ensemble of distances between analytic functions representing each of the distributions. The first test uses smoothed empirical characteristic functions to represent the distributions, the second uses distribution embeddings in a… (More)

This paper is about two related decision theoretic problems, nonparametric two-sample testing and independence testing. There is a belief that two recently proposed solutions, based on kernels and distances between pairs of points, behave well in high-dimensional settings. We identify different sources of misconception that give rise to the above belief.… (More)

- Dougal J. Sutherland, Hsiao-Yu Fish Tung, +4 authors Arthur Gretton
- ArXiv
- 2016

We propose a method to optimize the representation and distinguishability of samples from two probability distributions, by maximizing the estimated power of a statistical test based on the maximum mean discrepancy (MMD). This optimized MMD is applied to the setting of unsupervised learning by generative adversarial networks (GAN), in which a model attempts… (More)