#### Filter Results:

- Full text PDF available (142)

#### Publication Year

1997

2017

- This year (14)
- Last 5 years (84)
- Last 10 years (133)

#### Publication Type

#### Co-author

#### Journals and Conferences

#### Data Set Used

#### Key Phrases

#### Method

#### Organism

Learn More

- Jure Leskovec, Kevin J. Lang, Anirban Dasgupta, Michael W. Mahoney
- Internet Mathematics
- 2009

A large body of work has been devoted to defining and identifying clusters or communities in social and information networks, i.e., in graphs in which the nodes represent underlying social entities and the edges represent some sort of interaction between pairs of nodes. Most such research begins with the premise that a community or a cluster should be… (More)

- Petros Drineas, Michael W. Mahoney
- Journal of Machine Learning Research
- 2005

A problem for many kernel-based methods is that the amount of computation required to find the solution scales as O(n3), where n is the number of training examples. We develop and analyze an algorithm to compute an easily-interpretable low-rank approximation to an n× n Gram matrix G such that computations of interest may be performed more rapidly. The… (More)

A large body of work has been devoted to identifying community structure in networks. A community is often though of as a set of nodes that has more connections between its members than to the remainder of the network. In this paper, we characterize as a function of size the statistical and structural properties of such sets of nodes. We define the network… (More)

- Jure Leskovec, Kevin J. Lang, Michael W. Mahoney
- WWW
- 2010

Detecting clusters or communities in large real-world graphs such as large social or information networks is a problem of considerable interest. In practice, one typically chooses an objective function that captures the intuition of a network cluster as set of nodes with better internal connectivity than external connectivity, and then one applies… (More)

- Petros Drineas, Ravi Kannan, Michael W. Mahoney
- SIAM J. Comput.
- 2006

In many applications, the data consist of (or may be naturally formulated as) an m×n matrix A. It is often of interest to find a low-rank approximation to A, i.e., an approximation D to the matrix A of rank not greater than a specified rank k, where k is much smaller than m and n. Methods such as the singular value decomposition (SVD) may be used to find an… (More)

- Petros Drineas, Ravi Kannan, Michael W. Mahoney
- SIAM J. Comput.
- 2006

In many applications, the data consist of (or may be naturally formulated as) an m× n matrix A which may be stored on disk but which is too large to be read into random access memory (RAM) or to practically perform superlinear polynomial time computations on it. Two algorithms are presented which, when given an m×n matrix A, compute approximations to A… (More)

- Petros Drineas, Michael W. Mahoney, S. Muthukrishnan
- SIAM J. Matrix Analysis Applications
- 2008

Many data analysis applications deal with large matrices and involve approximating the matrix using a small number of “components.” Typically, these components are linear combinations of the rows and columns of the matrix, and are thus difficult to interpret in terms of the original features of the input data. In this paper, we propose and study matrix… (More)

- Petros Drineas, Ravi Kannan, Michael W. Mahoney
- SIAM J. Comput.
- 2006

Motivated by applications in which the data may be formulated as a matrix, we consider algorithms for several common linear algebra problems. These algorithms make more efficient use of computational resources, such as the computation time, random access memory (RAM), and the number of passes over the data, than do previously known algorithms for these… (More)

We consider the problem of selecting the “best” subset of exactly k columns from an m× n matrix A. In particular, we present and analyze a novel two-stage algorithm that runs in O(min{mn2,m2n}) time and returns as output an m × k matrix C consisting of exactly k columns of A. In the first stage (the randomized stage), the algorithm randomly selects O(k log… (More)

- Alex Gittens, Michael W. Mahoney
- ICML
- 2013

We reconsider randomized algorithms for the low-rank approximation of symmetric positive semi-definite (SPSD) matrices such as Laplacian and kernel matrices that arise in data analysis and machine learning applications. Our main results consist of an empirical evaluation of the performance quality and running time of sampling and projection methods on a… (More)