Chandler May

Learn More
OBJECTIVE To review the evidence that supports the use of certain cephalosporins in penicillin-allergic patients. DATA SOURCES Published articles were identified through Medline and EMBASE (1960-2007) using the search terms penicillin and allergy and cephalosporin and cross-reactivity. Additional sources were identified from the authors' personal(More)
The challenge of modelling cancer presents a major opportunity to improve our ability to reduce mortality from malignant neoplasms, improve treatments and meet the demands associated with the individualization of care needs. This is the central motivation behind the ContraCancrum project. By developing integrated multi-scale cancer models, ContraCancrum is(More)
We compare the multinomial i-vector framework from the speech community with LDA, SAGE, and LSA as feature learners for topic ID on multinomial speech and text data. We also compare the learned representations in their ability to discover topics, quantified by dis-tributional similarity to gold-standard topics and by human interpretability. We find that(More)
Previous research has established several methods of online learning for latent Dirichlet allocation (LDA). However , streaming learning for LDA— allowing only one pass over the data and constant storage complexity—is not as well explored. We use reservoir sampling to reduce the storage complexity of a previously-studied online algorithm, namely the(More)
The input data to a topic model is a collection of documents, each of which is represented as a vector of word counts. Even in a noisy environment, relatively little work is required to generate such count vectors from a raw corpus. On the other hand, generating count vectors (or soft–count vectors) over content-bearing features from raw audio data requires(More)
We develop a streaming (one-pass, bounded-memory) word embedding algorithm based on the canonical skip-gram with negative sampling algorithm implemented in word2vec. We compare our streaming algorithm to word2vec empirically by measuring the cosine similarity between word pairs under each algorithm and by applying each algorithm in the downstream task of(More)
  • 1