Towards a Practical Differentially Private Collaborative Phone Blacklisting System

  title={Towards a Practical Differentially Private Collaborative Phone Blacklisting System},
  author={Daniele Ucci and Roberto Perdisci and Jaewoo Lee and Mustaque Ahamad},
  journal={Annual Computer Security Applications Conference},
Spam phone calls have been rapidly growing from nuisance to an increasingly effective scam delivery tool. To counter this increasingly successful attack vector, a number of commercial smartphone apps that promise to block spam phone calls have appeared on app stores, and are now used by hundreds of thousands or even millions of users. However, following a business model similar to some online social network services, these apps often collect call records or other potentially sensitive… Expand

Figures from this paper


Augmenting Telephone Spam Blacklists by Mining Large CDR Datasets
A novel detection system that aims to discover telephone numbers involved in spam campaigns using a combination of unsupervised and supervised machine learning methods to mine new, previously unknown spam numbers from large datasets of call detail records (CDRs). Expand
A Machine Learning Approach to Prevent Malicious Calls over Telephony Networks
This work presents the first machine learning-based solution without relying on any particular assumptions on the underlying telephony network infrastructures to predict malicious calls, and extensively evaluates different state-of-the-art machine learning approaches using the proposed features. Expand
Locally Differentially Private Heavy Hitter Identification
In this paper, a proposed LDP protocol, which the authors call Prefix Extending Method (PEM), users are divided into groups, with each group reporting a prefix of her value and experiments show that under the same privacy guarantee and computational cost, PEM has better utility on both synthetic and real-world datasets than existing solutions. Expand
Building a RAPPOR with the Unknown: Privacy-Preserving Learning of Associations and Data Dictionaries
This paper proposes a novel decoding algorithm for the RAPPOR mechanism that enables the estimation of “unknown unknowns,” i.e., strings the authors do not know they should be estimating, to enable learning without explicit dictionary knowledge. Expand
Locally Differentially Private Frequent Itemset Mining
This paper formally defines padding and sample based frequency oracles (PSFO) and identifies the privacy amplification property in PSFO, and proposes SVIM, a protocol for finding frequent items in the set-valued LDP setting, which significantly improves over existing methods. Expand
Heavy Hitter Estimation over Set-Valued Data with Local Differential Privacy
The main idea is to first gather a candidate set of heavy hitters using a portion of the privacy budget, and focus the remaining budget on refining the candidate set in a second phase, which is much more efficient budget-wise than obtaining the heavy hitters directly from the whole dataset. Expand
RAPPOR: Randomized Aggregatable Privacy-Preserving Ordinal Response
This paper describes and motivates RAPPOR, details its differential-privacy and utility guarantees, discusses its practical deployment and properties in the face of different attack models, and gives results of its application to both synthetic and real-world data. Expand
Locally Differentially Private Protocols for Frequency Estimation
This paper introduces a framework that generalizes several LDP protocols proposed in the literature and yields a simple and fast aggregation algorithm, whose accuracy can be precisely analyzed, resulting in two new protocols that provide better utility than protocols previously proposed. Expand
Differential privacy under continual observation
This work identifies the problem of maintaining a counter in a privacy preserving manner and shows its wide applicability to many different problems. Expand
"Better Than Nothing" Privacy with Bloom Filters: To What Extent?
This work derives exact and (tightly) approximate formulae which permit to readily relate privacy properties with filter (and universe set) parameters, and quantitatively investigate the emerging privacy/utility trade-offs. Expand