TOD: GPU-accelerated Outlier Detection via Tensor Operations
@article{Zhao2021TODGO, title={TOD: GPU-accelerated Outlier Detection via Tensor Operations}, author={Yue Zhao and George H. Chen and Zhihao Jia}, journal={Proc. VLDB Endow.}, year={2021}, volume={16}, pages={546-560} }
Outlier detection (OD) is a key machine learning task for finding rare and deviant data samples, with many time-critical applications such as fraud detection and intrusion detection. In this work, we propose TOD , the first tensor-based system for efficient and scalable outlier detection on distributed multi-GPU machines. A key idea behind TOD is decomposing complex OD applications into a small collection of basic tensor algebra operators. This decomposition enables TOD to accelerate OD…
Figures and Tables from this paper
One Citation
ADMoE: Anomaly Detection with Mixture-of-Experts from Noisy Labels
- Computer ScienceArXiv
- 2022
ADMoE is model-agnostic to enable any neural network-based detection methods to handle noisy labels, where it showcases its results on both multiple-layer perceptron (MLP) and leading AD method DeepSAD.
References
SHOWING 1-10 OF 104 REFERENCES
SUOD: Accelerating Large-Scale Unsupervised Heterogeneous Outlier Detection
- Computer ScienceMLSys
- 2021
A three-module acceleration system called SUOD (scalable unsupervised outlier detection) that focuses on three complementary aspects to accelerate (dimensionality reduction for high-dimensional data, model approximation for complex models, and execution efficiency improvement for taskload imbalance within distributed systems), while controlling detection performance degradation.
GPU Strategies for Distance-Based Outlier Detection
- Computer ScienceIEEE Transactions on Parallel and Distributed Systems
- 2016
A family of parallel and distributed algorithms for graphic processing units (GPU) derived from two distance-based outlier detection algorithms: BruteForce and SolvingSet are proposed, which differ in the way they exploit the architecture and memory hierarchy of the GPU and guarantee significant improvements with respect to the CPU versions.
Sparx: Distributed Outlier Detection at Scale
- Computer ScienceKDD
- 2022
Sparx is designed, a data-parallel OD algorithm suitable for shared-nothing infrastructures, which it is shown that existing open-source solutions fail to scale up; either by large number of points or high dimensionality, whereas Sparx yields scalable and effective performance.
Accelerating the local outlier factor algorithm on a GPU for intrusion detection systems
- Computer ScienceGPGPU-3
- 2010
This paper explores how to utilize a CUDA-based GPU implementation of the k-nearest neighbor algorithm to accelerate LOF classification and achieves more than a 100X speedup over a multi-threaded dual-core CPU implementation.
GPU-Accelerated Feature Selection for Outlier Detection Using the Local Kernel Density Ratio
- Computer Science2012 IEEE 12th International Conference on Data Mining
- 2012
This work presents a novel non-parametric evaluation criterion for filter-based feature selection which caters to outlier detection problems, and implements the algorithm on a graphics processing unit (GPU) to gain significant speedup over the serial version.
A Tensor Compiler for Unified Machine Learning Prediction Serving
- Computer ScienceOSDI
- 2020
HUMMINGBIRD is a novel approach to model scoring, which compiles featurization operators and traditional ML models into a small set of tensor operations, which inherently reduces infrastructure complexity and directly leverages existing investments in Neural Network compilers and runtimes to generate efficient computations for both CPU and hardware accelerators.
Real-Time Distance-Based Outlier Detection in Data Streams
- Computer ScienceProc. VLDB Endow.
- 2020
This study proposes a new algorithm, CPOD, to improve the efficiency of outlier detections while reducing its memory requirements, and shows that with six real-world and one synthetic dataset,CPOD is, on average, 10, 19, and 73 times faster than M_MCOD, NETS, and MCOD, respectively, while consuming low memory.
Parallel processing for distance-based outlier detection on a multi-core CPU
- Computer Science2014 IEEE 7th International Workshop on Computational Intelligence and Applications (IWCIA)
- 2014
A new parallelization model for the parallel processing of Orca-based outlier detection on a multi-core CPU that utilizes data parallelism and a Multi-thread model and outperforms conventional parallelization models.
Distributed Local Outlier Detection in Big Data
- Computer ScienceKDD
- 2017
This work presents the first distributed solution for the Local Outlier Factor (LOF) method, and proposes a data assignment strategy which ensures that each machine is self-sufficient in all stages of the LOF pipeline, while minimizing the number of data replicas.
Estimating GPU memory consumption of deep learning models
- Computer ScienceESEC/SIGSOFT FSE
- 2020
DNNMem employs an analytic estimation approach to systematically calculate the memory consumption of both the computation graph and the DL framework runtime, and shows that DNNMem is effective in estimating GPU memory consumption.