Uniformization of Discrete Data

  title={Uniformization of Discrete Data},
  author={Lei Yang},
  • Lei Yang
  • Published in ISAAC 19 December 2005
  • Computer Science
Some kind of discrete data sets can be practically transformed into uniform by the related distribution function. By addressing the sparsity of data which measures the discreteness, this paper demonstrates that the sparsity decides the uniformity of the transformed data, and that could be a good reason to explain both the success of the bucket sort in PennySort 2003 and the failure for the same algorithm with the data modified. So the sparsity provides a good criterion to predict whether the… 
1 Citations
Employment of neural network and rough set in meta-learning
A quantitative meta-learning approach based on neural network and rough set theory in the selection of the best predictive model, which depends directly on the characteristic, meta-features of the input data sets.


Performance / Price Sort and PennySort
This paper documents this and proposes that the PennySort benchmark be revised to Performance/Price sort: a simple GB/$ sort metric based on a two-pass external sort.
THSORT: A Single-Processor Parallel Sorting Algorithm
Experimental results based on a computer with two RAIDs indicate that THSORT (Tsinghua SORT), a parallel sorting algorithm on a single computer, has almost doubled the performance of NTSORT, a famous sorting program.
The Sample-Seperators Based Distributing Scheme of the External Bucket Sort Algorithm
This scheme enables the bucket sort algorithm to be applied in the SheenkSort system to win the 2003 PennySort competition, and the probability to avoid memory overflow is calculated.