Inferring User Actions from Provenance Logs

  title={Inferring User Actions from Provenance Logs},
  author={Xin Li and Chaitanya Joshi and Yu Shyang Tan and Ryan Kok Leong Ko},
  journal={2015 IEEE Trustcom/BigDataSE/ISPA},
Progger, a kernel-spaced cloud data provenance logger which provides fine-grained data activity records, was recently developed to empower cloud stakeholders to trace data life cycles within and across clouds. Progger logs have the potential to allow analysts to infer user actions and create a data-centric behaviour history in a cloud computing environment. However, the Progger logs are complex and noisy and therefore, currently this potential can not be met. This paper proposes a statistical… 

A survey of simulation provenance systems: modeling, capturing, querying, visualization, and advanced utilization

A comprehensive survey of a wide range of existing systems to utilize provenance data produced by simulation and presents a taxonomy of scientific platforms regarding provenance support to holistically tabulate the major functionalities and supporting levels of the studied systems.

Data Leakage Detection Using System Call Provenance

The design and implementation of Peeper, a policy based system for data leakage detection that utilizes operating system call provenance, shows that it enables real-time data detection of data leakage.

QProv: A provenance system for quantum computing

QProv, a provenance system that automatically collects the identified provenance attributes and provides them in a uniform manner to the user is introduced and a case study with the collected provenance data is presented here.



Progger: An Efficient, Tamper-Evident Kernel-Space Logger for Cloud Data Provenance Tracking

  • R. KoM. Will
  • Computer Science
    2014 IEEE 7th International Conference on Cloud Computing
  • 2014
Progger (Provenance Logger), a kernel-space logger which potentially empowers all cloud stakeholders to trace their data, is presented, which provides high assurance of data security and data activity audit.

Characterizing users’ visual analytic activity for insight provenance

The concept of actions has been integrated into the labpsilas prototype visual analytic system, HARVEST, as the basis for its insight provenance capabilities and a taxonomy to categorize actions into three major classes based on their semantic intent is defined.

S2Logger: End-to-End Data Tracking Mechanism for Cloud Data Provenance

S2Logger is introduced, a data event logging mechanism which captures, analyses and visualizes data events in the cloud from the data point of view, and can detect critical data-related cloud security problems such as malicious actions, data leakages and data policy violations by analysing the data provenance.

A Framework for Fine-grained Data Integration and Curation, with Provenance, in a Dataspace

This paper defines a data curation setting characterized by data that are explicitly selected, copied, and then pasted into a target dataset where they can be confirmed or replaced.

Temporal representation for scientific data provenance

This work proposes a representation of the provenance data based on logical time that reduces the feature space and applies clustering, classification and association rule mining to the abstract representations to determine the usefulness of the temporal representation.

BURRITO: Wrapping Your Lab Notebook in Computational Infrastructure

Researchers in fields such as bioinformatics, CS, finance, and applied math have trouble managing the numerous code and data files generated by their computational experiments, comparing the results

Flogger: A File-Centric Logger for Monitoring File Access and Transfers within Cloud Computing Environments

Fogger records file-centric access and transfer information from within the kernel spaces of both virtual machines (VMs) and physical machines (PMs) in the Cloud, thus giving full transparency of the entire data landscape in the cloud.

The WEKA data mining software: an update

This paper provides an introduction to the WEKA workbench, reviews the history of the project, and, in light of the recent 3.6 stable release, briefly discusses what has been added since the last stable version (Weka 3.4) released in 2003.

Applying data mining techniques for descriptive phrase extraction in digital document collections

This paper shows that general data mining methods are applicable to text analysis tasks such as descriptive phrase extraction and presents a general framework for text mining, based on generalized episodes and episode rules.

Prediction of User's Web-Browsing Behavior: Application of Markov Model

  • M. AwadIssa M. Khalil
  • Computer Science
    IEEE Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics)
  • 2012
A new modified Markov model is proposed to alleviate the issue of scalability in the number of paths and a new two-tier prediction framework is presented that creates an example classifier EC, based on the training examples and the generated classifiers.