Privacy-aware environmental sound classification for indoor human activity recognition

  title={Privacy-aware environmental sound classification for indoor human activity recognition},
  author={Wei Wang and Fatjon Seraj and Nirvana Meratnia and Paul J. M. Havinga},
  journal={Proceedings of the 12th ACM International Conference on PErvasive Technologies Related to Assistive Environments},
  • Wei Wang, F. Seraj, P. Havinga
  • Published 5 June 2019
  • Computer Science
  • Proceedings of the 12th ACM International Conference on PErvasive Technologies Related to Assistive Environments
This paper presents a comparative study on different feature extraction and machine learning techniques for indoor environmental sound classification. Compared to outdoor environmental sound classification systems, indoor systems need to pay special attention to power consumption and privacy. We consider feature calculation complexity, classification accuracy and privacy as evaluation metrics. To ensure privacy, we strip voice bands from sound input to make human conversations unrecognizable… 
A Comparative Survey of Feature Extraction and Machine Learning Methods in Diverse Acoustic Environments
A comparative study on combining several feature extraction algorithms with a group of machine learning algorithms with the goal of detailing a best practice method and evaluating the reliability of this general-purpose algorithm for all the classes.
Sound Event Recognition Based in Feature Combination with Low SNR
A new method of sound event recognition based in feature combination with low SNR is proposed which improves the recognition of sound events in real scenes by adding noise to public sound data sets and drawing on the advantages of image recognition.
Audiovisual Classification of Group Emotion Valence Using Activity Recognition Networks
The results show that using activity recognition pretraining offers performance advantages for group-emotion recognition and that audio is essential to improve the accuracy and robustness of video-based recognition.
Outdoor Human Footsteps Event and Environment Joint Recognition
  • Fengkai Xu, P. Li
  • Computer Science
    2020 Chinese Control And Decision Conference (CCDC)
  • 2020
The cross double footsteps segmentation algorithm and fusion features designed in this paper simplify and balance the tasks, so that the multitask design of outdoor human footsteps event and environment joint recognition can achieve accurate recognition without complex models.
Ok Google, What Am I Doing?
This work explores how off-the-shelf conversational assistants can be enhanced with acoustic-based human activity recognition by leveraging the short interval after a voice command is given to the device.


Sound Event Recognition With Probabilistic Distance SVMs
  • H. D. Tran, Haizhou Li
  • Computer Science
    IEEE Transactions on Audio, Speech, and Language Processing
  • 2011
The results show that the proposed classification method significantly outperforms conventional SVM classifiers with Mel-frequency cepstral coefficients (MFCCs) and makes the proposed method an obvious choice for online sound event recognition.
TUT database for acoustic scene classification and sound event detection
The recording and annotation procedure, the database content, a recommended cross-validation setup and performance of supervised acoustic scene classification system and event detection baseline system using mel frequency cepstral coefficients and Gaussian mixture models are presented.
Audio-based context recognition
This paper investigates the feasibility of an audio-based context recognition system developed and compared to the accuracy of human listeners in the same task, with particular emphasis on the computational complexity of the methods.
Sound Event Detection in Multichannel Audio Using Spatial and Harmonic Features
The proposed SED system is compared against the state of the art mono channel method on the development subset of TUT sound events detection 2016 database and the usage of spatial and harmonic features are shown to improve the performance of SED.
Environmental sound classification with convolutional neural networks
  • Karol J. Piczak
  • Computer Science
    2015 IEEE 25th International Workshop on Machine Learning for Signal Processing (MLSP)
  • 2015
The model outperforms baseline implementations relying on mel-frequency cepstral coefficients and achieves results comparable to other state-of-the-art approaches.
A Dataset and Taxonomy for Urban Sound Research
A taxonomy of urban sounds and a new dataset, UrbanSound, containing 27 hours of audio with 18.5 hours of annotated sound event occurrences across 10 sound classes are presented.
Detection, classification and localization of acoustic events in the presence of background noise for acoustic surveillance of hazardous situations
It is found that the engineered algorithms provide a sufficient robustness in moderately intense noise in order to be applied to practical audio-visual surveillance systems.
Comparison of techniques for environmental sound recognition
Sensor-based occupancy behavioral pattern recognition for energy and comfort management in intelligent buildings
The objective is to develop and implement algorithms for sensor-based modelling and prediction of user behaviour in intelligent buildings and connect the behavioural patterns to building energy and comfort management systems through simulation tools.
Human Detection Based on the Generation of a Background Image by Using a Far-Infrared Light Camera
A new method for detecting human areas in thermal camera images is proposed and Experimental results with two types of databases confirm that the proposed method outperforms other methods.