Sound Sharing and Retrieval

@inproceedings{Font2018SoundSA,
  title={Sound Sharing and Retrieval},
  author={Frederic Font and Gerard Roma and Xavier Serra},
  year={2018}
}
Multimedia sharing has experienced an enormous growth in recent years, and sound sharing has not been an exception. Nowadays one can find online sound sharing sites in which users can search, browse, and contribute large amounts of audio content such as sound effects, field and urban recordings, music tracks, and music samples. This poses many challenges to enable search, discovery, and ultimately reuse of this content. In this chapter we give an overview of different ways to approach such… 
Search Result Clustering in Collaborative Sound Collections
TLDR
This work proposes a graph-based approach using audio features for clustering diverse sound collections obtained when querying large online databases, and shows that using a confidence measure for discarding inconsistent clusters improves the quality of the partitions.
Live Repurposing of Sounds: MIR Explorations with Personal and Crowdsourced Databases
TLDR
A live coding system built on SuperCollider enabling the use of audio content from online Creative Commons (CC) sound databases such as Freesound or personal sound databases and its potential in facilitating tailorability of the tool to their own creative workflows is demonstrated.
Music Information Retrieval in Live Coding: A Theoretical Framework
TLDR
It is found that it is still a technical challenge to use high-level features in real time, yet using rhythmic and tonal properties (midlevel features) in combination with text-based information (e.g., tags) helps to achieve a closer perceptual level centered on pitch and rhythm when using MIR in live coding.
Leveraging Online Audio Commons Content for Media Production
With the advent of online audio resources and web technologies, digital tools for sound designers and music producers are changing. The Internet provides access to hundreds of thousands of digital
Towards Duration Robust Weakly Supervised Sound Event Detection
TLDR
This work proposes a duration robust CRNN (CDur) framework, which aims to achieve competitive performance in terms of segment- and event-level localization, and investigates two data augmentation methods along with a label smoothing method within the scope of WSSED.
THE SMALLRICE SUBMISSION TO THE DCASE2021 TASK 4 CHALLENGE: A LIGHTWEIGHT APPROACH FOR SEMI-SUPERVISED SOUND EVENT DETECTION WITH UNSUPERVISED DATA AUGMENTATION Technical Report
TLDR
This work focuses on training a lightweight and well-performing model which can be used in real-world applications, and which can achieve competitive performance to the more common mean teacher paradigm.
MULTI-SCALE CONVOLUTION BASED ATTENTION NETWORK FOR SEMI-SUPERVISED SOUND EVENT DETECTION Technical Report
TLDR
This work proposed a Multi-Scale Convolution based Attention Network (MSCA), which can naturally learn coarse-to-fine multi-scale features to helps the model recognize different sound events and a channel-wise attention module is designed, which can adaptively recalibrates channels-wise feature responses by explicitly modelling interdependencies between channels.
Duration Robust Weakly Supervised Sound Event Detection
  • Heinrich Dinkel, Kai Yu
  • Computer Science
    ICASSP 2020 - 2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)
  • 2020
TLDR
It is shown that for this task subsampling the temporal resolution by a neural network enhances the F1 score as well as its robustness towards short, sporadic sound events and the use of double thresholding as a more robust and predictable post-processing method.
Capsule Routing for Sound Event Detection
TLDR
This work proposes a neural network architecture that uses the recently-proposed capsule routing mechanism to train a network that can learn global coherence implicitly, thereby improving generalization performance.
De la creación de un archivo a la práctica artística en el proyecto Intimal Lecciones aprendidas de la escucha de un archivo de historia oral de mujeres migrantes colombianas
This paper describes a multidisciplinary encounter with oral testimony archives and their incorporation in the artistic research project INTIMAL. It explores ways in which to creatively listen to
...
...

References

SHOWING 1-10 OF 76 REFERENCES
ENHANCING SONIC BROWSING USING AUDIO INFORMATION RETRIEVAL
TLDR
This paper describes how audio information retrieval can be utilized to create novel user interfaces for browsing of audio collections and reports on recent work on two system prototypes: the Sonic Browser and Marsyas and current work on merging the two systems in a common flexible system.
Can all tags be used for search?
TLDR
This paper is the first to present an in-depth study of tagging behavior for very different kinds of resources and systems - Web pages, music, and images - and compares the results with anchor text characteristics, and provides statistics on tag distributions in all three tagging environments.
An overview of audio information retrieval
  • J. Foote
  • Computer Science
    Multimedia Systems
  • 1999
TLDR
The state of the art in audio information retrieval is reviewed, and recent advances in automatic speech recognition, word spotting, speaker and music identification, and audio similarity are presented with a view towards making audio less “opaque”.
Tag Recommendation using Folksonomy Information for Online Sound Sharing Platforms
TLDR
A new perspective for tag recommendation is explored which, besides taking advantage of information from the folksonomy, employs a sound-specific ontology to guide users during the annotation process, and can be easily generalised and thus be of use to other information sharing platforms.
SOUND SEARCH BY CONTENT-BASED NAVIGATION IN LARGE DATABASES
We propose to apply the principle of interactive real-time corpus-based concatenative synthesis to search in effects or instrument sound databases, which becomes content-based navigation in a space
MARSYAS: a framework for audio analysis
TLDR
This paper describes MARSYAS, a framework for experimenting, evaluating and integrating techniques for audio content analysis in restricted domains and a new method for temporal segmentation based on audio texture that is combined with audio analysis techniques and used for hierarchical browsing, classification and annotation of audio files.
SoundTorch: Quick Browsing in Large Audio Collections
TLDR
User tests show that this method can leverage the human brain’s capability to single out sounds from a spatial mixture and enhance browsing in large collections of audio content.
A Survey of Automatic Query Expansion in Information Retrieval
TLDR
This survey presents a unified view of a large number of recent approaches to AQE that leverage various data sources and employ very different principles and techniques.
Acoustic lexemes for organizing internet audio
In this article, a method is proposed for automatic fine-scale audio description that draws inspiration from ontological sound description methods such as Shaeffer's Objets Sonores and Smalley's
Querying Freesound with a microphone
TLDR
This work presents a demonstration system that allows users to make queries on Freesound.org by recording audio in the browser, enabling Query by Example (QbE) applications.
...
...