• Corpus ID: 236493514

Pitch-Informed Instrument Assignment Using a Deep Convolutional Network with Multiple Kernel Shapes

@article{Lordelo2021PitchInformedIA,
  title={Pitch-Informed Instrument Assignment Using a Deep Convolutional Network with Multiple Kernel Shapes},
  author={Carlos Lordelo and Emmanouil Benetos and Simon Dixon and Sven Ahlback},
  journal={ArXiv},
  year={2021},
  volume={abs/2107.13617}
}
This paper proposes a deep convolutional neural network for performing note-level instrument assignment. Given a polyphonic multi-instrumental music signal along with its ground truth or predicted notes, the objective is to assign an instrumental source for each note. This problem is addressed as a pitch-informed classification task where each note is analysed individually. We also propose to utilise several kernel shapes in the convolutional layers in order to facilitate learning of efficient… 
1 Citations

Figures and Tables from this paper

Voice Assignment in Vocal Quartets Using Deep Learning Models Based on Pitch Salience

  • Helena CuestaE. Gómez
  • Computer Science
    Transactions of the International Society for Music Information Retrieval
  • 2022

References

SHOWING 1-10 OF 23 REFERENCES

Frame-level Instrument Recognition by Timbre and Pitch

This paper builds and evaluates a convolutional neural network for making frame-level instrument prediction, and considers it as a multi-label classification problem for each frame and uses frame- level annotations as the supervisory signal in training the network.

Deep Convolutional Neural Networks for Predominant Instrument Recognition in Polyphonic Music

The analysis on the instrument-wise performance found that the onset type is a critical factor for recall and precision of each instrument, and convolutional neural networks are more robust than conventional methods that exploit spectral features and source separation with support vector machines.

Multitask Learning for Frame-level Instrument Recognition

A large-scale dataset that contains synthetic polyphonic music with frame-level pitch and instrument labels is presented and a simple yet novel network architecture is proposed to jointly predict the Pitch and instrument for each frame and the effectiveness of the proposed method is validated.

Deep Salience Representations for F0 Estimation in Polyphonic Music

A fully convolutional neural network for learning salience representations for estimating fundamental frequencies, trained using a large, semi-automatically generated f0 dataset is described and shown to achieve state-of-the-art performance on several multi-f0 and melody datasets.

Music instrument recognition using deep convolutional neural networks

A deep convolution neural network framework for predominant instrument recognition in real-world polyphonic music is accomplished and the research excellent result with 92.8% accuracy.

An Attention Mechanism for Musical Instrument Recognition

The proposed attention model is compared to multiple models which include a baseline binary relevance random forest, recurrent neural network, and fully connected neural networks to show that incorporating attention leads to an overall improvement in classification accuracy metrics across all 20 instruments in the OpenMIC dataset.

Investigating Kernel Shapes and Skip Connections for Deep Learning-Based Harmonic-Percussive Separation

An efficient deep learning encoder-decoder network for performing Harmonic-Percussive Source Separation is proposed and it is shown that the number of model trainable parameters is greatly reduced by using a dense arrangement of skip connections between the model layers.

Multi-Instrument Automatic Music Transcription With Self-Attention-Based Instance Segmentation

This article proposes a multi-instrument AMT method, with signal processing techniques specifying pitch saliency, novel deep learning techniques, and concepts partly inspired by multi-object recognition, instance segmentation, and image-to-image translation in computer vision.

Multi-Instrument Music Transcription Based on Deep Spherical Clustering of Spectrograms and Pitchgrams

The proposed clustering-based music transcription method can transcribe musical pieces including unknown musical instruments as well as those containing only predefined instruments, at the state-of-the-art transcription accuracy.

Learning Features of Music from Scratch

A multi-label classification task to predict notes in musical recordings is defined, along with an evaluation protocol, and several machine learning architectures for this task are benchmarked.