2014 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2014
DOI: 10.1109/icassp.2014.6854173
|View full text |Cite
|
Sign up to set email alerts
|

Polyphonic piano transcription using non-negative Matrix Factorisation with group sparsity

Abstract: Non-negative Matrix Factorisation (NMF) is a popular tool in musical signal processing. However, problems using this methodology in the context of Automatic Music Transcription (AMT) have been noted resulting in the proposal of supervised and constrained variants of NMF for this purpose. Group sparsity has previously been seen to be effective for AMT when used with stepwise methods. In this paper group sparsity is introduced to supervised NMF decompositions and a dictionary tuning approach to AMT is proposed b… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
16
0

Year Published

2016
2016
2024
2024

Publication Types

Select...
5
2
1

Relationship

2
6

Authors

Journals

citations
Cited by 15 publications
(16 citation statements)
references
References 16 publications
0
16
0
Order By: Relevance
“…This is usually not true in practice, where the instrument models/sources at test time are unknown and usually do not coincide with the instruments used for training. A majority of experiments with the MAPS dataset train and test model on disjoint instrument types [3], [2], [44]. We thus perform a second set of experiments to compare performance of the different neural network acoustic models in a more realistic setting.…”
Section: A Datasetmentioning
confidence: 99%
“…This is usually not true in practice, where the instrument models/sources at test time are unknown and usually do not coincide with the instruments used for training. A majority of experiments with the MAPS dataset train and test model on disjoint instrument types [3], [2], [44]. We thus perform a second set of experiments to compare performance of the different neural network acoustic models in a more realistic setting.…”
Section: A Datasetmentioning
confidence: 99%
“…Improvements of ∼ 5% for β (0.5) and ∼ 4% for KL are observed using dictionary tuning. In the case of β (0.5) , all metrics increase by 1% relative to β-NMD using the optimal dictionary, and also by ∼ 2% relative to our previous results [17], using the monotonic descent algorithm.…”
Section: Experiments Cmentioning
confidence: 74%
“…We then employ this approach in a dictionary tuning method, applied to a restructured version of the harmonic dictionary used in [4], whereby the hard harmonic constraint is dropped. Part of this work was described in [17] and is augmented here through monotonic descent algorithm with scale invariant group sparse penalty and further evaluation. We also propose a new onset detector for NMF-based AMT.…”
Section: A Contributions Of This Papermentioning
confidence: 99%
See 1 more Smart Citation
“…The proposed approach gives improved performance for the AMT task and is simple to implement, which makes it easy to add to existing NMF-based systems. For future work, similar ideas could be applied to more complicated NMF AMT algorithms involving group sparsity [22] or non-negative dictionary learning. So far, we have used a setting of 0.5 for β which has previously found to work well for AMT, however it is possible that the performance could be further improved by fine-tuning the β parameter for the low-rank approach.…”
Section: A Results and Discussionmentioning
confidence: 99%