|
--- |
|
license: mit |
|
datasets: |
|
- ccmusic-database/pianos |
|
language: |
|
- en |
|
tags: |
|
- music |
|
- art |
|
metrics: |
|
- accuracy |
|
pipeline_tag: audio-classification |
|
library_name: https://github.com/monetjoe/Piano-Classification |
|
--- |
|
# Intro |
|
This study, based on deep learning technology, draws inspiration from classical backbone network structures in the computer vision domain to construct an innovative 8-class piano timbre discriminator model through audio data processing. The model focuses on eight brands and types of pianos, including Kawai, Kawai Grand, YOUNG CHANG, HSINGHAI, Steinway Theatre, Steinway Grand, Pearl River, and Yamaha. By transforming audio data into Mel spectrograms and conducting supervised learning in the fine-tuning phase, the model accurately distinguishes different piano timbres and performs well in practical testing. In the training process, a large-scale annotated audio dataset is utilized, and the introduction of deep learning technology provides crucial support for improving the model's performance by progressively learning to extract key features from audio. The piano timbre discriminator model has broad potential applications in music assessment, audio engineering, and other fields, offering an advanced and reliable solution for piano timbre discrimination. This study expands new possibilities for the application of deep learning in the audio domain, providing valuable references for future research and applications in related fields. |
|
|
|
## Demo |
|
<https://huggingface.co/spaces/ccmusic-database/pianos> |
|
|
|
## Usage |
|
```python |
|
from modelscope import snapshot_download |
|
model_dir = snapshot_download('ccmusic-database/pianos') |
|
``` |
|
|
|
## Maintenance |
|
```bash |
|
GIT_LFS_SKIP_SMUDGE=1 git clone [email protected]:ccmusic-database/pianos |
|
cd pianos |
|
``` |
|
|
|
## Results |
|
A demo result of SqueezeNet fine-tuning: |
|
<style> |
|
#pianos td { |
|
vertical-align: middle !important; |
|
text-align: center; |
|
} |
|
#pianos th { |
|
text-align: center; |
|
} |
|
</style> |
|
<table id="pianos"> |
|
<tr> |
|
<th>Loss curve</th> |
|
<td><img src="https://www.modelscope.cn/api/v1/models/ccmusic-database/pianos/repo?Revision=master&FilePath=loss.jpg"></td> |
|
</tr> |
|
<tr> |
|
<th>Training and validation accuracy</th> |
|
<td><img src="https://www.modelscope.cn/api/v1/models/ccmusic-database/pianos/repo?Revision=master&FilePath=acc.jpg"></td> |
|
</tr> |
|
<tr> |
|
<th>Confusion matrix</th> |
|
<td><img src="https://www.modelscope.cn/api/v1/models/ccmusic-database/pianos/repo?Revision=master&FilePath=mat.jpg"></td> |
|
</tr> |
|
</table> |
|
|
|
## Dataset |
|
<https://huggingface.co/datasets/ccmusic-database/pianos> |
|
|
|
## Mirror |
|
<https://www.modelscope.cn/models/ccmusic-database/pianos> |
|
|
|
## Evaluation |
|
<https://github.com/monetjoe/Piano-Classification> |
|
|
|
## Cite |
|
```bibtex |
|
@article{Zhou2023AHE, |
|
author = {Monan Zhou and Shangda Wu and Shaohua Ji and Zijin Li and Wei Li}, |
|
title = {A Holistic Evaluation of Piano Sound Quality}, |
|
booktitle = {Proceedings of the 10th Conference on Sound and Music Technology (CSMT)}, |
|
year = {2023}, |
|
publisher = {Springer Singapore}, |
|
address = {Singapore} |
|
} |
|
``` |