Datasets:
audio
audioduration (s) 1
60
| label
class label 2
classes |
|---|---|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
|
0testing
|
Dataset Card for Dataset Name
This dataset is a collection of audioclips used for training machinelearning models on instrument based sound classification.
Dataset Details
Uses
This dataset can be used for training sound classification models on basic instruments
Direct Use
The Direct use of this dataset is to train a model on Sound Classification data based on instruments, using it to help those with Hearing Impairment identifying instruments.
Out-of-Scope Use
This dataset will not be usefull for anything other than sound based models
Dataset Structure
Clap 89% / 11% (5m 10s / 38s) Guitar 75% / 25% (42s / 14s) Kling 80% / 20% (2m 0s / 30s) Maracas 75% / 25% (3m 31s / 1m 11s) Noise 90% / 10% (8m 54s / 1m 0s) Snap 83% / 17% (4m 21s / 52s) Whistle 85% / 15% (1m 51s / 19s)
Dataset Creation
Curation Rationale
This dataset it aimed at training models to help people with impaired hearing identify instruments based on sound in a public space.
Source Data
.Wav audio files
Data Collection and Processing
Data was collected with a mobile microphone to best simulate the usecase for users. Data was cleaned as to minimally include high volume background noise.
Who are the source data producers?
The data was collected by students at Medialogy Masters at AAU copenhagen
Annotation process
The process included data cleaning, removing data that was unclear or duplicated.
Who are the annotators?
Student at Medialogy Masters AAU Copenhagen.
Personal and Sensitive Information
This dataset does not include private or sensitive data.
Bias, Risks, and Limitations
Cannot be used for anything non-soundbased, Additional data can be recomended for higher scale models.
Recommendations
Additional data can be recomended for higher scale models.
BibTeX: PROBLEMS ENCOUNTERED!! TRY APA!! @misc{huggingfaceAthorl22MLME_Instrument_ClassificationMain, author = {Athorl22}, title = {{A}thorl22/{M}{L}{M}{E}_{I}nstrument_{C}lassification at main --- huggingface.co}, howpublished = {\url{https://huggingface.co/datasets/Athorl22/MLME_Instrument_Classification/tree/main}}, year = {2025}, note = {[Accessed 04-12-2025]}, }
APA: WORKS!!! Athorl22/MLME_Instrument_Classification at main. (n.d.). https://huggingface.co/datasets/Athorl22/MLME_Instrument_Classification/tree/main
Dataset Card Contact
Github: AlfredVThor
- Downloads last month
- 28