Intonation: A Dataset of Quality Vocal Performances Refined by Spectral Clustering on Pitch Congruence
Author(s): Wager, Sanna; Tzanetakis, George; Sullivan, Stefan; Wang, Cheng-i; Shimmin, John; et al
DownloadTo refer to this page use:
http://arks.princeton.edu/ark:/88435/pr1f83g
Abstract: | We introduce the "Intonation" dataset of amateur vocal performances with a tendency for good intonation, collected from Smule, Inc. The dataset can be used for music information retrieval tasks such as autotuning, query by humming, and singing style analysis. It is available upon request on the Stanford CCRMA DAMP website. 1 We describe a semi-supervised approach to selecting the audio recordings from a larger collection of performances based on intonation patterns. The approach can be applied in other situations where a researcher needs to extract a subset of data samples from a large database. A comparison of the "Intonation" dataset and the remaining collection of performances shows that the two have different intonation behavior distributions. |
Publication Date: | 2019 |
Citation: | Wager, Sanna, George Tzanetakis, Stefan Sullivan, Cheng-I. Wang, John Shimmin, Minje Kim, and Perry Cook. "Intonation: A Dataset of Quality Vocal Performances Refined by Spectral Clustering on Pitch Congruence." IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) (2019): pp. 476-480. doi: 10.1109/ICASSP.2019.8683554 |
DOI: | 10.1109/ICASSP.2019.8683554 |
ISSN: | 1520-6149 |
EISSN: | 2379-190X |
Pages: | 476 - 480 |
Type of Material: | Conference Article |
Journal/Proceeding Title: | IEEE International Conference on Acoustics, Speech and Signal Processing |
Version: | Author's manuscript |
Items in OAR@Princeton are protected by copyright, with all rights reserved, unless otherwise indicated.