论文标题
使用数据增强评估深度音乐生成方法
Evaluating Deep Music Generation Methods Using Data Augmentation
论文作者
论文摘要
尽管深度算法的音乐产生进展,但对生成样本的评估通常依赖于人类评估,这是主观且昂贵的。我们专注于设计一个均匀的客观框架,用于评估算法生成音乐的样本。评估生成音乐的任何工程措施通常都试图定义样品的音乐性,但不会捕捉音乐的素质,例如主题或情绪。我们不寻求评估产生音乐的音乐功能,而是探索生成的样本是否包含与情感或情绪/主题有关的有意义的信息。我们通过在使用生成的样本增强其培训数据后测量音乐情绪/主题分类器的预测性能的变化来实现这一目标。我们分析了由Samplernn,Jukebox和DDSP产生的三种模型生成的音乐样本,并在所有方法中采用均匀的框架以进行客观比较。这是通过有条件生成的音乐增强音乐流派分类数据集的首次尝试。我们使用深度音乐的生成以及发电机的能力来研究分类性能的改善,并使用数据集的其他情感注释来制作情感音乐。最后,我们使用经过实际数据训练的分类器来评估班级条件生成的样本的标签有效性。
Despite advances in deep algorithmic music generation, evaluation of generated samples often relies on human evaluation, which is subjective and costly. We focus on designing a homogeneous, objective framework for evaluating samples of algorithmically generated music. Any engineered measures to evaluate generated music typically attempt to define the samples' musicality, but do not capture qualities of music such as theme or mood. We do not seek to assess the musical merit of generated music, but instead explore whether generated samples contain meaningful information pertaining to emotion or mood/theme. We achieve this by measuring the change in predictive performance of a music mood/theme classifier after augmenting its training data with generated samples. We analyse music samples generated by three models -- SampleRNN, Jukebox, and DDSP -- and employ a homogeneous framework across all methods to allow for objective comparison. This is the first attempt at augmenting a music genre classification dataset with conditionally generated music. We investigate the classification performance improvement using deep music generation and the ability of the generators to make emotional music by using an additional, emotion annotation of the dataset. Finally, we use a classifier trained on real data to evaluate the label validity of class-conditionally generated samples.