Listening to features

M Moussallam, A Liutkus, L Daudet - arXiv preprint arXiv:1501.04981, 2015 - arxiv.org
arXiv preprint arXiv:1501.04981, 2015arxiv.org
This work explores nonparametric methods which aim at synthesizing audio from low-
dimensionnal acoustic features typically used in MIR frameworks. Several issues prevent
this task to be straightforwardly achieved. Such features are designed for analysis and not
for synthesis, thus favoring high-level description over easily inverted acoustic
representation. Whereas some previous studies already considered the problem of
synthesizing audio from features such as Mel-Frequency Cepstral Coefficients, they mainly …
This work explores nonparametric methods which aim at synthesizing audio from low-dimensionnal acoustic features typically used in MIR frameworks. Several issues prevent this task to be straightforwardly achieved. Such features are designed for analysis and not for synthesis, thus favoring high-level description over easily inverted acoustic representation. Whereas some previous studies already considered the problem of synthesizing audio from features such as Mel-Frequency Cepstral Coefficients, they mainly relied on the explicit formula used to compute those features in order to inverse them. Here, we instead adopt a simple blind approach, where arbitrary sets of features can be used during synthesis and where reconstruction is exemplar-based. After testing the approach on a speech synthesis from well known features problem, we apply it to the more complex task of inverting songs from the Million Song Dataset. What makes this task harder is twofold. First, that features are irregularly spaced in the temporal domain according to an onset-based segmentation. Second the exact method used to compute these features is unknown, although the features for new audio can be computed using their API as a black-box. In this paper, we detail these difficulties and present a framework to nonetheless attempting such synthesis by concatenating audio samples from a training dataset, whose features have been computed beforehand. Samples are selected at the segment level, in the feature space with a simple nearest neighbor search. Additionnal constraints can then be defined to enhance the synthesis pertinence. Preliminary experiments are presented using RWC and GTZAN audio datasets to synthesize tracks from the Million Song Dataset.
arxiv.org