I'm trying to understand in detail exactly how the results of the adaptive spectrogram matches up with musical time. On sonic annotator, I ran the vamp:qm-vamp-plugins:qm-adaptivespectrogram:output transform on some music (127.663 seconds), using the -d option. "with plugin step size 1024 and block size 2048 (adapter step and block size 16384)" The output is a 22106 X 1024 matrix.
My question: how do I interpret the data in terms of time? For each of the 22106 vectors, I would like to associate it with the proper span of musical time, and I'm not sure how to do that.
The documentation of the adaptive spectrogram does help me understand what the algorithm is doing, but it doesn't really answer my question.