I just found the Docs, oops:
I notice in the API docs you have a submodule dedicated to RNN, but not one for CNN.
Is it fair to say that RNN is adequate for MIR from monophonic sources and that CNN provides an improved performance for polyphonic material?
If so, would it be true to say that CNN methods would not yield an improved result for monophonic signals?
I am not an academic or even a Python developer, but I have been developing software for 15 years. My goal is to prototype some processing workflows which will give a really nice tempo map of monophonic sources, with the ability to produce a midi output from an audio input.
As I'm new to MIR, I don't know how tempo mapping is done, but I would think you could model underlying tempo variation as a separate thing from local note deviation. I would think that these would feed into the produced map with different weights, but just speculating.
I ultimately I would anticipate training the network as part of an ongoing project, but I have no clue about the extent to which any trained networks you have are converged for this kind of task.
Now I see from the docs your library has utilities for interfacing with MIDI, it seems to me it's all there. But python and signal processing are outside the areas I've worked in commercially (I have written a real-time graphical spectrum analyser , so know a bit about Nyquist and FFT, and also written a successful captcha breaking prototype using fast forward neural networks. These were both just for fun).
I now plan to work through the API docs and try to implement some stuff, but if you did have any sample projects you think might help that are in the public domain, I would be very grateful.