I just got back from the INCF meeting in the Netherlands, where I had the opportunity to speak (along with Stephen) about open collaboration in computational neuroscience, and to see all the pieces of the puzzle coming together that will help make this work. I think some of the complementary technologies that will help model validation against data become achievable across scales and platforms are starting to reach maturity. This will also help make the work that we start here (with behavioral testing) relevant to future worm tests (e.g. neuron tests), and to parts of models that we haven't yet imagined.
Michael mentioned testing of the movement_validation repository itself. We've called this sort of testing (e.g. are features being generated from the data, do they match across implementations, etc) "verification testing", whereas tests of the scientific validity of the model (e.g. do the features in the data match the features in the model) we called "validation testing". While in principle this verification testing could be done with SciUnit, I've in the past chosen not to do this sort of thing with SciUnit, simply because there are other reasonable tools for the job, or because someone else had already implemented it. For example, neuron models in Open Source Brain already have verification tests (written by Padraig Gleeson) that help ensure that some given simulation output (e.g. spike times) is identical as simulators change. So for that project I am focusing on validation tests which check whether that simulator output is consistent with experimental data.
So if you also want to implement the worm movement verification tests in SciUnit, I'd be OK with that; it might be a good test case before moving on to validation tests of the sim against those features. But if you have something else in mind for the verification tests, or want to use tests you've already begun developing for the verification, that would be fine, too.
Thoughts?