Part of the issue here is that Dent was an author we me on our Cactus
alignment program, I helped him develop the Evolver scripts,
I'm privileged to work with Dent and, finally, we (myself and others
at UCSC, not Dent) are planning on submitting Cactus entries to the
Alignathon! So, to avoid any real or imagined bias towards UCSC, we
wanted everything to be available to everyone, despite this creating
the problems you outline.
Note, no one knows the true answer to the 12 fly alignments, and as
long as people submit answers for them, these should provide
interesting complementary metrics. That said, I think it would be good
to do a run with novel simulations - perhaps this could be done after
the initial submission stage - as a check for overtraining to the
provided simulations. I also agree that Dent should provide the exact
recipe he used for generating the alignments, so that people can make
there own sets and play around.
Benedict
Great question. We went with the totally open approach to solutions for Alignathon because as a lab we have a strong interest in alignment and some of our lab members will be working on their own submissions. I have no involvement with them in their submissions. However, we wanted to be totally above-board on everything so instead of dealing with any doubts about fairness we thought it would be best to let everyone see exactly the same thing: everything.
Incidentally, the size of the "true" alignments will likely throw a person off, those alignments contain oodles more detail than an aligner could hope to recover since they hold all of the intermediate branch-point species that were simulated in order to get the leaf-node species that are actually provided.
I do share your hope that people don't over-fit to the solutions (or really "fit" at all) but I was more worried about the specter of impropriety.
d
Thanks for the suggestion! I think this is a great idea and I would love to be able to include such a framework in future Alignathons; doing so would be fairly simple from my point of view. I can envision a submission that is geared specifically to a set of known input sequences where the code submitted is essentially a recipe script or Makefile that performs all the requisite steps to make the alignment. This sort of submission is something could be possible under a future Alignathon. As an aside, short of actually providing a Makefile we are expecting that teams that are submitting to Alignathon will produce a written recipe of how they created their alignments.
For the question "What are we evaluating, the tool or the product of the tool," I know we tried to be careful in the Assemblathon paper to refer to assemblies rather than assemblers because we weren't looking at the results of the asemblers alone. There were data preprocessing steps, some assemblers did their own scaffolding, some used off the shelf scaffolding, and there were data post-processing steps. Some teams were comprised of a single individual and there were a few teams that were staffed by multiple engineers. We thought it would have been misleading to focus on assembers at that point and decided to focus on the product, the assemblies. I imagine we'll be taking a similar approach for Alignathon. Best,
d