Hi Alexei,
Thanks for the quick reply. At the moment, I am doing some testing with datasets of 3-4k taxa and ~10kb using Tesla Kepler K40M GPUs with 12GB memory. The analyses start and 1m generations take 3-4h, so it may be feasible to run at least 100m generations. I don't know if mixing will be good, though. If things look bad, I may consider the possibility of using topological backbones or even fixed topologies. Anyway, my point is that with the development of more and more powerful GPUs, it will make sense to give BEAST a shot with very large datasets, and therefore this limit on the number of taxa will become a more serious issue.
Cheers,
Victor