Hi Xabi,
So '-C' parameter can sometimes help but with the sheer number of contigs your inputing I am not sure if that will actually affect the run times. So I'll recommend two potential angles you can take:
1. First your working with ~1.6M contigs and a cutoff 1kbp. While I have tested Binsanity down to contigs of 1000bp I find that setting a cuttoff at ~ 2000bp typically speeds up the run significantly and does not remove any bin quality. Ultimately below 2000kbp while these contigs can often be useful they also often have more variable coverage profiles and composition metrics that may not align directly with the actual source genome, often when I include contigs this small most end up unbinned or I end up having to do quite a lot more manual genome refinement using anvio to confirm contig assignment. Increasing your cut-off to 2000bp would be the quickest way to speed up the run and reduce complexity.
2. Your other option in terms of getting BinSanity to run with that many contigs gets a little trickier. One of the main reasons I haven't pushed much further from the implementation of 'Binsanity-lc' in trying to reduce memory is that all the methods I have tried beyond the current implementation start to see some amount of loss in the quality of resultant bins. I give you this work around with that caveat. So from what I understand your first attempt completed the K-MEANS clustering step before canceling out meaning that you have produced a directory 'BINSANITY-RESULTS' in which you should see the following directories (or something similar depending on where it canceled:
BINSANITY-INITIAL BinSanityLC_binsanity_checkm BinSanityLC-BinsanityLC-log.txt BinSanityLC.checkm.logfile BinSanityLC-KMEAN-BINS
Now the initial K-MEANS bins can be found in 'BinSanityLC-KMEAN-BINS'. If you cannot get 'Binsanity-lc' to run on your system in any other way you can take the 'BINS' in this directory. Then run individual instances of `Binsanity-lc` or `Binsanity-wf`. Take the final-genomes from each individual run on each individual KMEAN BIN and combine them at the the end to get your final genome set. Again I will say doing this may cause you to loose some amount of genome quality, but it is one of the few ways I have found to get around memory related issues. You may still though run into the 200 hour time limit though depending on your system configuration.
Please let me know if any of this works or if you need more clarification!
-Elaina