Hi all,
I am attempting to remove OTUs that are present at less than 1% of the total read counts on a per-sample basis, not across the whole dataset. (In other words, going through each sample one at a time and replacing any OTUs that account for <1% of the reads in that sample with zeros.) I'm trying to find a way to get summarize_taxa.py to calculate relative abundances of each OTU instead of each taxa, so that I can then open the table in Excel and use a function to replace all numbers less than 0.01 with 0. I think I should be able to get this to work using the --md_identifier option but I am running into an error.
[awillcox@kure-login1 december_8]$ bsub summarize_taxa.py -i ../otu_table_no_singletons.biom --md_identifier 'OTU ID' -o relative_abundance_otu_table
The output (if any) follows:
Traceback (most recent call last):
File "/nas02/apps/qiime-1.9.0/python/bin/summarize_taxa.py", line 259, in <module>
main()
File "/nas02/apps/qiime-1.9.0/python/bin/summarize_taxa.py", line 235, in main
md_identifier)
File "/nas02/apps/qiime-1.9.0/python/lib/python2.7/site-packages/qiime/summarize_taxa.py", line 43, in make_summary
md_identifier)
File "/nas02/apps/qiime-1.9.0/python/lib/python2.7/site-packages/qiime/summarize_taxa.py", line 101, in sum_counts_by_consensus
"identifier?" % (md_identifier, otu_id))
KeyError: u"Metadata category 'OTU ID' not in OTU denovo84068. Can't continue. Did you pass the correct metadata identifier?"
I've also tried calling '#OTU ID' and get the same error. Any suggestions for how to fix this or a better way to go about removing these OTUs?
Thanks,
Alexandra