I'll take a stab at this. Others please jump in and/or pile on.
Do keep in mind that 97% threshold is based on alignments of the ENTIRE 16S gene. v4 region with 515/806 primers is only about 250 bases with above average entropy so really a lower threshold should be considered. As to taxonomic assignments, I'm not sure why you were getting more deep assignments at 99%, but I would hold any assignments below family with this region (sometimes genus is OK) with relative suspicion and contempt. If I can hazard a guess, I would say that miseq data has enough errors when sequencing amplicons, especially toward the 3' end that perhaps at a higher threshold, there were more representatives of spurious OTU sequences which offered the tax assigner the chance to provide assignments where maybe it wouldn't otherwise. At any rate, clustering at a high threshold is a great way to provide diversity estimates that are wickedly inflated. I particularly like the article by Kunin (2010) on this topic (using 454 data and a different region, but I think the result should still make miseq users be cautious) where they used a SINGLE e coli isolate as a "community" and then clustered at different values. Data that was untrimmed (as with QIIME split_library_fastq default settings -- not entirely true, but close) and clustered at 99% observed nearly 100 OTUs. They found that if they aggressively filtered the data where a base must have <0.2% probability of being an error was required to achieve the desired result, and they could cluster up to 98%. This means filtering data at q27 or greater (set -q to 26 in split_libraries_fastq.py). In doing this you will lose a LOT of data! However, if that data causes you to say something that you shouldn't (we observed a billion jillion OTUs!), that should be data you are happy to shed.
On a typical data set I will retain about 85% of reads using default settings, or maybe 25% of reads if I set -q to 19 (q20 or better), -r to 0, and -p to 0.95. I get much better results personally when all the reads are near to or exactly the same length. I suspect there is a reference for this effect, but I've not found one. Anyone else know?
Nelson et al (2014) put out a cool paper in PLOS (yea yea, those guys suck right now -- PLOS, not Nelson et al) where they reanalyzed miseq data substituting a mock community for the phix component. PhiX is usually how run metrics are determined during sequencing. They found that the amplicon data did not behave similarly to the balanced sequence of PhiX during sequencing and that the error rate was indeed higher than reported by the instrument. They go on to do a nice analysis about how one should process amplicon data, but the error rate observation was the really interesting part to me.
Maybe, like me, your data isn't good enough to filter at q27. If you did that, you would have nothing left to process and then you still can't say anything. Some runs are good enough to filter at q30, others not so much. Because I don't have a mock community to use to more accurately assess error rates during sequencing, I have to rely on the rates reported by the miseq for better or (more likely) for worse. q20 leaves me enough data to do substantive work, but I acknowledge that by amplicon size is 253 (v4). If I filter at q20 (5% error rate), I might expect 2.5 errors per sequence. Because of the issues with similarity clustering (attracting divergent reads to the same cluster seed), I favor distance-based clustering. This is available in swarm (Mahe et al 2014) which is de novo and in qiime 1.9.1 and is super fast due to the use of edit distances for calculation rather than local alignments. Another notable distance-based clustering method is minimum entropy decomposition (Eren et al, 2015). The approach used in dada2 (Callahan et al, 2015) is the same (I think) as the Tikhonov reference you mention, but also offers a denoising function to try to address the problem of systematic errors which occur on the miseq. Sure, we used to hate waiting for denoising 454 data, but it also may be a mistake by the microbial ecology community to have essentially ignored this analogous problem for so long on the new platform. I haven't had the time to make dada2 work for myself yet, but I have had some success using the bayeshammer denoiser (Nikolenko et al, 2013), available in spades 3.6 (Bankevich et al, 2012). That and using read overlap to "correct" errors through consensus can really improve the outcome of Illumina data.
Here are some screenshots from a presentation I gave with regard to data treatment. The first are quality plots of the first two reads (from fastqc). You can see how terrible read 2 is all by itself. Some might find this enough cause to use only the first read, but I was determined. The next shows read1 again (for comparison) and joined data. If you relax fastq-join enough (I use -p 30 -m 30) and do your quality filtering subsequently, you should get a lot of reads to overlap without sacrificing quality. As indicated in the image, I then trimmed the data to 220 bases prior to analysis. This gave me plenty of reads to play with, and the average quality to start with was much better than with read1 alone. These are data from fungal ITS2, so there wasn't a complete overlap for 2x250 data as with 16S v4. I outline some of this process in my wiki on ITS analysis considerations here:
https://github.com/alk224/akutils-v1.2/wiki/ITS-analysis
References below. Hopefully you find this helpful.


Bankevich A., Nurk S., Antipov D., Gurevich A a., Dvorkin M., Kulikov AS., Lesin VM., Nikolenko SI., Pham S., Prjibelski AD., Pyshkin A V., Sirotkin A V., Vyahhi N., Tesler G., Alekseyev M a., Pevzner P a. 2012. SPAdes: A New Genome Assembly Algorithm and Its Applications to Single-Cell Sequencing. Journal of Computational Biology 19:455–477.
Callahan BJ., Mcmurdie PJ., Rosen MJ., Han AW., Johnson AJ., Holmes SP. 2015. DADA2: High resolution sample inference from amplicon data. bioRxiv:0–14.
Eren AM., Morrison HG., Lescault PJ., Reveillaud J., Vineis JH., Sogin ML. 2015. Minimum entropy decomposition : Unsupervised oligotyping for sensitive partitioning of high- throughput marker gene sequences. Isme J 9:968–979.
Kunin V., Engelbrektson A., Ochman H., Hugenholtz P. 2010. Wrinkles in the rare biosphere: Pyrosequencing errors can lead to artificial inflation of diversity estimates. Environmental Microbiology 12:118–123.
Mahé F., Rognes T., Quince C., de Vargas C., Dunthorn M. 2014. Swarm: robust and fast clustering method for amplicon-based studies. PeerJ 2:e593.
Nelson MC., Morrison HG., Benjamino J., Grim SL., Graf J. 2014. Analysis, optimization and verification of Illumina-generated 16S rRNA gene amplicon surveys. PloS one 9:e94249.
Nikolenko SI., Korobeynikov AI., Alekseyev MA. 2013. BayesHammer: Bayesian clustering for error correction in single-cell sequencing. BMC Genomics 14:1–11.