Since nobody in the know has commented, I'll venture some comments:
The nltk's brown module breaks up sentences using an automatic
sentence tokenizer, so perhaps it is simply wrong. But the NLTK
total is about 10% higher than the counts you cite, so I really
doubt that's the case. I've worked with the corpus and have not
noticed anything like 10% tokenization errors.
Is the figure 52108 authoritative? That's not clear in the article
you link to. If it's based on the parsed version of the corpus it
should be reliable, but then why does the author cite alternative
estimates as well? The NLTK's sents() list includes headlines and
other interspersed textual material (a quick count gave me 2615
"sentences" that do not end with punctuation), and it's likely
following different rules about how to treat quotations etc.
I'd take a good look at the NLTK's sentences, and if you don't
notice pervasive errors, maybe you can just accept it. If you need
something authoritative, your best option might be to get your hands
on the parsed version of the Brown corpus (it's not in the NLTK).
Best,
Alexis