Hi folks!
First off, great library, super useful. Wanted to ask if there had been any discussions on the next NN architecture for tesseract 5:
1. Will we move the LSTM-based code to a transformer-based architecture? [0]
2. Going further, we know that large pre-trained Transformer models have good next-word prediction properties which can correct text filled with OCR errors[1]. Have we tried fine-tuning such models for the OCR task itself? If not, would a small demo be welcome?
Thanks,
Brad