BERT-based Models for Next Version?

245 views
Skip to first unread message

Brad Windsor

unread,
Sep 10, 2020, 1:51:10 AM9/10/20
to tesseract-dev
Hi folks!

First off, great library, super useful. Wanted to ask if there had been any discussions on the next NN architecture for tesseract 5:

1. Will we move the LSTM-based code to a transformer-based architecture? [0] 
2. Going further, we know that large pre-trained Transformer models have good next-word prediction properties which can correct text filled with OCR errors[1]. Have we tried fine-tuning such models for the OCR task itself? If not, would a small demo be welcome? 


Thanks,
Brad
Reply all
Reply to author
Forward
0 new messages