Current state-of-the-art in OCR (e.g. what Google does in Drive/Vision) is already to incorporate a language model to generate reasonable output. (So the equivalent of "oECLARATION" in the English above is already not going to be generated.)
So the equivalent of what ChatGPT is doing here would be useful in cases where OCR-ed text exists but the image itself is not available: because in the latter case it's always going to be better to throw away the existing poor OCR and just do OCR from scratch.