--
You received this message because you are subscribed to the Google Groups "SIGARAB: Special Interest Group on Arabic Natural Language Processing" group.
To unsubscribe from this group and stop receiving emails from it, send an email to sigarab+u...@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/sigarab/CABRLrSpbkpaX6kqqWMmPowfACbgOqAWYMFNiS%3Do82H%2B9jj6XHA%40mail.gmail.com.
SA Tariq,
You may benefit from CL-AraBERT (CLassical AraBERT). It is an AraBERT-based pre-trained model that is further pre-trained on about 1.05B-word Classical Arabic dataset (after being initially pre-trained on MSA datasets), to make it a better fit for NLP tasks on Classical Arabic text.
I think it is inherently suitable for your task because the model is pre-trained using two unsupervised tasks (as in pretraining BERT/AraBERT): the Masked Language Model task (MLM), and the Next Sentence Prediction (NSP) task.
You can download CL-AraBERT’s checkpoints from this GitHub link https://github.com/RanaMalhas/QRCD/blob/main/README.md#cl-arabert-pre-trained-language-model
Regards,
Rana Malhas
--