Research Article

Arabic Syntactic Diacritics Restoration Using BERT Models

Table 5

The comparison between the proposed BERT model and the state-of-the-art systems on the ATB test dataset.

MethodAll WERMorph WERCEER (%)

MaxEnt tagger [10]18.00%7.90%10.10
Rule-based tagger [41]9.97
MADA tagger [42]14.90%5.50%9.40
Random forest tagger [28]13.70%4.30%9.40
Scoring of a language model [5]12.50%3.10%9.11
Confused subset resolution [43]11.60%3.00%8.60
Scoring of a language model [16]10.87%3.00%7.87
SVM tagger [29]6.8
MADAMIRA + character RNN tagger [21]8.40%2.30%6.10
Character RNN tagger [20]9.07%4.34%4.73
Word level MaxEnt/BiLSTM tagger [29]5.3
Word level MaxEnt/BiLSTM tagger + distillation of knowledge + embeddings based on characters [4]4.3
BERT tagger (two steps fine-tuning)2.94