@article {10.3844/jcssp.2023.599.609, article_type = {journal}, title = {Improving Arabic Named Entity Recognition with a Modified Transformer Encoder}, author = {Alsultani, Hamid Sadeq Mahdi and Aliwy, Ahmed H.}, volume = {19}, number = {5}, year = {2023}, month = {Apr}, pages = {599-609}, doi = {10.3844/jcssp.2023.599.609}, url = {https://thescipub.com/abstract/jcssp.2023.599.609}, abstract = {This article investigates the use of a transformer encoder for Arabic Named Entity Recognition (NER). The classic transformer that was originally proposed for machine translation adopts the absolute sinusoidal position embedding which is aware of distance but unfortunately is not aware of the directionality. However, in the NER task, both distance and orientation are crucial. Therefore, in this study, instead of using absolute sinusoidal position encoding, we employ relative positional encoding and incorporate the directionality information in our NER model. More specifically, our proposed model uses Bidirectional Long Short-Term Memory (BiLSTM) for encoding every input token. Then, the output of the encoder is fed to the multi-head attention where both the distance and directionality information are incorporated. The decoder layer with a simple fully connected layer takes as input, the result of the attention layer, and the prediction layer with Conditional Random Fields (CRF) predicts the tag of each token. We validate our proposed approach on two merged public datasets, namely, ANER corp and AQMAR. Our experiment results demonstrate significant improvements when compare to the vanilla Transformer with absolute sinusoidal position encoding while achieving a state-of-the-art result on a merged two Arabic public datasets.}, journal = {Journal of Computer Science}, publisher = {Science Publications} }