z-logo
open-access-imgOpen Access
Impact of Using Bidirectional Encoder Representations from Transformers (BERT) Models for Arabic Dialogue Acts Identification
Author(s) -
Alaa Joukhadar,
Nada Ghneim,
Ghaida Rebdawi
Publication year - 2021
Publication title -
ingénierie des systèmes d'information/ingénierie des systèmes d'information
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 0.161
H-Index - 8
eISSN - 2116-7125
pISSN - 1633-1311
DOI - 10.18280/isi.260506
Subject(s) - computer science , transformer , natural language processing , encoder , artificial intelligence , representation (politics) , utterance , language model , arabic , identification (biology) , task (project management) , speech recognition , linguistics , philosophy , physics , botany , management , quantum mechanics , voltage , politics , political science , law , economics , biology , operating system
In Human-Computer dialogue systems, the correct identification of the intent underlying a speaker's utterance is crucial to the success of a dialogue. Several researches have studied the Dialogue Act Classification (DAC) task to identify Dialogue Acts (DA) for different languages. Recently, the emergence of Bidirectional Encoder Representations from Transformers (BERT) models, enabled establishing state-of-the-art results for a variety of natural language processing tasks in different languages. Very few researches have been done in the Arabic Dialogue acts identification task. The BERT representation model has not been studied yet in Arabic Dialogue acts detection task. In this paper, we propose a model using BERT language representation to identify Arabic Dialogue Acts. We explore the impact of using different BERT models: AraBERT Original (v0.1, v1), AraBERT Base (v0.2, and v2) and AraBERT Large (v0.2, and v2), which are pretrained on different Arabic corpora (different in size, morphological segmentation, language model window, …). The comparison was performed on two available Arabic datasets. Using AraBERTv0.2-base model for dialogue representations outperformed all other pretrained models. Moreover, we compared the performance of AraBERTv0.2-base model to the state-of-the-art approaches applied on the two datasets. The comparison showed that this representation model outperformed the performance both state-of-the-art models.

The content you want is available to Zendy users.

Already have an account? Click here to sign in.
Having issues? You can contact us here