
Word Embeddings for Constructive Comments Classification
Author(s) -
Diego Uribe
Publication year - 2021
Publication title -
journal of physics. conference series
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 0.21
H-Index - 85
eISSN - 1742-6596
pISSN - 1742-6588
DOI - 10.1088/1742-6596/1824/1/012005
Subject(s) - computer science , natural language processing , word (group theory) , artificial intelligence , word embedding , constructive , focus (optics) , identification (biology) , convolutional neural network , semantics (computer science) , embedding , deep learning , linguistics , philosophy , physics , botany , process (computing) , optics , biology , programming language , operating system
Word embeddings are so significant today that it is common to see their application in multiple natural language tasks. Indeed, word embeddings as the first layer of a deep learning model are widely adopted and they can also be found in multiple natural language tasks such as classification of texts and named entity recognition. The focus of this paper is the identification of constructive online comments through the use of dense vector semantics such as word embeddings. We specifically explore two approaches: learning distributed word representations and the use of a pre-trained word embedding model. We evaluate these word embedding methods on a recently created constructive comments corpus comprised of 12,000 annotated news comments, intended to improve the quality of online discussions. The obtained results show how the performance of complicated architectures like recurrent and convolutional neural networks can be matched by a language model based on learning embeddings.