z-logo
open-access-imgOpen Access
Research Trends on Deep Transformation Neural Models for Text Analysis in NLP Applications
Author(s) -
T. Chellatamilan,
B. Valarmathi,
K. Santhi
Publication year - 2020
Publication title -
international journal of recent technology and engineering
Language(s) - English
Resource type - Journals
ISSN - 2277-3878
DOI - 10.35940/ijrte.b3838.079220
Subject(s) - computer science , artificial intelligence , automatic summarization , recurrent neural network , deep learning , natural language processing , machine translation , language model , convolutional neural network , artificial neural network , sentiment analysis , transformation (genetics) , machine learning , biochemistry , chemistry , gene
In the recent few years, text analyses with neural models have become more popular due its versatile usages in different software applications. In order to improve the performance of text analytics, there is a huge collection of methods that have been identified and justified by the researchers. Most of these techniques have been efficiently used for text categorization, text generation, text summarization, query formulation, query answering, sentiment analysis and etc. In this review paper, we consolidate a recent literature along with the technical survey on different neural models such as Neural Language Model (NLM), sequence to sequence model (seq2seq), text generation, Bidirectional Encoder Representations from Transformers (BERT), machine translation model (MT), transformation model, attention model from the perception of applying deep machine learning algorithms for text analysis. Applied extensive experiments were conducted on the deep learning model such as Recurrent Neural Network (RNN) / Long Short-Term Memory (LSTM) / Convolutional Neural Network (CNN) and Attentive Transformation model to examine the efficacy of different neural models with the implementation using tensor flow and keras.

The content you want is available to Zendy users.

Already have an account? Click here to sign in.
Having issues? You can contact us here