z-logo
open-access-imgOpen Access
Mix Multiple Features to Evaluate the Content and the Linguistic Quality of Text Summaries
Author(s) -
Samira Ellouze,
Maher Jaoua,
Lamia Hadrich Belguith
Publication year - 2017
Publication title -
journal of computing and information technology
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 0.169
H-Index - 27
eISSN - 1846-3908
pISSN - 1330-1136
DOI - 10.20532/cit.2017.1003398
Subject(s) - automatic summarization , computer science , quality (philosophy) , natural language processing , artificial intelligence , correlation , content (measure theory) , machine learning , quality score , baseline (sea) , metric (unit) , oceanography , geometry , mathematics , economics , geology , mathematical analysis , philosophy , operations management , epistemology
In this article, we propose a method of text summary's content and linguistic quality evaluation that is based on a machine learning approach. This method operates by combining multiple features to build predictive models that evaluate the content and the linguistic quality of new summaries (unseen) constructed from the same source documents as the summaries used in the training and the validation of models. To obtain the best model, many single and ensemble learning classifiers are tested. Using the constructed models, we have achieved a good performance in predicting the content and the linguistic quality scores. In order to evaluate the summarization systems, we calculated the system score as the average of the score of summaries that are built from the same system. Then, we evaluated the correlation of the system score with the manual system score. The obtained correlation indicates that the system score outperforms the baseline scores.

The content you want is available to Zendy users.

Already have an account? Click here to sign in.
Having issues? You can contact us here
Accelerating Research

Address

John Eccles House
Robert Robinson Avenue,
Oxford Science Park, Oxford
OX4 4GP, United Kingdom