z-logo
open-access-imgOpen Access
Bridging the gap between machine translation output and images in multimodal documents
Author(s) -
Thiago Blanch Pires,
Augusto Velloso dos Santos Espindola
Publication year - 2021
Publication title -
cadernos de tradução
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 0.1
H-Index - 1
eISSN - 2175-7968
pISSN - 1414-526X
DOI - 10.5007/2175-7968.2021.e75483
Subject(s) - multimodality , bridging (networking) , computer science , machine translation , focus (optics) , natural language processing , translation (biology) , artificial intelligence , human–computer interaction , interface (matter) , information retrieval , data science , world wide web , computer network , biochemistry , chemistry , physics , bubble , maximum bubble pressure method , messenger rna , parallel computing , optics , gene
The aim of this article is to report on recent findings concerning the use of Google Translate outputs in multimodal contexts. Development and evaluation of machine translation often focus on verbal mode, but accounts by the area on the exploration text-image relations in multimodal documents translated automatically are rare. Thus, this work seeks to describe just what are such relations and how to describe them, organized in two parts: firstly, by exploring the problem through an interdisciplinary interface, involving Machine Translation and Multimodality to analyze some examples from the Wikihow website; secondly, by reporting on recent investigation on suitable tools and methods to properly annotate these issues from within a long-term purpose to assemble a corpus. Finally, this article provides a discussion on the findings, including some limitations and perspectives for future research.

The content you want is available to Zendy users.

Already have an account? Click here to sign in.
Having issues? You can contact us here