Auditory Graphs from Denoising Real Images Using Fully Symmetric Convolutional Neural Networks
Author(s) -
Rodrigo F. Cádiz,
Lothar Droppelmann,
Max Guzmán,
Cristián Tejos
Publication year - 2021
Language(s) - English
Resource type - Conference proceedings
DOI - 10.21785/icad2021.028
Subject(s) - computer science , sonification , convolutional neural network , usability , pixel , artificial intelligence , graph , noise reduction , chart , pattern recognition (psychology) , computer vision , speech recognition , human–computer interaction , theoretical computer science , statistics , mathematics
Auditory graphs are a very useful way to deliver numerical information to visually impaired users. Several tools have been proposed for chart data sonification, including audible spreadsheets, custom interfaces, interactive tools and automatic models. In the case of the latter, most of these models are aimed towards the extraction of contextual information and not many solutions have been proposed for the generation of an auditory graph directly from the pixels of an image by the automatic extraction of the underlying data. These kind of tools can dramatically augment the availability and usability of auditory graphs for the visually impaired community. We propose a deep learning-based approach for the generation of an automatic sonification of an image containing a bar or a line chart using only pixel information. In particular, we took a denoising approach to this problem, based on a fully symmetric convolutional neural network architecture. Our results show that this approach works as a basis for the automatic sonification of charts directly from the information contained in the pixels of an image.
Accelerating Research
Robert Robinson Avenue,
Oxford Science Park, Oxford
OX4 4GP, United Kingdom
Address
John Eccles HouseRobert Robinson Avenue,
Oxford Science Park, Oxford
OX4 4GP, United Kingdom