z-logo
open-access-imgOpen Access
Adapting Models to Signal Degradation using Distillation
Author(s) -
Jong-Chyi Su,
Subhransu Maji
Publication year - 2017
Publication title -
computer vision and pattern recognition
Language(s) - English
Resource type - Conference proceedings
DOI - 10.5244/c.31.21
Subject(s) - degradation (telecommunications) , signal (programming language) , distillation , computer science , signal processing , environmental science , process engineering , biological system , chemistry , chromatography , engineering , digital signal processing , telecommunications , programming language , computer hardware , biology
Model compression and knowledge distillation have been successfully applied for cross-architecture and cross-domain transfer learning. However, a key requirement is that training examples are in correspondence across the domains. We show that in many scenarios of practical importance such aligned data can be synthetically generated using computer graphics pipelines allowing domain adaptation through distillation. We apply this technique to learn models for recognizing low-resolution images using labeled high-resolution images, non-localized objects using labeled localized objects, line-drawings using labeled color images, etc. Experiments on various fine-grained recognition datasets demonstrate that the technique improves recognition performance on the low-quality data and beats strong baselines for domain adaptation. Finally, we present insights into workings of the technique through visualizations and relating it to existing literature.

The content you want is available to Zendy users.

Already have an account? Click here to sign in.
Having issues? You can contact us here
Accelerating Research

Address

John Eccles House
Robert Robinson Avenue,
Oxford Science Park, Oxford
OX4 4GP, United Kingdom