Adapting Models to Signal Degradation using Distillation
Author(s) -
Jong-Chyi Su,
Subhransu Maji
Publication year - 2017
Publication title -
computer vision and pattern recognition
Language(s) - English
Resource type - Conference proceedings
DOI - 10.5244/c.31.21
Subject(s) - degradation (telecommunications) , signal (programming language) , distillation , computer science , signal processing , environmental science , process engineering , biological system , chemistry , chromatography , engineering , digital signal processing , telecommunications , programming language , computer hardware , biology
Model compression and knowledge distillation have been successfully applied for cross-architecture and cross-domain transfer learning. However, a key requirement is that training examples are in correspondence across the domains. We show that in many scenarios of practical importance such aligned data can be synthetically generated using computer graphics pipelines allowing domain adaptation through distillation. We apply this technique to learn models for recognizing low-resolution images using labeled high-resolution images, non-localized objects using labeled localized objects, line-drawings using labeled color images, etc. Experiments on various fine-grained recognition datasets demonstrate that the technique improves recognition performance on the low-quality data and beats strong baselines for domain adaptation. Finally, we present insights into workings of the technique through visualizations and relating it to existing literature.
Accelerating Research
Robert Robinson Avenue,
Oxford Science Park, Oxford
OX4 4GP, United Kingdom
Address
John Eccles HouseRobert Robinson Avenue,
Oxford Science Park, Oxford
OX4 4GP, United Kingdom