
Gradient Descent is a Technique for Learning to Learn
Author(s) -
Taposh Kumar Neogy,
Naresh Babu Bynagari
Publication year - 2018
Publication title -
asian journal of humanity, art and literature
Language(s) - English
Resource type - Journals
eISSN - 2312-2021
pISSN - 2311-8636
DOI - 10.18034/ajhal.v5i2.578
Subject(s) - computer science , exploit , artificial intelligence , machine learning , artificial neural network , simple (philosophy) , gradient descent , variety (cybernetics) , philosophy , computer security , epistemology
In machine learning, the transition from hand-designed features to learned features has been a huge success. Regardless, optimization methods are still created by hand. In this study, we illustrate how an optimization method's design can be recast as a learning problem, allowing the algorithm to automatically learn to exploit structure in the problems of interest. On the tasks for which they are taught, our learning algorithms, implemented by LSTMs, beat generic, hand-designed competitors, and they also adapt well to other challenges with comparable structure. We show this on a variety of tasks, including simple convex problems, neural network training, and visual styling with neural art.