Fast Training for Large-Scale One-versus-All Linear Classifiers using Tree-Structured Initialization
Author(s) -
Fang Huang,
Minhao Cheng,
ChoJui Hsieh,
Michael P. Friedlander
Publication year - 2019
Publication title -
society for industrial and applied mathematics ebooks
Language(s) - English
Resource type - Book series
DOI - 10.1137/1.9781611975673.32
Subject(s) - initialization , computer science , training (meteorology) , scale (ratio) , artificial intelligence , decision tree , tree (set theory) , training set , machine learning , pattern recognition (psychology) , mathematics , geography , cartography , combinatorics , programming language , meteorology
We consider the problem of training one-versus-all (OVA) linear classifiers for multiclass or multilabel classification when the number of labels is large. A naive extension of OVA to this problem, even with hundreds of cores, usually requires hours for training on large real world datasets. We propose a novel algorithm called OVA-Primal++ that speeds up the training of OVA by using a tree-structured training order, where each classifier is trained using its parent’s classifier as initialization. OVA-Primal++ is both theoretically and empirically faster than the naive OVA algorithm, and yet still enjoys the same highly parallelizability and small memory footprint. Extensive experiments on multiclass and multilabel classification datasets validate the effectiveness of our method.
Accelerating Research
Robert Robinson Avenue,
Oxford Science Park, Oxford
OX4 4GP, United Kingdom
Address
John Eccles HouseRobert Robinson Avenue,
Oxford Science Park, Oxford
OX4 4GP, United Kingdom