We study in this paper flat and hierarchical classification strategies in the context of large-scale taxonomies. To this end, we first propose a multiclass, hierarchical data dependent bound on the generalization error of classifiers deployed in large-scale taxonomies. This bound provides an explanation to several empirical results reported in the literature, related to the performance of flat and hierarchical classifiers. We then introduce another type of bound targeting the approximation error of a family of classifiers, and derive from it features used in a meta-classifier to decide which nodes to prune (or flatten) in a large-scale taxonomy. We finally illustrate the theoretical developments through several experiments conducted on two widely used taxonomies.
|Title of host publication||NIPS'13: Proceedings of the 26th International Conference on Neural Information Processing Systems |
|Publication status||Published - 2013|
|MoE publication type||A4 Article in a conference publication|
|Event||IEEE Conference on Neural Information Processing Systems - Lake Tahoe, United States|
Duration: 5 Dec 2013 → 10 Dec 2013
|Conference||IEEE Conference on Neural Information Processing Systems|
|Period||05/12/2013 → 10/12/2013|