Generalized Sparse Learning of Linear Models Over the Complete Subgraph Feature Set

Ichigaku Takigawa, Hiroshi Mamitsuka

Research output: Contribution to journalArticleScientificpeer-review

4 Citations (Scopus)

Abstract

Supervised learning over graphs is an intrinsically difficult problem: simultaneous learning of relevant features from the complete subgraph feature set, in which enumerating all subgraph features occurring in given graphs is practically intractable due to combinatorial explosion. We show that 1) existing graph supervised learning studies, such as Adaboost, LPBoost, and LARS/LASSO, can be viewed as variations of a branch-and-bound algorithm with simple bounds, which we call Morishita-Kudo bounds; 2) We present a direct sparse optimization algorithm for generalized problems with arbitrary twice-differentiable loss functions, to which Morishita-Kudo bounds cannot be directly applied; 3) We experimentally showed that i) our direct optimization method improves the convergence rate and stability, and ii) L1-penalized logistic regression (L1-LogReg) by our method identifies a smaller subgraph set, keeping the competitive performance, iii) the learned subgraphs by L1-LogReg are more size-balanced than competing methods, which are biased to small-sized subgraphs.
Original languageEnglish
Article number7469410
Pages (from-to)617-624
Number of pages8
JournalIEEE Transactions on Pattern Analysis and Machine Intelligence
Volume39
Issue number3
DOIs
Publication statusPublished - 13 Feb 2017
MoE publication typeA1 Journal article-refereed

Keywords

  • supervised learning for graphs
  • graph mining
  • sparsity-inducing regularization
  • block coordinate gradient descent
  • simultaneous feature learning

Fingerprint

Dive into the research topics of 'Generalized Sparse Learning of Linear Models Over the Complete Subgraph Feature Set'. Together they form a unique fingerprint.

Cite this