2016
DOI: 10.48550/arxiv.1602.05916
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Local Rademacher Complexity-based Learning Guarantees for Multi-Task Learning

Niloofar Yousefi,
Yunwen Lei,
Marius Kloft
et al.

Abstract: We show a Talagrand-type concentration inequality for Multi-Task Learning (MTL), using which we establish sharp excess risk bounds for MTL in terms of distribution-and data-dependent versions of the Local Rademacher Complexity (LRC). We also give a new bound on the LRC for norm regularized as well as strongly convex hypothesis classes, which applies not only to MTL but also to the standard i.i.d. setting. Combining both results, one can now easily derive fast-rate bounds on the excess risk for many prominent M… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...

Citation Types

0
0
0

Publication Types

Select...

Relationship

0
0

Authors

Journals

citations
Cited by 0 publications
references
References 27 publications
(69 reference statements)
0
0
0
Order By: Relevance

No citations

Set email alert for when this publication receives citations?