Skip to yearly menu bar Skip to main content


Poster

Transductive Sample Complexities Are Compact

Julian Asilis · Siddartha Devic · Shaddin Dughmi · Vatsal Sharan · Shang-Hua Teng

[ ]
Thu 12 Dec 4:30 p.m. PST — 7:30 p.m. PST

Abstract: We demonstrate a compactness result holding broadly across supervised learning with a general class of loss functions: Any hypothesis class $\mathcal{H}$ is learnable with transductive sample complexity $m$ precisely when all of its finite projections are learnable with sample complexity $m$. We prove that this exact form of compactness holds for realizable and agnostic learning with respect to all \emph{proper} metric loss functions (e.g., any norm on $\mathbb{R}^d$) and any continuous loss on a compact space (e.g., cross-entropy, squared loss). For realizable learning with \emph{improper} metric losses, we show that exact compactness of sample complexity can fail, and provide matching upper and lower bounds of a factor of 2 on the extent to which such sample complexities can differ. We conjecture that larger gaps are possible for the agnostic case. Furthermore, invoking the equivalence between sample complexities in the PAC and transductive models (up to lower order factors, in the realizable case) permits us to directly port our results to the PAC model, revealing an almost-exact form of compactness holding broadly in PAC learning.

Live content is unavailable. Log in and register to view live content