Skip to yearly menu bar Skip to main content


Poster

Fast Iterative Hard Thresholding Methods with Pruning Gradient Computations

Yasutoshi Ida · Sekitoshi Kanai · Atsutoshi Kumagai · Tomoharu Iwata · Yasuhiro Fujiwara

[ ]
Thu 12 Dec 11 a.m. PST — 2 p.m. PST

Abstract:

We accelerate the iterative hard thresholding (IHT) method, which finds (k) important elements from a parameter vector in a linear regression model. Although the plain IHT repeatedly updates the parameter vector during the optimization, computing gradients is the main bottleneck. Our method safely prunes unnecessary gradient computations to reduce the processing time.The main idea is to efficiently construct a candidate set, which contains (k) important elements in the parameter vector, for each iteration. Specifically, before computing the gradients, we prune unnecessary elements in the parameter vector for the candidate set by utilizing upper bounds on absolute values of the parameters. Our method guarantees the same optimization results as the plain IHT because our pruning is safe. Experiments show that our method is up to 73 times faster than the plain IHT without degrading accuracy.

Live content is unavailable. Log in and register to view live content