Fully Understanding The Hashing Trick | Awesome Learning to Hash Add your paper to Learning2Hash

Fully Understanding The Hashing Trick

Freksen Casper Benjamin, Kamma Lior, Larsen Kasper Green. Arxiv 2018

[Paper]    
ARXIV Independent

Feature hashing, also known as {\em the hashing trick}, introduced by Weinberger et al. (2009), is one of the key techniques used in scaling-up machine learning algorithms. Loosely speaking, feature hashing uses a random sparse projection matrix A:RnRm (where mn) in order to reduce the dimension of the data from n to m while approximately preserving the Euclidean norm. Every column of A contains exactly one non-zero entry, equals to either 1 or 1. Weinberger et al. showed tail bounds on \(|Ax|2^2\). Specifically they showed that for every ϵ,δ, if \(|x|{\infty} / |x|2\) is sufficiently small, and m is sufficiently large, then $Pr[||Ax|22|x|22|<ϵ|x|22]1δ.TheseboundswerelaterextendedbyDasgupta\etal(2010)andmostrecentlyrefinedbyDahlgaardetal.(2017),however,thetruenatureoftheperformanceofthiskeytechnique,andspecificallythecorrecttradeoffbetweenthepivotalparameters|x|{\infty} / |x|_2, m, \epsilon, \deltaremainedanopenquestion.Wesettlethisquestionbygivingtightasymptoticboundsontheexacttradeoffbetweenthecentralparameters,thusprovidingacompleteunderstandingoftheperformanceoffeaturehashing.Wecomplementtheasymptoticboundwithempiricaldata,whichshowsthattheconstantshidingintheasymptoticnotationare,infact,verycloseto1$, thus further illustrating the tightness of the presented bounds in practice.

Similar Work