Last weeks learnings at the Benelearn conference about how to compress your Neural Networks to fit into less memory
Compressing your NN by 1/64 without hurting your results too-much, it just sounds like dark magic! Great talk by +Kilian Weinberger.
Compressing your NN by 1/64 without hurting your results too-much, it just sounds like dark magic! Great talk by +Kilian Weinberger.
3 Replies to “Last weeks learnings at the Benelearn conference about how to compress your Neural Networks to fit into less memory”
I always wondered if you could go the opposite way and use that as seed for a larger network.
+Henk Poley did you read this? http://arxiv.org/abs/1503.02531
Hmm, as a sort of hack you could train dozens of small predictors, put them in an ensemble, and then distill a "larger" predictor that is smaller than the ensemble together.
Interesting results nonetheless. Using the lesser probability predictions as a representation of internal state (yes a car looks a lot like a truck).