Bayesian Compression for Deep Learning
                                            conference paper
                                        
                                    
                                            Compression and computational efficiency in deep learning have become a problem of great significance. In this work, we argue that the most principled and effective way to attack this problem is by adopting a Bayesian point of view, where through sparsity inducing priors we prune large parts of the network. We introduce two novelties in this paper: 1) we use hierarchical priors to prune nodes instead of individual weights, and 2) we use the posterior uncertainties to determine the optimal fixed point precision to encode the weights. Both factors significantly contribute to achieving the state of the art in terms of compression rates, while still staying competitive with methods designed to optimize for speed or energy
efficiency.
                                    efficiency.
TNO Identifier
                                        
                                            782558
                                        
                                    Source title
                                        
                                            31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA.
                                        
                                    Files
                                        
                                            
                                                To receive the publication files, please send an e-mail request to TNO Repository.