Aghasi, Alireza, Nam Nguyen, and Justin Romberg. 2016. “Net-Trim: A Layer-Wise Convex Pruning of Deep Neural Networks.” arXiv:1611.05162 [Cs, Stat]
Blalock, Davis, Jose Javier Gonzalez Ortiz, Jonathan Frankle, and John Guttag. 2020. “What Is the State of Neural Network Pruning?” arXiv:2003.03033 [Cs, Stat]
Borgerding, Mark, and Philip Schniter. 2016. “Onsager-Corrected Deep Networks for Sparse Linear Inverse Problems.” arXiv:1612.01183 [Cs, Math]
Cai, Han, Chuang Gan, Tianzhe Wang, Zhekai Zhang, and Song Han. 2020. “Once-for-All: Train One Network and Specialize It for Efficient Deployment.”
Chen, Tianqi, Ian Goodfellow, and Jonathon Shlens. 2015. “Net2Net: Accelerating Learning via Knowledge Transfer.” arXiv:1511.05641 [Cs]
Chen, Wenlin, James T. Wilson, Stephen Tyree, Kilian Q. Weinberger, and Yixin Chen. 2015. “Compressing Convolutional Neural Networks.” arXiv:1506.04449 [Cs]
Cheng, Yu, Duo Wang, Pan Zhou, and Tao Zhang. 2017. “A Survey of Model Compression and Acceleration for Deep Neural Networks.” arXiv:1710.09282 [Cs]
Cutajar, Kurt, Edwin V. Bonilla, Pietro Michiardi, and Maurizio Filippone. 2017. “Random Feature Expansions for Deep Gaussian Processes.”
Daniely, Amit. 2017. “Depth Separation for Neural Networks.” arXiv:1702.08489 [Cs, Stat]
Frankle, Jonathan, and Michael Carbin. 2019. “The Lottery Ticket Hypothesis: Finding Sparse, Trainable Neural Networks.” arXiv:1803.03635 [Cs]
Garg, Sahil, Irina Rish, Guillermo Cecchi, and Aurelie Lozano. 2017. “Neurogenesis-Inspired Dictionary Learning: Online Model Adaption in a Changing World.”
In arXiv:1701.06106 [Cs, Stat]
Gelder, Maxwell Van, Mitchell Wortsman, and Kiana Ehsani. n.d. “Deconstructing the Structure of Sparse Neural Networks.” In, 6.
Globerson, Amir, and Roi Livni. 2016. “Learning Infinite-Layer Networks: Beyond the Kernel Trick.” arXiv:1606.05316 [Cs]
Gray, Scott, Alec Radford, and Diederik P Kingma. n.d. “GPU Kernels for Block-Sparse Weights,” 12.
Gu, Albert, Isys Johnson, Karan Goel, Khaled Saab, Tri Dao, Atri Rudra, and Christopher Ré. 2021. “Combining Recurrent, Convolutional, and Continuous-Time Models with Linear State Space Layers.”
In Advances in Neural Information Processing Systems
, 34:572–85. Curran Associates, Inc.
Ha, David, Andrew Dai, and Quoc V. Le. 2016. “HyperNetworks.” arXiv:1609.09106 [Cs]
Hardt, Moritz, Benjamin Recht, and Yoram Singer. 2015. “Train Faster, Generalize Better: Stability of Stochastic Gradient Descent.” arXiv:1509.01240 [Cs, Math, Stat]
Hayou, Soufiane, Jean-Francois Ton, Arnaud Doucet, and Yee Whye Teh. 2020. “Pruning Untrained Neural Networks: Principles and Analysis.” arXiv:2002.08797 [Cs, Stat]
Hazimeh, Hussein, Natalia Ponomareva, Petros Mol, Zhenyu Tan, and Rahul Mazumder. 2020. “The Tree Ensemble Layer: Differentiability Meets Conditional Computation,”
He, Yihui, Ji Lin, Zhijian Liu, Hanrui Wang, Li-Jia Li, and Song Han. 2019. “AMC: AutoML for Model Compression and Acceleration on Mobile Devices.” arXiv:1802.03494 [Cs]
Howard, Andrew G., Menglong Zhu, Bo Chen, Dmitry Kalenichenko, Weijun Wang, Tobias Weyand, Marco Andreetto, and Hartwig Adam. 2017. “MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications.” arXiv:1704.04861 [Cs]
Iandola, Forrest N., Song Han, Matthew W. Moskewicz, Khalid Ashraf, William J. Dally, and Kurt Keutzer. 2016. “SqueezeNet: AlexNet-Level Accuracy with 50x Fewer Parameters and <0.5MB Model Size.” arXiv:1602.07360 [Cs]
LeCun, Yann, John S. Denker, and Sara A. Solla. 1990. “Optimal Brain Damage.”
In Advances in Neural Information Processing Systems
Lee, Holden, Rong Ge, Tengyu Ma, Andrej Risteski, and Sanjeev Arora. 2017. “On the Ability of Neural Nets to Express Distributions.”
In arXiv:1702.07028 [Cs]
Lemhadri, Ismael, Feng Ruan, Louis Abraham, and Robert Tibshirani. 2021. “LassoNet: A Neural Network with Feature Sparsity.” Journal of Machine Learning Research
22 (127): 1–29.
Liebenwein, Lucas, Cenk Baykal, Brandon Carter, David Gifford, and Daniela Rus. 2021. “Lost in Pruning: The Effects of Pruning Neural Networks Beyond Test Accuracy.” arXiv:2103.03014 [Cs]
Lobacheva, Ekaterina, Nadezhda Chirkova, and Dmitry Vetrov. 2017. “Bayesian Sparsification of Recurrent Neural Networks.”
In Workshop on Learning to Generate Natural Language
Louizos, Christos, Max Welling, and Diederik P. Kingma. 2017. “Learning Sparse Neural Networks Through \(L_0\) Regularization.” arXiv:1712.01312 [Cs, Stat]
Mariet, Zelda Elaine. 2016. “Learning and enforcing diversity with Determinantal Point Processes.”
Thesis, Massachusetts Institute of Technology.
Molchanov, Dmitry, Arsenii Ashukha, and Dmitry Vetrov. 2017. “Variational Dropout Sparsifies Deep Neural Networks.”
In Proceedings of ICML
Narang, Sharan, Eric Undersander, and Gregory Diamos. 2017. “Block-Sparse Recurrent Neural Networks.” arXiv:1711.02782 [Cs, Stat]
Pan, Wei, Hao Dong, and Yike Guo. 2016. “DropNeuron: Simplifying the Structure of Deep Neural Networks.” arXiv:1606.07326 [Cs, Stat]
Renda, Alex, Jonathan Frankle, and Michael Carbin. 2020. “Comparing Rewinding and Fine-Tuning in Neural Network Pruning.” arXiv:2003.02389 [Cs, Stat]
Scardapane, Simone, Danilo Comminiello, Amir Hussain, and Aurelio Uncini. 2016. “Group Sparse Regularization for Deep Neural Networks.” arXiv:1607.00485 [Cs, Stat]
Shi, Lei, Shikun Feng, and ZhifanZhu. 2016. “Functional Hashing for Compressing Neural Networks.” arXiv:1605.06560 [Cs]
Srinivas, Suraj, and R. Venkatesh Babu. 2016. “Generalized Dropout.” arXiv:1611.06791 [Cs]
Steeg, Greg Ver, and Aram Galstyan. 2015. “The Information Sieve.” arXiv:1507.02284 [Cs, Math, Stat]
Ullrich, Karen, Edward Meeds, and Max Welling. 2017. “Soft Weight-Sharing for Neural Network Compression.” arXiv Preprint arXiv:1702.04008
Urban, Gregor, Krzysztof J. Geras, Samira Ebrahimi Kahou, Ozlem Aslan, Shengjie Wang, Rich Caruana, Abdelrahman Mohamed, Matthai Philipose, and Matt Richardson. 2016. “Do Deep Convolutional Nets Really Need to Be Deep (Or Even Convolutional)?” arXiv:1603.05691 [Cs, Stat]
Venturi, Daniele, and Xiantao Li. 2022. “The Mori-Zwanzig Formulation of Deep Learning.”
Wang, Yunhe, Chang Xu, Chao Xu, and Dacheng Tao. 2019. “Packing Convolutional Neural Networks in the Frequency Domain.” IEEE transactions on pattern analysis and machine intelligence
41 (10): 2495–2510.
Wang, Yunhe, Chang Xu, Shan You, Dacheng Tao, and Chao Xu. 2016. “CNNpack: Packing Convolutional Neural Networks in the Frequency Domain.”
In Advances in Neural Information Processing Systems 29
, edited by D. D. Lee, M. Sugiyama, U. V. Luxburg, I. Guyon, and R. Garnett, 253–61. Curran Associates, Inc.
Wang, Zhangyang, Shiyu Chang, Qing Ling, Shuai Huang, Xia Hu, Honghui Shi, and Thomas S. Huang. 2016. “Stacked Approximated Regression Machine: A Simple Deep Learning Approach.”
Warden, Pete, and Daniel Situnayake. 2020. TinyML: Machine Learning with TensorFlow Lite on Arduino and Ultra-Low-Power Microcontrollers. O’Reilly Media, Incorporated.
You, Haoran, Chaojian Li, Pengfei Xu, Yonggan Fu, Yue Wang, Xiaohan Chen, Richard G. Baraniuk, Zhangyang Wang, and Yingyan Lin. 2019. “Drawing Early-Bird Tickets: Toward More Efficient Training of Deep Networks.”
No comments yet. Why not leave one?