Aghasi, Alireza, Nam Nguyen, and Justin Romberg. 2016.

“Net-Trim: A Layer-Wise Convex Pruning of Deep Neural Networks.” November 16, 2016.

http://arxiv.org/abs/1611.05162.

Borgerding, Mark, and Philip Schniter. 2016.

“Onsager-Corrected Deep Networks for Sparse Linear Inverse Problems.” December 4, 2016.

http://arxiv.org/abs/1612.01183.

Cai, Han, Chuang Gan, Tianzhe Wang, Zhekai Zhang, and Song Han. 2020.

“Once-for-All: Train One Network and Specialize It for Efficient Deployment.” In.

http://arxiv.org/abs/1908.09791.

Chen, Tianqi, Ian Goodfellow, and Jonathon Shlens. 2015.

“Net2Net: Accelerating Learning via Knowledge Transfer.” November 17, 2015.

http://arxiv.org/abs/1511.05641.

Chen, Wenlin, James T. Wilson, Stephen Tyree, Kilian Q. Weinberger, and Yixin Chen. 2015.

“Compressing Convolutional Neural Networks.” June 14, 2015.

http://arxiv.org/abs/1506.04449.

Cheng, Yu, Duo Wang, Pan Zhou, and Tao Zhang. 2017.

“A Survey of Model Compression and Acceleration for Deep Neural Networks.” October 23, 2017.

http://arxiv.org/abs/1710.09282.

Cutajar, Kurt, Edwin V. Bonilla, Pietro Michiardi, and Maurizio Filippone. 2017.

“Random Feature Expansions for Deep Gaussian Processes.” In

*PMLR*.

http://proceedings.mlr.press/v70/cutajar17a.html.

Daniely, Amit. 2017.

“Depth Separation for Neural Networks.” February 27, 2017.

http://arxiv.org/abs/1702.08489.

Frankle, Jonathan, and Michael Carbin. 2019.

“The Lottery Ticket Hypothesis: Finding Sparse, Trainable Neural Networks.” March 4, 2019.

http://arxiv.org/abs/1803.03635.

Garg, Sahil, Irina Rish, Guillermo Cecchi, and Aurelie Lozano. 2017.

“Neurogenesis-Inspired Dictionary Learning: Online Model Adaption in a Changing World.” In.

http://arxiv.org/abs/1701.06106.

Gelder, Maxwell Van, Mitchell Wortsman, and Kiana Ehsani. n.d. “Deconstructing the Structure of Sparse Neural Networks.” In, 6.

Ghosh, Tapabrata. 2017.

“QuickNet: Maximizing Efficiency and Efficacy in Deep Architectures.” January 9, 2017.

http://arxiv.org/abs/1701.02291.

Globerson, Amir, and Roi Livni. 2016.

“Learning Infinite-Layer Networks: Beyond the Kernel Trick.” June 16, 2016.

http://arxiv.org/abs/1606.05316.

Gray, Scott, Alec Radford, and Diederik P Kingma. n.d. “GPU Kernels for Block-Sparse Weights,” 12.

Ha, David, Andrew Dai, and Quoc V. Le. 2016.

“HyperNetworks.” September 27, 2016.

http://arxiv.org/abs/1609.09106.

Hardt, Moritz, Benjamin Recht, and Yoram Singer. 2015.

“Train Faster, Generalize Better: Stability of Stochastic Gradient Descent.” September 3, 2015.

http://arxiv.org/abs/1509.01240.

Hayou, Soufiane, Jean-Francois Ton, Arnaud Doucet, and Yee Whye Teh. 2020.

“Pruning Untrained Neural Networks: Principles and Analysis.” June 24, 2020.

http://arxiv.org/abs/2002.08797.

Hazimeh, Hussein, Natalia Ponomareva, Petros Mol, Zhenyu Tan, and Rahul Mazumder. 2020.

“The Tree Ensemble Layer: Differentiability Meets Conditional Computation,” February.

https://arxiv.org/abs/2002.07772v2.

He, Yihui, Ji Lin, Zhijian Liu, Hanrui Wang, Li-Jia Li, and Song Han. 2019.

“AMC: AutoML for Model Compression and Acceleration on Mobile Devices.” January 15, 2019.

http://arxiv.org/abs/1802.03494.

Howard, Andrew G., Menglong Zhu, Bo Chen, Dmitry Kalenichenko, Weijun Wang, Tobias Weyand, Marco Andreetto, and Hartwig Adam. 2017.

“MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications.” April 16, 2017.

http://arxiv.org/abs/1704.04861.

Iandola, Forrest N., Song Han, Matthew W. Moskewicz, Khalid Ashraf, William J. Dally, and Kurt Keutzer. 2016.

“SqueezeNet: AlexNet-Level Accuracy with 50x Fewer Parameters and \(<\)0.5MB Model Size.” February 23, 2016.

http://arxiv.org/abs/1602.07360.

Lee, Holden, Rong Ge, Tengyu Ma, Andrej Risteski, and Sanjeev Arora. 2017.

“On the Ability of Neural Nets to Express Distributions.” In.

http://arxiv.org/abs/1702.07028.

Lemhadri, Ismael, Feng Ruan, Louis Abraham, and Robert Tibshirani. 2021.

“LassoNet: A Neural Network with Feature Sparsity.” *Journal of Machine Learning Research* 22 (127): 1–29.

http://jmlr.org/papers/v22/20-848.html.

Lobacheva, Ekaterina, Nadezhda Chirkova, and Dmitry Vetrov. 2017.

“Bayesian Sparsification of Recurrent Neural Networks.” In

*Workshop on Learning to Generate Natural Language*.

http://arxiv.org/abs/1708.00077.

Louizos, Christos, Max Welling, and Diederik P. Kingma. 2017.

“Learning Sparse Neural Networks Through $L_0$ Regularization.” December 4, 2017.

http://arxiv.org/abs/1712.01312.

Molchanov, Dmitry, Arsenii Ashukha, and Dmitry Vetrov. 2017.

“Variational Dropout Sparsifies Deep Neural Networks.” In

*Proceedings of ICML*.

http://arxiv.org/abs/1701.05369.

Narang, Sharan, Eric Undersander, and Gregory Diamos. 2017.

“Block-Sparse Recurrent Neural Networks.” November 7, 2017.

http://arxiv.org/abs/1711.02782.

Pan, Wei, Hao Dong, and Yike Guo. 2016.

“DropNeuron: Simplifying the Structure of Deep Neural Networks.” June 23, 2016.

http://arxiv.org/abs/1606.07326.

Renda, Alex, Jonathan Frankle, and Michael Carbin. 2020.

“Comparing Rewinding and Fine-Tuning in Neural Network Pruning.” March 4, 2020.

http://arxiv.org/abs/2003.02389.

Scardapane, Simone, Danilo Comminiello, Amir Hussain, and Aurelio Uncini. 2016.

“Group Sparse Regularization for Deep Neural Networks.” July 2, 2016.

http://arxiv.org/abs/1607.00485.

Shi, Lei, Shikun Feng, and ZhifanZhu. 2016.

“Functional Hashing for Compressing Neural Networks.” May 20, 2016.

http://arxiv.org/abs/1605.06560.

Srinivas, Suraj, and R. Venkatesh Babu. 2016.

“Generalized Dropout.” November 21, 2016.

http://arxiv.org/abs/1611.06791.

Ullrich, Karen, Edward Meeds, and Max Welling. 2017.

“Soft Weight-Sharing for Neural Network Compression.” 2017.

https://arxiv.org/abs/1702.04008.

Urban, Gregor, Krzysztof J. Geras, Samira Ebrahimi Kahou, Ozlem Aslan, Shengjie Wang, Rich Caruana, Abdelrahman Mohamed, Matthai Philipose, and Matt Richardson. 2016.

“Do Deep Convolutional Nets Really Need to Be Deep (Or Even Convolutional)?” March 17, 2016.

http://arxiv.org/abs/1603.05691.

Wang, Yunhe, Chang Xu, Chao Xu, and Dacheng Tao. 2019.

“Packing Convolutional Neural Networks in the Frequency Domain.” *IEEE Transactions on Pattern Analysis and Machine Intelligence* 41 (10): 2495–2510.

https://doi.org/10.1109/TPAMI.2018.2857824.

Wang, Yunhe, Chang Xu, Shan You, Dacheng Tao, and Chao Xu. 2016.

“CNNpack: Packing Convolutional Neural Networks in the Frequency Domain.” In

*Advances in Neural Information Processing Systems 29*, edited by D. D. Lee, M. Sugiyama, U. V. Luxburg, I. Guyon, and R. Garnett, 253–61.

Curran Associates, Inc. http://papers.nips.cc/paper/6390-cnnpack-packing-convolutional-neural-networks-in-the-frequency-domain.pdf.

Wang, Zhangyang, Shiyu Chang, Qing Ling, Shuai Huang, Xia Hu, Honghui Shi, and Thomas S. Huang. 2016.

“Stacked Approximated Regression Machine: A Simple Deep Learning Approach.” In.

https://arxiv.org/abs/1608.04062.

Warden, Pete, and Daniel Situnayake. 2020.

*TinyML: Machine Learning with TensorFlow Lite on Arduino and Ultra-Low-Power Microcontrollers*.

O’Reilly Media, Incorporated.

http://books.google.com?id=sB3mxQEACAAJ.

You, Haoran, Chaojian Li, Pengfei Xu, Yonggan Fu, Yue Wang, Xiaohan Chen, Richard G. Baraniuk, Zhangyang Wang, and Yingyan Lin. 2019.

“Drawing Early-Bird Tickets: Toward More Efficient Training of Deep Networks.” In.

https://openreview.net/forum?id=BJxsrgStvr.

Zhao, Liang. 2017.

“Fast Algorithms on Random Matrices and Structured Matrices.” http://academicworks.cuny.edu/gc_etds/2073/.

## No comments yet. Why not leave one?