{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,6]],"date-time":"2026-03-06T18:58:01Z","timestamp":1772823481728,"version":"3.50.1"},"reference-count":45,"publisher":"IEEE","license":[{"start":{"date-parts":[[2023,10,1]],"date-time":"2023-10-01T00:00:00Z","timestamp":1696118400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/linproxy.fan.workers.dev:443\/https\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2023,10,1]],"date-time":"2023-10-01T00:00:00Z","timestamp":1696118400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/linproxy.fan.workers.dev:443\/https\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2023,10,1]]},"DOI":"10.1109\/iccv51070.2023.01600","type":"proceedings-article","created":{"date-parts":[[2024,1,15]],"date-time":"2024-01-15T20:55:59Z","timestamp":1705352159000},"page":"17401-17411","source":"Crossref","is-referenced-by-count":25,"title":["Efficient Joint Optimization of Layer-Adaptive Weight Pruning in Deep Neural Networks"],"prefix":"10.1109","author":[{"given":"Kaixin","family":"Xu","sequence":"first","affiliation":[{"name":"Technology and Research (A*STAR),Institute for Infocomm Research (I2R), Agency for Science,Singapore,138632"}]},{"given":"Zhe","family":"Wang","sequence":"additional","affiliation":[{"name":"Technology and Research (A*STAR),Institute for Infocomm Research (I2R), Agency for Science,Singapore,138632"}]},{"given":"Xue","family":"Geng","sequence":"additional","affiliation":[{"name":"Technology and Research (A*STAR),Institute for Infocomm Research (I2R), Agency for Science,Singapore,138632"}]},{"given":"Min","family":"Wu","sequence":"additional","affiliation":[{"name":"Technology and Research (A*STAR),Institute for Infocomm Research (I2R), Agency for Science,Singapore,138632"}]},{"given":"Xiaoli","family":"Li","sequence":"additional","affiliation":[{"name":"Technology and Research (A*STAR),Institute for Infocomm Research (I2R), Agency for Science,Singapore,138632"}]},{"given":"Weisi","family":"Lin","sequence":"additional","affiliation":[{"name":"Nanyang Technological University,Singapore"}]}],"member":"263","reference":[{"key":"ref1","article-title":"Net-trim: Convex pruning of deep neural networks with performance guarantee","volume":"30","author":"Aghasi","year":"2017","journal-title":"Advances in neural information processing systems"},{"key":"ref2","article-title":"Post training 4-bit quantization of convolutional networks for rapid-deployment","volume":"32","author":"Banner","year":"2019","journal-title":"Advances in Neural Information Processing Systems"},{"key":"ref3","article-title":"Deep rewiring: Training very sparse deep networks","volume-title":"International Conference on Learning Representations","author":"Bellec"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v36i1.19888"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00890"},{"key":"ref6","first-page":"19637","article-title":"Only train once: A one-shot neural network training and pruning framework","volume":"34","author":"Chen","year":"2021","journal-title":"Advances in Neural Information Processing Systems"},{"key":"ref7","article-title":"Progressive skeletonization: Trimming more fat from a network at initialization","volume-title":"International Conference on Learning Representations","author":"de Jorge"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"ref9","article-title":"Learning to prune deep neural networks via layer-wise optimal brain surgeon","volume":"30","author":"Dong","year":"2017","journal-title":"Advances in Neural Information Processing Systems"},{"key":"ref10","first-page":"2943","article-title":"Rigging the lottery: Making all tickets winners","volume-title":"International Conference on Machine Learning","author":"Evci"},{"key":"ref11","article-title":"Fighting quantization bias with bias","author":"Finkelstein","year":"2019"},{"key":"ref12","article-title":"The lottery ticket hypothesis: Finding sparse, trainable neural networks","volume-title":"International Conference on Learning Representations","author":"Frankle"},{"key":"ref13","article-title":"The state of sparsity in deep neural networks","author":"Gale","year":"2019"},{"key":"ref14","article-title":"Dynamic network surgery for efficient dnns","volume":"29","author":"Guo","year":"2016","journal-title":"Advances in neural information processing systems"},{"key":"ref15","article-title":"Deep compression: Compressing deep neural networks with pruning, trained quantization and huffman coding","volume-title":"ICLR","author":"Han"},{"key":"ref16","article-title":"Learning both weights and connections for efficient neural network","volume":"28","author":"Han","year":"2015","journal-title":"Advances in neural information processing systems"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.90"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.155"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.243"},{"key":"ref20","first-page":"3821","article-title":"An information-theoretic justification for model pruning","volume-title":"International Conference on Artificial Intelligence and Statistics","author":"Isik"},{"key":"ref21","author":"Krizhevsky","year":"2009","journal-title":"Learning multiple layers of features from tiny images"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1145\/3065386"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1109\/ICCVW54120.2021.00094"},{"key":"ref24","article-title":"Optimal brain damage","volume":"2","author":"LeCun","year":"1989","journal-title":"Advances in neural information processing systems"},{"key":"ref25","article-title":"Layer-adaptive sparsity for the magnitude-based pruning","volume-title":"International Conference on Learning Representations","author":"Lee"},{"key":"ref26","article-title":"Snip: Single-shot network pruning based on connection sensitivity","volume-title":"International Conference on Learning Representations","author":"Lee"},{"key":"ref27","article-title":"Pruning filters for efficient convnets","author":"Li","year":"2016"},{"key":"ref28","article-title":"Runtime neural pruning","volume":"30","author":"Lin","year":"2017","journal-title":"Advances in neural information processing systems"},{"key":"ref29","article-title":"Dynamic model pruning with feedback","volume-title":"International Conference on Learning Representations","author":"Lin"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.298"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.541"},{"key":"ref32","article-title":"Pruning convolutional neural networks for resource efficient inference","author":"Molchanov","year":"2016"},{"key":"ref33","article-title":"One ticket to win them all: generalizing lottery ticket initializations across datasets and optimizers","volume":"32","author":"Morcos","year":"2019","journal-title":"Advances in neural information processing systems"},{"key":"ref34","article-title":"Very deep convolutional networks for large-scale image recognition","volume-title":"International Conference on Learning Representations","author":"Simonyan"},{"key":"ref35","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2015.7298594"},{"key":"ref36","article-title":"Picking winning tickets before training by preserving gradient flow","volume-title":"International Conference on Learning Representations","author":"Wang"},{"key":"ref37","doi-asserted-by":"publisher","DOI":"10.24963\/ijcai.2022\/786"},{"key":"ref38","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-19775-8_10"},{"key":"ref39","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.643"},{"key":"ref40","article-title":"Optg: Optimizing gradient-driven criteria in network sparsity","author":"Zhang","year":"2022"},{"key":"ref41","doi-asserted-by":"publisher","DOI":"10.1109\/DCC50243.2021.00033"},{"key":"ref42","doi-asserted-by":"publisher","DOI":"10.1109\/ICIP.2019.8803498"},{"key":"ref43","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.00360"},{"key":"ref44","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v32i1.11623"},{"key":"ref45","author":"Zhu","year":"2018","journal-title":"To prune, or not to prune: Exploring the efficacy of pruning for model compression"}],"event":{"name":"2023 IEEE\/CVF International Conference on Computer Vision (ICCV)","location":"Paris, France","start":{"date-parts":[[2023,10,1]]},"end":{"date-parts":[[2023,10,6]]}},"container-title":["2023 IEEE\/CVF International Conference on Computer Vision (ICCV)"],"original-title":[],"link":[{"URL":"https:\/\/linproxy.fan.workers.dev:443\/http\/xplorestaging.ieee.org\/ielx7\/10376473\/10376477\/10376733.pdf?arnumber=10376733","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,1,18]],"date-time":"2024-01-18T01:02:00Z","timestamp":1705539720000},"score":1,"resource":{"primary":{"URL":"https:\/\/linproxy.fan.workers.dev:443\/https\/ieeexplore.ieee.org\/document\/10376733\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,10,1]]},"references-count":45,"URL":"https:\/\/linproxy.fan.workers.dev:443\/https\/doi.org\/10.1109\/iccv51070.2023.01600","relation":{},"subject":[],"published":{"date-parts":[[2023,10,1]]}}}