{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,31]],"date-time":"2025-12-31T12:07:41Z","timestamp":1767182861062,"version":"3.44.0"},"publisher-location":"New York, NY, USA","reference-count":28,"publisher":"ACM","license":[{"start":{"date-parts":[[2024,6,12]],"date-time":"2024-06-12T00:00:00Z","timestamp":1718150400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/linproxy.fan.workers.dev:443\/https\/creativecommons.org\/licenses\/by\/4.0\/"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2024,6,12]]},"DOI":"10.1145\/3649476.3660380","type":"proceedings-article","created":{"date-parts":[[2024,6,10]],"date-time":"2024-06-10T12:29:41Z","timestamp":1718022581000},"page":"620-625","update-policy":"https:\/\/linproxy.fan.workers.dev:443\/https\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":5,"title":["IRET: Incremental Resolution Enhancing Transformer"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/linproxy.fan.workers.dev:443\/https\/orcid.org\/0000-0003-3735-9191","authenticated-orcid":false,"given":"Banafsheh","family":"Saber Latibari","sequence":"first","affiliation":[{"name":"Electrical and Computer Engineering, University of California Davis, USA"}]},{"ORCID":"https:\/\/linproxy.fan.workers.dev:443\/https\/orcid.org\/0000-0001-5998-8795","authenticated-orcid":false,"given":"Soheil","family":"Salehi","sequence":"additional","affiliation":[{"name":"Electrical and Computer Engineering, University of Arizona, USA"}]},{"ORCID":"https:\/\/linproxy.fan.workers.dev:443\/https\/orcid.org\/0000-0001-8904-4699","authenticated-orcid":false,"given":"Houman","family":"Homayoun","sequence":"additional","affiliation":[{"name":"University of California Davis, USA"}]},{"ORCID":"https:\/\/linproxy.fan.workers.dev:443\/https\/orcid.org\/0000-0002-4052-8075","authenticated-orcid":false,"given":"Avesta","family":"Sasan","sequence":"additional","affiliation":[{"name":"University of California Davis, USA"}]}],"member":"320","published-online":{"date-parts":[[2024,6,12]]},"reference":[{"key":"e_1_3_2_1_1_1","volume-title":"Multi-exit vision transformer for dynamic inference. arXiv preprint arXiv:2106.15183","author":"Arian Bakhtiarnia","year":"2021","unstructured":"Arian Bakhtiarnia 2021. Multi-exit vision transformer for dynamic inference. arXiv preprint arXiv:2106.15183 (2021)."},{"key":"e_1_3_2_1_2_1","volume-title":"Proc., Part I 16","author":"Carion Nicolas","year":"2020","unstructured":"Nicolas Carion 2020. End-to-end object detection with transformers. In Computer Vision\u2013ECCV 2020: 16th European conf., Glasgow, UK, August 23\u201328, 2020, Proc., Part I 16. Springer, 213\u2013229."},{"key":"e_1_3_2_1_3_1","volume-title":"Rethinking attention with performers. arXiv preprint arXiv:2009.14794","author":"Krzysztof Choromanski","year":"2020","unstructured":"Krzysztof Choromanski 2020. Rethinking attention with performers. arXiv preprint arXiv:2009.14794 (2020)."},{"key":"e_1_3_2_1_4_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"e_1_3_2_1_5_1","unstructured":"Jacob Devlin 2019. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. arxiv:1810.04805\u00a0[cs.CL]"},{"key":"e_1_3_2_1_6_1","volume-title":"An image is worth 16x16 words: Transformers for image recognition at scale. arXiv preprint arXiv:2010.11929","author":"Alexey Dosovitskiy","year":"2020","unstructured":"Alexey Dosovitskiy 2020. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv preprint arXiv:2010.11929 (2020)."},{"key":"e_1_3_2_1_7_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.00675"},{"key":"e_1_3_2_1_8_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-20083-0_24"},{"key":"e_1_3_2_1_9_1","volume-title":"Magic pyramid: Accelerating inference with early exiting and token pruning. arXiv preprint arXiv:2111.00230","author":"Xuanli He","year":"2021","unstructured":"Xuanli He 2021. Magic pyramid: Accelerating inference with early exiting and token pruning. arXiv preprint arXiv:2111.00230 (2021)."},{"key":"e_1_3_2_1_10_1","volume-title":"Multi-scale dense networks for resource efficient image classification. arXiv preprint arXiv:1703.09844","author":"Gao Huang","year":"2017","unstructured":"Gao Huang 2017. Multi-scale dense networks for resource efficient image classification. arXiv preprint arXiv:1703.09844 (2017)."},{"key":"e_1_3_2_1_11_1","volume-title":"SimA: Simple Softmax-free Attention for Vision Transformers. arXiv preprint arXiv:2206.08898","author":"Soroush\u00a0Abbasi","year":"2022","unstructured":"Soroush\u00a0Abbasi Koohpayegani 2022. SimA: Simple Softmax-free Attention for Vision Transformers. arXiv preprint arXiv:2206.08898 (2022)."},{"key":"e_1_3_2_1_12_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.naacl-main.162"},{"key":"e_1_3_2_1_13_1","volume-title":"PatchDropout: Economizing Vision Transformers Using Patch Dropout. arXiv preprint arXiv:2208.07220","author":"Yue Liu","year":"2022","unstructured":"Yue Liu 2022. PatchDropout: Economizing Vision Transformers Using Patch Dropout. arXiv preprint arXiv:2208.07220 (2022)."},{"key":"e_1_3_2_1_14_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.00986"},{"key":"e_1_3_2_1_15_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.01199"},{"key":"e_1_3_2_1_16_1","volume-title":"cosFormer: Rethinking Softmax in Attention. arXiv preprint arXiv:2202.08791","author":"Zhen Qin","year":"2022","unstructured":"Zhen Qin 2022. cosFormer: Rethinking Softmax in Attention. arXiv preprint arXiv:2202.08791 (2022)."},{"key":"e_1_3_2_1_17_1","volume-title":"Dynamicvit: Efficient vision transformers with dynamic token sparsification. Advances in neural information processing systems 34","author":"Yongming Rao","year":"2021","unstructured":"Yongming Rao 2021. Dynamicvit: Efficient vision transformers with dynamic token sparsification. Advances in neural information processing systems 34 (2021), 13937\u201313949."},{"key":"e_1_3_2_1_18_1","doi-asserted-by":"publisher","DOI":"10.1109\/WACV45572.2020.9093580"},{"key":"e_1_3_2_1_19_1","doi-asserted-by":"publisher","DOI":"10.1109\/DAC18074.2021.9586134"},{"key":"e_1_3_2_1_20_1","volume-title":"Int. conf. on machine learning. PMLR, 10347\u201310357","author":"Touvron Hugo","year":"2021","unstructured":"Hugo Touvron 2021. Training data-efficient image transformers & distillation through attention. In Int. conf. on machine learning. PMLR, 10347\u201310357."},{"key":"e_1_3_2_1_21_1","volume-title":"Attention is all you need. Advances in neural information processing systems 30","author":"Ashish Vaswani","year":"2017","unstructured":"Ashish Vaswani 2017. Attention is all you need. Advances in neural information processing systems 30 (2017)."},{"key":"e_1_3_2_1_22_1","volume-title":"Efficient Softmax Approximation for Deep Neural Networks with Attention Mechanism. arXiv preprint arXiv:2111.10770","author":"Ihor Vasyltsov","year":"2021","unstructured":"Ihor Vasyltsov 2021. Efficient Softmax Approximation for Deep Neural Networks with Attention Mechanism. arXiv preprint arXiv:2111.10770 (2021)."},{"key":"e_1_3_2_1_23_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.00061"},{"key":"e_1_3_2_1_24_1","doi-asserted-by":"publisher","DOI":"10.1007\/s41095-022-0274-8"},{"key":"e_1_3_2_1_25_1","first-page":"11960","article-title":"Not all images are worth 16x16 words: Dynamic transformers for efficient image recognition","volume":"34","author":"Wang Yulin","year":"2021","unstructured":"Yulin Wang 2021. Not all images are worth 16x16 words: Dynamic transformers for efficient image recognition. Advances in Neural Information Processing Systems 34 (2021), 11960\u201311973.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_26_1","doi-asserted-by":"publisher","unstructured":"Ross Wightman. 2019. PyTorch Image Models. https:\/\/linproxy.fan.workers.dev:443\/https\/github.com\/rwightman\/pytorch-image-models. https:\/\/linproxy.fan.workers.dev:443\/https\/doi.org\/10.5281\/zenodo.4414861","DOI":"10.5281\/zenodo.4414861"},{"key":"e_1_3_2_1_27_1","volume-title":"Wave-ViT: Unifying Wavelet and Transformers for Visual Representation Learning. arXiv preprint arXiv:2207.04978","author":"Ting Yao","year":"2022","unstructured":"Ting Yao 2022. Wave-ViT: Unifying Wavelet and Transformers for Visual Representation Learning. arXiv preprint arXiv:2207.04978 (2022)."},{"key":"e_1_3_2_1_28_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.01054"}],"event":{"name":"GLSVLSI '24: Great Lakes Symposium on VLSI 2024","sponsor":["SIGDA ACM Special Interest Group on Design Automation"],"location":"Clearwater FL USA","acronym":"GLSVLSI '24"},"container-title":["Proceedings of the Great Lakes Symposium on VLSI 2024"],"original-title":[],"link":[{"URL":"https:\/\/linproxy.fan.workers.dev:443\/https\/dl.acm.org\/doi\/10.1145\/3649476.3660380","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/linproxy.fan.workers.dev:443\/https\/dl.acm.org\/doi\/pdf\/10.1145\/3649476.3660380","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,8,22]],"date-time":"2025-08-22T02:32:53Z","timestamp":1755829973000},"score":1,"resource":{"primary":{"URL":"https:\/\/linproxy.fan.workers.dev:443\/https\/dl.acm.org\/doi\/10.1145\/3649476.3660380"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,6,12]]},"references-count":28,"alternative-id":["10.1145\/3649476.3660380","10.1145\/3649476"],"URL":"https:\/\/linproxy.fan.workers.dev:443\/https\/doi.org\/10.1145\/3649476.3660380","relation":{},"subject":[],"published":{"date-parts":[[2024,6,12]]},"assertion":[{"value":"2024-06-12","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}