{“状态”:“确定”,“消息类型”:“工作”,“信息版本”:“1.0.0”,“邮件”:{“索引”:{“日期-部件”:[[2024,6,14]],“日期-时间”:“2024-06-14T07:13:17Z”,“时间戳”:1718349197903},“引用-计数”:37,“发布者”:“IEEE”,“许可证”:[{“开始”:{-“日期-零件”:[2020,6,1]],“时间”:”2020-06-01T00:00:00 Z“,”timestamp“:1590969600000},”content-version“:”vor“,“delay-in-days”:0,“URL”:“https:\/\/ieeexplore.iee.org\/Xplorehelp\/downloads\/license-information\/ieee.html”},{“start”:{“date-parts”:[[2020,6,1]],“date-time”:“2020-06-01T00:00:00Z”,“timestamp”:1590969600000},“content-version”:“stm-asf”,“delay-in-days“:0,”URL“https:\\/doi.org\/10.1 5223\/policy-029“},{“开始”:{“日期部分”:[[2020,6,1]],“date-time”:“2020-06-01T00:00:00Z”,“timestamp”:1590969600000},“content-version”:“stm-asf”,“delay-in-days”:0,“URL”:“https:\\/doi.org\/10.15223\/policy-037”}],“content-domain”:{“domain”:[],“crossmark-restriction”:false},”short-container-title“:[],”published-print“:{”date-parts“:[2020,6]},“doi”:“10.1109\/cvpr42600.2020.00198”,“类型”:“procesdings-article”,“created”:{“date-parts”:[[2020,8,5]],“date-time”:“2020-08-05T21:20:05Z”,“timestamp”:1596662405000},“source”:“Crossref”,“is-referenced-by-count”:17,“title”:[“结构化压缩重量加密用于非结构化删减和量化”],“prefix”:“10.1109”,”author“:[{”given“:”Se Jung“,“family”:“Kwon”,“sequence”:“first”,“affiliation”:[]},{“given”:“Dongsoo”,“family”:“Lee”,“sequence”:“additional”,“affaliation”(附属):[]}附加“,”从属“:[]},{“给定”:“Gu-Yeon”,“family”:“Wei”,“sequence”:“additional”,“affiliation”:[]}],“member”:“263”,“reference”:[{“key”:“ref33”,“author”:“ye”,“year”:“2018”,“journal-title”:“Rethinking the smaller-norm-less-informative assemption in channel prunning of convolutional layers”},{“密钥”:“ref32”,“article-title“递归神经网络的交替多位量化”,“author”:“xu”,“year”:“0”,“journal-title”:“International Conference on Learning Representations(ICLR)”},{“key”:“ref31”,“article-title“:“HitNet:Hybrid triumary Recurrive neural network”,“author”:“wang”,“年份”:“2018”,“日记标题”:“Advances in neural information processing systems”}、{“密钥”:“ref30”,“doi-asserted-by”:“publisher”:“10.1109\/MDT.2006.105”},{“key”:“ref37”,“article-title”:“修剪,还是不修剪:探索修剪对模型压缩的效果”,“volume”:”abs 1710 1878“,”author“:”zhu“,”year“:”2017“,”journal-title“:”CoRR“}:“国际学习表征会议(ICLR)”},{“key”:“ref35”,“doi-asserted-by”:“publisher”,“doi”:“10.1109\/MICRO.2018.0011”}、{“密钥”:“ref34”、“doi-sserted-by“:”publisher“,”doi“:”10.1145\/3140659.3080215“}:“深度压缩:通过剪枝、训练量化和哈夫曼编码压缩深度神经网络”,“author”:“han”,“year”:“0”,“journal-title”:“International Conference on Learning Representations(ICLR)”},{“key”:“ref11”,“first-page”:“1135”,“article-title(文章标题):“Learning both weights and connections for effective neural networks”,“作者”:“汉”,“年份”:“2015年”,“日志标题”:“神经信息处理系统的进展”},{“key”:“ref12”,“doi-asserted-by”:“publisher”,“doi”:“10.1109\/CVPR.2016.90”}:“量化神经网络训练具有低精度权重和激活的神经网络”},{“key”:“ref15”,“author”:“kapoor”,“year”:“2019”,“journal-title”:“deep neural networks的计算效率量化方法”}:“从微小图像中学习多层特征”},{“key”:“ref17”,“首页”:“1097”,“article-title”:“深度卷积神经网络的Imagenet分类”,“author”:“krizhevsky”,“year”:“2012”,“journal-title“:“Advances in neural Information Processing Systems 25”}:“最佳脑损伤”,“author”:“lecun”,“year”:“1990”,“journal-title”:“Advances in neural information processing systems”},{“key”:“ref19”,“article-title“:“Viterbi-based pruning for sparse matrix with fixed and high index compression ratio”,“作者”:“lee”,“年份”:“0”,“日记标题”:“International Conference on Learning Representations”(ICLR)},{“key”:“ref28”,“article-title”:“XNOR-Net:使用二进制卷积神经网络对Imagenet进行分类”,“author”:“rastegari”,“year”:“0”,“journal-title“:”ECCV“},{“key”:”ref4“,“first page”:“3123”,“article-tiple”:,“journal-title”:“神经信息处理系统的进展”},{“key”:“ref27”,“first page”:《2498》,“article-title》:“变分辍学稀疏化深层神经网络”,“author”:“molchanov”,“year”:“0”,“jornal-tittle”:”国际机器学习会议(ICML)“}”,{:“ref3”,“doi-asserted-by”:“publisher”,“doi”:“10.1145\/378239.378388”},{“key”:“ref6”,“author”:“frankle”,“year”:“2018”,“journal-title”:“彩票假设发现稀疏可训练神经网络”}:“denil”,“year”:“2013”,“journal-title”:“Advances in neural information processing systems”},{“key”:“ref8”,“article-title“:“Dynamic network surgery for efficient DNNs”,“author”:“guo”,“year”:”2016“,”journal-title“:”Advances on neural information processingsystems“},”{“key”:“ref7”,“author”:“goodfellow”,“年份”:“2016”,“journal-ttitle”:”:,{“key”:“ref2”,“doi-asserted-by”:“publisher”,“doi”:“10.1145\/3005348”},{“密钥”:“ref9”,“doi-asserte-by”:“publisher”,“DI:”10.1145\/3007787.3001163“},}“密钥“:”ref1“,”article-title“:”Double Viterbi:“重量编码,用于深层神经网络的高压缩比和快速片上重建”,“author”:“ahn”,“year”:“0”,“journal-title”:“国际学习表征会议(ICLR)”},{“key”:“ref20”,“author”:“lee”,“year”:“2018”,“journal-title”:“Deeptwist Learning model compression via percurrous weight distoration”}、{“key”:”ref22“,”author“:”lee“,”year“:”2019“,”journal-title“:”低秩二进制索引的网络剪枝,“journal-title”:“深度神经网络基于检索的迭代权重量化”},{“key”:“ref24”,“article-title“:“高效convents的剪枝过滤器”,“author”:“li”,“year”:“0”,“jornal-tittle”:《学习表征国际会议》},}“key:”ref23“author:”li“,”year“:”2016“,”journal-Tittle“:”三元权重网络“},{“key”:“ref26”,“doi-asserted-by”:“publisher”,“doi”:“10.3115\/1075812.1075835”},{“key”:《ref25》,“author”:“mao”,“year”:“2017”,“journal-title”:“Exploring the regulatory of sparse structure in convolutional neural networks”}],“event”:{“name”:“2020 IEEE\/CVF Conference on Computer Vision and Pattern Recognition(CVPR)”,“location”:“Seattle,WA,USA”,“start”:{“date-parts”:[[2020,6,13]]},“end”:{“date-parts”:[2020,6,19]]}},”container-title“:[”2020 IEEE \/CVF计算机视觉和模式识别会议(CVPR)“],“original-title”:[],“link”:[{“URL”:“http://\xplorestaging.IEEE.org\/ielx7\/9142308\/9156271\/091156820.pdf?arnumber=9156820“,”content-type“:”unspecified“,”content-version“:”vor“,“intended-application”:“相似性检查”}],“存放”:{“date-parts”:[[2022,6,27]],“date-time”:“2022-06-27T15:55:58Z”,“时间戳”:1656345358000},“分数”:1,“资源”:{primary“:{”URL“:”https:\/\/ieeexplore.iee.org\/document\/9156820\/“}”,“副标题”:[],“短标题”:[],“已发布”“:[[2020,6]]},”references-count“:37,”URL“:“http:\/\/dx.doi.org/10.1109\/cvpr42600.2020.00198”,“关系”:{},“主题”:[],“已发布”:{“日期部分”:[[2020,6]]}}}