中文版 | English
Title

FAT: Frequency-Aware Transformation for Bridging Full-Precision and Low-Precision Deep Representations

Author
Corresponding AuthorTao, Chaofan
Publication Years
2022-07-01
DOI
Source Title
ISSN
2162-237X
EISSN
2162-2388
VolumePPIssue:99Pages:1-15
Abstract
Learning low-bitwidth convolutional neural networks (CNNs) is challenging because performance may drop significantly after quantization. Prior arts often quantize the network weights by carefully tuning hyperparameters such as nonuniform stepsize and layerwise bitwidths, which are complicated since the full-and low-precision representations have large discrepancies. This work presents a novel quantization pipeline, named frequency-aware transformation (FAT), that features important benefits: 1) instead of designing complicated quantizers, FAT learns to transform network weights in the frequency domain to remove redundant information before quantization, making them amenable to training in low bitwidth with simple quantizers; 2) FAT readily embeds CNNs in low bitwidths using standard quantizers without tedious hyperparameter tuning and theoretical analyses show that FAT minimizes the quantization errors in both uniform and nonuniform quantizations; and 3) FAT can be easily plugged into various CNN architectures. Using FAT with a simple uniform/logarithmic quantizer can achieve the state-of-the-art performance in different bitwidths on various model architectures. Consequently, FAT serves to provide a novel frequency-based perspective for model quantization.
Keywords
URL[Source Record]
Indexed By
Language
English
SUSTech Authorship
Others
Funding Project
General Research Fund (GRF)["17206020","17209721"]
WOS Research Area
Computer Science ; Engineering
WOS Subject
Computer Science, Artificial Intelligence ; Computer Science, Hardware & Architecture ; Computer Science, Theory & Methods ; Engineering, Electrical & Electronic
WOS Accession No
WOS:000833054400001
Publisher
Data Source
Web of Science
PDF urlhttps://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=9837828
Citation statistics
Cited Times [WOS]:0
Document TypeJournal Article
Identifierhttp://kc.sustech.edu.cn/handle/2SGJ60CL/364975
DepartmentSUSTech Institute of Microelectronics
Affiliation
1.Univ Hong Kong, Dept Elect & Elect Engn, Hong Kong, Peoples R China
2.Southern Univ Sci & Technol, Sch Microelect, Shenzhen 518055, Peoples R China
3.Chinese Univ Hong Kong, Dept Elect Engn, Hong Kong, Peoples R China
4.Univ Hong Kong, Dept Comp Sci, Hong Kong, Peoples R China
Recommended Citation
GB/T 7714
Tao, Chaofan,Lin, Rui,Chen, Quan,et al. FAT: Frequency-Aware Transformation for Bridging Full-Precision and Low-Precision Deep Representations[J]. IEEE Transactions on Neural Networks and Learning Systems,2022,PP(99):1-15.
APA
Tao, Chaofan,Lin, Rui,Chen, Quan,Zhang, Zhaoyang,Luo, Ping,&Wong, Ngai.(2022).FAT: Frequency-Aware Transformation for Bridging Full-Precision and Low-Precision Deep Representations.IEEE Transactions on Neural Networks and Learning Systems,PP(99),1-15.
MLA
Tao, Chaofan,et al."FAT: Frequency-Aware Transformation for Bridging Full-Precision and Low-Precision Deep Representations".IEEE Transactions on Neural Networks and Learning Systems PP.99(2022):1-15.
Files in This Item:
There are no files associated with this item.
Related Services
Recommend this item
Bookmark
Usage statistics
Export to Endnote
Export to Excel
Export to Csv
Altmetrics Score
Google Scholar
Similar articles in Google Scholar
[Tao, Chaofan]'s Articles
[Lin, Rui]'s Articles
[Chen, Quan]'s Articles
Baidu Scholar
Similar articles in Baidu Scholar
[Tao, Chaofan]'s Articles
[Lin, Rui]'s Articles
[Chen, Quan]'s Articles
Bing Scholar
Similar articles in Bing Scholar
[Tao, Chaofan]'s Articles
[Lin, Rui]'s Articles
[Chen, Quan]'s Articles
Terms of Use
No data!
Social Bookmark/Share
No comment.

Items in the repository are protected by copyright, with all rights reserved, unless otherwise indicated.