Learning to Quantize Deep Networks by Optimizing Quantization Intervals with Task Loss

Cited 165 time in webofscience Cited 0 time in scopus
  • Hit : 761
  • Download : 0
DC FieldValueLanguage
dc.contributor.authorJung Sang Gilko
dc.contributor.authorSon, Cahng Tongko
dc.contributor.authorLee, Seo Hyungko
dc.contributor.authorSon, Jin Wooko
dc.contributor.authorKwak, Young Junko
dc.contributor.authorHan, Jae Joonko
dc.contributor.authorHwang, Sung Juko
dc.contributor.authorChoi, Chang Gyuko
dc.date.accessioned2020-02-10T07:21:19Z-
dc.date.available2020-02-10T07:21:19Z-
dc.date.created2019-12-03-
dc.date.created2019-12-03-
dc.date.created2019-12-03-
dc.date.issued2019-06-16-
dc.identifier.citation32nd IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp.1 - 10-
dc.identifier.issn1063-6919-
dc.identifier.urihttp://hdl.handle.net/10203/272204-
dc.description.abstractReducing bit-widths of activations and weights of deep networks makes it efficient to compute and store them in memory, which is crucial in their deployments to resource-limited devices, such as mobile phones. However, decreasing bit-widths with quantization generally yields drastically degraded accuracy. To tackle this problem, we propose to learn to quantize activations and weights via a trainable quantizer that transforms and discretizes them. Specifically, we parameterize the quantization intervals and obtain their optimal values by directly minimizing the task loss of the network. This quantization-interval-learning (QIL) allows the quantized networks to maintain the accuracy of the full-precision (32-bit) networks with bit-width as low as 4-bit and minimize the accuracy degeneration with further bit-width reduction (i.e., 3 and 2-bit). Moreover, our quantizer can be trained on a heterogeneous dataset, and thus can be used to quantize pretrained networks without access to their training data. We demonstrate the effectiveness of our trainable quantizer on ImageNet dataset with various network architectures such as ResNet-18, -34 and AlexNet, on which it outperforms existing methods to achieve the state-of-the-art accuracy.-
dc.languageEnglish-
dc.publisherConference on Computer Vision and Pattern Recognition(CVPR)-
dc.titleLearning to Quantize Deep Networks by Optimizing Quantization Intervals with Task Loss-
dc.typeConference-
dc.identifier.wosid000529484004054-
dc.identifier.scopusid2-s2.0-85076874705-
dc.type.rimsCONF-
dc.citation.beginningpage1-
dc.citation.endingpage10-
dc.citation.publicationname32nd IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)-
dc.identifier.conferencecountryUS-
dc.identifier.conferencelocationLong Beach Convention Center-
dc.identifier.doi10.1109/CVPR.2019.00448-
dc.contributor.localauthorHwang, Sung Ju-
dc.contributor.nonIdAuthorJung Sang Gil-
dc.contributor.nonIdAuthorSon, Cahng Tong-
dc.contributor.nonIdAuthorLee, Seo Hyung-
dc.contributor.nonIdAuthorSon, Jin Woo-
dc.contributor.nonIdAuthorKwak, Young Jun-
dc.contributor.nonIdAuthorHan, Jae Joon-
dc.contributor.nonIdAuthorChoi, Chang Gyu-
Appears in Collection
AI-Conference Papers(학술대회논문)
Files in This Item
There are no files associated with this item.
This item is cited by other documents in WoS
⊙ Detail Information in WoSⓡ Click to see webofscience_button
⊙ Cited 165 items in WoS Click to see citing articles in records_button

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0