Memory access reduction techniques for CNN accelerating systemsCNN 가속 시스템을 위한 메모리 접근 감소 기법

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 239
  • Download : 0
DC FieldValueLanguage
dc.contributor.advisorPark, In-Cheol-
dc.contributor.advisor박인철-
dc.contributor.authorKim, Suchang-
dc.date.accessioned2023-06-23T19:33:48Z-
dc.date.available2023-06-23T19:33:48Z-
dc.date.issued2023-
dc.identifier.urihttp://library.kaist.ac.kr/search/detail/view.do?bibCtrlNo=1030535&flag=dissertationen_US
dc.identifier.urihttp://hdl.handle.net/10203/309116-
dc.description학위논문(박사) - 한국과학기술원 : 전기및전자공학부, 2023.2,[v, 76 p. :]-
dc.description.abstractConvolutional neural networks (CNNs) have been actively applied to computer vision applications using deep-layered architecture in order to achieve high accuracy. However, the deep-layered architecture increases energy consumption required in implementing CNN accelerating systems, so reducing memory access that consumes large energy is important to achieve high energy efficiency. The computer vision applications can be classified into image-level labeling application, such as image classification and object detection, and pixel-level labeling application, such as super-resolution and image-to-image translation. This dissertation proposes memory reduction techniques and hardware architecture for both applications by analyzing unique characteristics of each application. In the image-level labeling application, an input image is processed by a feature extraction network to generate inferences for objects in the image. In the feature extraction network, the number of features involved in the convolution of a shallow layer is larger than that of kernels. Due to the feature extraction network, however, the number of features decreases while that of kernels increases as the layer deepens. By taking into account the numbers of data, a hybrid convolution technique, which selects either kernel-stay convolution or feature-stay convolution, is proposed to reduce memory access. In the pixel-level labeling application, on the other hand, an input image is translated to another image of the same or higher resolution by using a encoder-decoder network that infers every pixel. In the encoder-decoder network, the number of features is maintained even if the layer deepens. As high-resolution images are becoming mainstream, in addition, the number of features generated is very large. To reduce memory access, CNN compression and layer-chaining convolution techniques are proposed. Realizing the proposed techniques, a neural processing unit that accelerates CNNs is designed for each application, and CNN accelerating systems have been implemented.-
dc.languageeng-
dc.publisher한국과학기술원-
dc.subjectComputer vision▼aConvolutional neural networks▼aEnergy efficiency▼aNeural processing units▼aConvolutional neural network accelerating systems-
dc.subject컴퓨터 비전▼a컨볼루션 신경망▼a에너지 효율▼a신경망 처리장치▼a컨볼루션 신경망 가속 시스템-
dc.titleMemory access reduction techniques for CNN accelerating systems-
dc.title.alternativeCNN 가속 시스템을 위한 메모리 접근 감소 기법-
dc.typeThesis(Ph.D)-
dc.identifier.CNRN325007-
dc.description.department한국과학기술원 :전기및전자공학부,-
dc.contributor.alternativeauthor김수창-
Appears in Collection
EE-Theses_Ph.D.(박사논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0