Energy-efficient acceleration of DNN-based image recognition by leveraging task- and token-level redundancies작업 및 토큰 수준의 중복성을 활용한 DNN 기반 이미지 인식의 에너지 효율적 가속

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 24
  • Download : 0
The advancement of Deep Neural Networks (DNNs) has significantly transformed our daily lives through various Computer Vision (CV) applications. Tasks such as image classification, object recognition, and motion detection, previously handled by distinct algorithms, have been integrated into DNN-based algorithms, exhibiting superior performance. Furthermore, the development of DNN architectures, including the Vision Transformer (ViT), has continuously improved the performance of computer vision applications. However, the substantial computational requirements and data volumes necessary for DNN-based image processing pose challenges to the commercialization of such applications. Especially for real-time interactive computer vision applications, which are primarily processed on resource-constrained edge devices like mobile and IoT devices, the computational demands and data sizes of DNNs become significant obstacles. For instance, energy consumption to load data from memory shortens device battery life, and the execution time of applications is extended due to numerous multiplication operations on limited computational resources. Therefore, this thesis proposes a hardware-algorithm co-optimization technique to reduce energy consumption and execution time required for DNN-based computer vision applications. Firstly, for energy reduction, the goal is to share data, namely model weights and feature values, among multiple computer vision tasks. To achieve weight sharing between tasks, a transfer learning technique is introduced, which avoids altering the backbone network's weights during model training for a specific task. Additionally, a feature value sharing technique, utilizing image characteristics, reduces memory requirements for storing feature values. To maximize the benefits of these algorithm techniques, a hardware architecture for weight and feature value processing per task, along with a data flow that enables data sharing between tasks, is proposed, resulting in significant energy savings. Secondly, to reduce execution time required for computer vision applications, methods for reducing the number of feature values are presented. A Token Merging technique, adapted and optimized for computer vision applications, is used to address the limitations of Token Pruning, commonly used in transformer model compression. Furthermore, a hardware architecture is proposed to efficiently process the light-weight Vision Transformer. Unit designs for Token Merging, along with a new pipeline architecture to minimize associated overheads, significantly reduce the overall execution time of deep neural network models. These strategies collectively contribute to mitigating the challenges posed by the computational demands and data sizes of DNNs, making them more practical and efficient for various computer vision applications.
Advisors
김이섭researcher
Description
한국과학기술원 :전기및전자공학부,
Publisher
한국과학기술원
Issue Date
2024
Identifier
325007
Language
eng
Description

학위논문(박사) - 한국과학기술원 : 전기및전자공학부, 2024.2,[vi, 58 p. :]

Keywords

딥 뉴럴 네트워크 가속기▼a알고리즘-하드웨어 공동 최적화▼a컴퓨터 비전▼a다중 작업 처리▼a비전 트랜스포머; DNN accelerator▼aAlgorithm-hardware co-optimization▼aComputer vision▼aMulti-task prcoessing▼aVision transformer

URI
http://hdl.handle.net/10203/322155
Link
http://library.kaist.ac.kr/search/detail/view.do?bibCtrlNo=1100055&flag=dissertation
Appears in Collection
EE-Theses_Ph.D.(박사논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0