A GPU-based tensor decomposition method for large-scale tensors

Cited 1 time in webofscience Cited 0 time in scopus
  • Hit : 73
  • Download : 0
Recently, as the sizes of real tensors have become overwhelmingly large including billions of nonzeros, fast and scalable Tucker decomposition methods have become increasingly important. Tucker decomposition has been widely used to analyze multidimensional data modeled as tensors. Several GPU-based Tucker decomposition methods have been proposed to enhance the decomposition speed. However, they easily fail to process large-scale tensors owing to the high memory requirements, which are larger than the GPU memory. This paper presents a scalable GPU-based Tucker decomposition method called GTucker, which carefully partitions large-scale tensors into subtensors and processes them with reduced overhead on a single machine. The results of the experiments indicate that GTucker outperforms state-of-the-art methods in terms of scalability and decomposition speed.
Publisher
IEEE
Issue Date
2023-02
Language
English
Citation

2023 IEEE International Conference on Big Data and Smart Computing (BigComp), pp.77 - 80

DOI
10.1109/bigcomp57234.2023.00020
URI
http://hdl.handle.net/10203/310987
Appears in Collection
CS-Conference Papers(학술회의논문)
Files in This Item
There are no files associated with this item.
This item is cited by other documents in WoS
⊙ Detail Information in WoSⓡ Click to see webofscience_button
⊙ Cited 1 items in WoS Click to see citing articles in records_button

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0