Fully Scalable Methods for Distributed Tensor Factorization

Cited 52 time in webofscience Cited 47 time in scopus
  • Hit : 480
  • Download : 0
DC FieldValueLanguage
dc.contributor.authorShin, Kijungko
dc.contributor.authorSael, Leeko
dc.contributor.authorKang, Uko
dc.date.accessioned2019-03-04T10:56:19Z-
dc.date.available2019-03-04T10:56:19Z-
dc.date.created2019-03-04-
dc.date.created2019-03-04-
dc.date.created2019-03-04-
dc.date.issued2017-01-
dc.identifier.citationIEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, v.29, no.1, pp.100 - 113-
dc.identifier.issn1041-4347-
dc.identifier.urihttp://hdl.handle.net/10203/250520-
dc.description.abstractGiven a high-order large-scale tensor, how can we decompose it into latent factors? Can we process it on commodity computers with limited memory? These questions are closely related to recommender systems, which have modeled rating data not as a matrix but as a tensor to utilize contextual information such as time and location. This increase in the order requires tensor factorization methods scalable with both the order and size of a tensor. In this paper, we propose two distributed tensor factorization methods, CDTF and SALS. Both methods are scalable with all aspects of data and show a trade-off between convergence speed and memory requirements. CDTF, based on coordinate descent, updates one parameter at a time, while SALS generalizes on the number of parameters updated at a time. In our experiments, only our methods factorized a five-order tensor with 1 billion observable entries, 10M mode length, and 1 K rank, while all other state-of-the-art methods failed. Moreover, our methods required several orders of magnitude less memory than their competitors. We implemented our methods on MAPREDUCE with two widely-applicable optimization techniques: local disk caching and greedy row assignment. They speeded up our methods up to 98.2 x and also the competitors up to 5.9 x.-
dc.languageEnglish-
dc.publisherIEEE COMPUTER SOC-
dc.titleFully Scalable Methods for Distributed Tensor Factorization-
dc.typeArticle-
dc.identifier.wosid000390556100008-
dc.identifier.scopusid2-s2.0-85020124755-
dc.type.rimsART-
dc.citation.volume29-
dc.citation.issue1-
dc.citation.beginningpage100-
dc.citation.endingpage113-
dc.citation.publicationnameIEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING-
dc.identifier.doi10.1109/TKDE.2016.2610420-
dc.contributor.localauthorShin, Kijung-
dc.contributor.nonIdAuthorSael, Lee-
dc.contributor.nonIdAuthorKang, U-
dc.description.isOpenAccessN-
dc.type.journalArticleArticle-
dc.subject.keywordAuthorTensor factorization-
dc.subject.keywordAuthortensor completion-
dc.subject.keywordAuthordistributed computing-
dc.subject.keywordAuthorMapReduce-
dc.subject.keywordAuthorhadoop-
dc.subject.keywordPlusMATRIX FACTORIZATION-
dc.subject.keywordPlusRECOMMENDER SYSTEMS-
dc.subject.keywordPlusDECOMPOSITIONS-
dc.subject.keywordPlusPARAFAC-
Appears in Collection
AI-Journal Papers(저널논문)
Files in This Item
There are no files associated with this item.
This item is cited by other documents in WoS
⊙ Detail Information in WoSⓡ Click to see webofscience_button
⊙ Cited 52 items in WoS Click to see citing articles in records_button

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0