A sample decreasing threshold greedy-based algorithm for big data summarisation

Cited 1 time in webofscience Cited 0 time in scopus
  • Hit : 17
  • Download : 0
As the scale of datasets used for big data applications expands rapidly, there have been increased efforts to develop faster algorithms. This paper addresses big data summarisation problems using the submodular maximisation approach and proposes an efficient algorithm for maximising general non-negative submodular objective functions subject to k-extendible system constraints. Leveraging a random sampling process and a decreasing threshold strategy, this work proposes an algorithm, named Sample Decreasing Threshold Greedy (SDTG). The proposed algorithm obtains an expected approximation guarantee of 1/1+k-epsilon for maximising monotone submodular functions and of k/(1+k)(2) - epsilon in non-monotone cases with expected computational complexity of O(n/(1+k)epsilon lnr/epsilon). Here, r is the largest size of feasible solutions, and epsilon is an element of(0, 1/1+k) is an adjustable designing parameter for the trade-off between the approximation ratio and the computational complexity. The performance of the proposed algorithm is validated and compared with that of benchmark algorithms through experiments with a movie recommendation system based on a real database.
Publisher
SPRINGERNATURE
Issue Date
2021-02
Language
English
Article Type
Article
Citation

JOURNAL OF BIG DATA, v.8, no.1

ISSN
2196-1115
DOI
10.1186/s40537-021-00416-y
URI
http://hdl.handle.net/10203/318575
Appears in Collection
GT-Journal Papers(저널논문)
Files in This Item
There are no files associated with this item.
This item is cited by other documents in WoS
⊙ Detail Information in WoSⓡ Click to see webofscience_button
⊙ Cited 1 items in WoS Click to see citing articles in records_button

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0