Towards Validating Long-Term User Feedbacks in Interactive Recommendation Systems

Cited 2 time in webofscience Cited 0 time in scopus
  • Hit : 89
  • Download : 0
DC FieldValueLanguage
dc.contributor.authorLee, Hojoonko
dc.contributor.authorHwang, Dongyoonko
dc.contributor.authorMin, Kyushikko
dc.contributor.authorChoo, Jaegulko
dc.date.accessioned2023-09-19T11:01:04Z-
dc.date.available2023-09-19T11:01:04Z-
dc.date.created2023-09-19-
dc.date.issued2022-07-
dc.identifier.citation45th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, SIGIR 2022, pp.2607 - 2611-
dc.identifier.urihttp://hdl.handle.net/10203/312773-
dc.description.abstractInteractive Recommender Systems (IRSs) have attracted a lot of attention, due to their ability to model interactive processes between users and recommender systems. Numerous approaches have adopted Reinforcement Learning (RL) algorithms, as these can directly maximize users' cumulative rewards. In IRS, researchers commonly utilize publicly available review datasets to compare and evaluate algorithms. However, user feedback provided in public datasets merely includes instant responses (e.g., a rating), with no inclusion of delayed responses (e.g., the dwell time and the lifetime value). Thus, the question remains whether these review datasets are an appropriate choice to evaluate the long-term effects in IRS. In this work, we revisited experiments on IRS with review datasets and compared RL-based models with a simple reward model that greedily recommends the item with the highest one-step reward. Following extensive analysis, we can reveal three main findings: First, a simple greedy reward model consistently outperforms RL-based models in maximizing cumulative rewards. Second, applying higher weighting to long-term rewards leads to degradation of recommendation performance. Third, user feedbacks have mere long-term effects in the benchmark datasets. Based on our findings, we conclude that a dataset has to be carefully verified and that a simple greedy baseline should be included for a proper evaluation of RL-based IRS approaches. Our code and dataset are available at https: //github.com/dojeon-ai/irs_validation.-
dc.languageEnglish-
dc.publisherAssociation for Computing Machinery, Inc-
dc.titleTowards Validating Long-Term User Feedbacks in Interactive Recommendation Systems-
dc.typeConference-
dc.identifier.wosid000852715902117-
dc.identifier.scopusid2-s2.0-85135025185-
dc.type.rimsCONF-
dc.citation.beginningpage2607-
dc.citation.endingpage2611-
dc.citation.publicationname45th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, SIGIR 2022-
dc.identifier.conferencecountrySP-
dc.identifier.conferencelocationMadrid-
dc.identifier.doi10.1145/3477495.3531869-
dc.contributor.localauthorChoo, Jaegul-
dc.contributor.nonIdAuthorMin, Kyushik-
Appears in Collection
AI-Conference Papers(학술대회논문)
Files in This Item
There are no files associated with this item.
This item is cited by other documents in WoS
⊙ Detail Information in WoSⓡ Click to see webofscience_button
⊙ Cited 2 items in WoS Click to see citing articles in records_button

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0