Residential Collegefalse
Status已發表Published
Multi-Task Learning with Personalized Transformer for Review Recommendation
Haiming Wang1,2; Wei Liu1,2,3; Jian Yin1,2
2021-10
Conference Name22nd International Conference on Web Information Systems Engineering, WISE 2021
Source PublicationLecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
Volume13081 LNCS
Pages162-176
Conference Date26-29 October 2021
Conference PlaceMelbourne, Australia
CountryAustralia
PublisherSpringer, Cham
Abstract

Drastic increase in item/product review volume has caused serious information overload. Traditionally, product reviews are exhibited in chronological or popularity order without personalization. The review recommendation model provides users with attractive reviews and efficient consumption experience, allowing users to grasp the characteristics of items in seconds. However, the sparsity of interactions between users and reviews appears to be a major challenge. And the multi-relationship context, especially potential semantic feature in reviews, is not fully exploited. To address these problems, Multi-Task Learning model incorporating Personalized Transformer (MTL-PT) is proposed to provide users with an interesting review list. It contains three tasks: the main task models user’s preference to reviews with the proposed poly aggregator, incorporating the user-item-aware semantic feature. Two auxiliary tasks model the quality of reviews, and user-item interactions, respectively. These tasks collaboratively learn the multi-relationship among user, item, and review. The shared latent features of user/item link the three tasks together. Especially, the personalized semantic features of reviews are also fused into the tasks with the proposed personalized transformer. Two new real-world datasets for personalized review recommendation are collected and constructed. Extensive experiments are conducted on them. Compared with the state-of-the-arts, the results validate the effectiveness of our model for review recommendation.

KeywordMulti-task Learning Personalized Transformer Review Recommendation
DOI10.1007/978-3-030-91560-5_12
URLView the original
Indexed ByCPCI-S
Language英語English
WOS Research AreaComputer Science
WOS SubjectComputer Science, Artificial Intelligence ; Computer Science, Information Systems ; Computer Science, Theory & Methods
WOS IDWOS:000897917000012
Scopus ID2-s2.0-85121904814
Fulltext Access
Citation statistics
Document TypeConference paper
CollectionTHE STATE KEY LABORATORY OF INTERNET OF THINGS FOR SMART CITY (UNIVERSITY OF MACAU)
Corresponding AuthorWei Liu
Affiliation1.School of Artificial Intelligence, Sun Yat-sen University, Guangzhou, China
2.Guangdong Key Laboratory of Big Data Analysis and Processing, Guangzhou, China
3.SKL of Internet of Things for Smart City, University of Macau, Macau, China
Corresponding Author AffilicationUniversity of Macau
Recommended Citation
GB/T 7714
Haiming Wang,Wei Liu,Jian Yin. Multi-Task Learning with Personalized Transformer for Review Recommendation[C]:Springer, Cham, 2021, 162-176.
APA Haiming Wang., Wei Liu., & Jian Yin (2021). Multi-Task Learning with Personalized Transformer for Review Recommendation. Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), 13081 LNCS, 162-176.
Files in This Item:
There are no files associated with this item.
Related Services
Recommend this item
Bookmark
Usage statistics
Export to Endnote
Google Scholar
Similar articles in Google Scholar
[Haiming Wang]'s Articles
[Wei Liu]'s Articles
[Jian Yin]'s Articles
Baidu academic
Similar articles in Baidu academic
[Haiming Wang]'s Articles
[Wei Liu]'s Articles
[Jian Yin]'s Articles
Bing Scholar
Similar articles in Bing Scholar
[Haiming Wang]'s Articles
[Wei Liu]'s Articles
[Jian Yin]'s Articles
Terms of Use
No data!
Social Bookmark/Share
All comments (0)
No comment.
 

Items in the repository are protected by copyright, with all rights reserved, unless otherwise indicated.