Transcription

Fundamentals of DeepRecommender SystemsWenqi FanThe Hong Kong Polytechnic Universityhttps://wenqifan03.github.io, [email protected] website: https://deeprs-tutorial.github.ioData Science and Engineering Lab1

A General Architecture of Deep Recommender SystemPrediction layerHidden layer(e.g., MLP, CNN, RNN, etc. )Embedding layer0011Field 1User00Field mItemContext010Field MInteraction2

NeuMFNeuMF unifies the strengths of MF and MLP in modeling user-item interactions.p MF uses an inner product as the interaction functionp MLP is more sufficient to capture the complex structure of user interaction dataNeural Collaborative Filtering, WWW, 20173

Wide&DeepStandard MLPEmbeddingConcatenationp The wide linear models can memorize seen feature interactions using cross-product featuretransformations.p The deep models can generalize to previously unseen feature interactions through low- dimensionalembeddings.Wide & Deep Learning for Recommender Systems, 1st DLRS, 20164

Neural FMNeural Factorization Machines (NFMs) “deepens”FM by placing hidden layers above second-orderfeature interaction modeling.categorical variablesNeural Factorization Machines for Sparse Predictive Analytics, SIGIR, 20175

Neural FMNeural Factorization Machines (NFMs) “deepens”FM by placing hidden layers above second-orderfeature interaction modeling.“Deep layers” learn higher-order featureinteractions only, being much easier to train.Bilinear Interaction Pooling:''𝑓!" 𝑉# 𝑥 𝐯 𝑥( v(categorical variables %& (% )&Neural Factorization Machines for Sparse Predictive Analytics, SIGIR, 20176

DeepFMDeepFM ensembles FM and DNN and to low- and high-order feature interactionssimultaneously from the input raw features.FM component (low-order)Deep component (high-order)Prediction Model:DeepFM: A Factorization-Machine based Neural Network for CTR Prediction, IJCAI, 20177

DeepFMDeepFM ensembles FM and DNN and to low- and high-order feature interactionssimultaneously from the input raw features.FM component (low-order)Deep component (high-order)Share theembedding layerPrediction Model:DeepFM: A Factorization-Machine based Neural Network for CTR Prediction, IJCAI, 20178

DSCFCollaborative Filtering with users’ social relations(Social Recommendation)Dr. Anthony FauciNewsUS could see millions of coronaviruscases and 100,000 or more deathsDeep Social Collaborative Filtering, RecSys, 20199

DSCFCollaborative Filtering with users’ social relations(Social Recommendation)Users might be affected by direct/distant neighbors.Ø Information diffusionØ Users with high reputationsr'User EmbeddingRating Embedding Item EmbeddingConcatenationOutput Layer Concatenationβ1β2β3Attention Networkβ4.α1SequenceLearning Layerα2α4Attention Networkα3Dr. Anthony FaucihNews1h1h2h2US could see millions of coronaviruscases and 100,000 or more deathsh.hL -1L -1hLhL.Embedding Layer u[1]u[2]Random Walk LayerDeep Social Collaborative Filtering, RecSys, 2019. .u[l]10

DSCFCollaborative Filtering with users’ social relations(Social Recommendation)Users might be affected by direct/distant neighbors.Ø Information diffusionØ Users with high reputationsr'User EmbeddingRating Embedding Item EmbeddingConcatenationOutput Layer Concatenationβ1β2β3Attention Networkβ4.Dr. Anthony FauciBi-LSTM withattentionmechanismsα1SequenceLearning Layerα2α4Attention Networkα3hNews1h1h2h2US could see millions of coronaviruscases and 100,000 or more deaths.h.hL -1L -1hLhL.Social Sequencesvia Random WalktechniquesEmbedding Layer u[1]u[2]Random Walk LayerDeep Social Collaborative Filtering, RecSys, 2019 .u[l]11

DASOCollaborative Filtering with users’ social relations(Social Recommendation)Item DomainSocial Domain3p User behave and interact differently in the item/social domains.51Deep Adversarial Social Recommendation, IJCAI, 201912

DASOCollaborative Filtering with users’ social relations(Social Recommendation)Item DomainSocial Domain3p User behave and interact differently in the item/social domains.5Learning separated user representations in two domains.Deep Adversarial Social Recommendation, IJCAI, 2019113

DASOCollaborative Filtering with users’ social relations(Social Recommendation)Item DomainSocial Domain3p User behave and interact differently in the item/social domains.5Learning separated user representations in two domains.1Bidirectional Knowledge Transfer with Cycle ReconstructionDeep Adversarial Social Recommendation, IJCAI, 201914

Optimization for Ranking Tasksp Negative Sampling’s Main Issue: It often generates low-quality negative samples that do not help you learn goodrepresentation.Deep Adversarial Social Recommendation, IJCAI, 201915

Optimization for Ranking Tasksp Negative Sampling’s Main Issue: It often generates low-quality negative samples that do not help you learn goodrepresentation [Cai and Wang, 2018; Wang et al., ed“easy”/unrelatedinformativeDressDeep Adversarial Social Recommendation, IJCAI, 2019Men’s Wallets16

Optimization for Ranking Tasksp Negative Sampling’s Main Issue: It often generates low-quality negative samples that do not help you learn goodrepresentation [Cai and Wang, 2018; Wang et al., ed“easy”/unrelatedinformativeDressDeep Adversarial Social Recommendation, IJCAI, 2019Dynamically generate“difficult” negative samplesOptimization withAdversarial Learning(GAN)Men’s Wallets17

DASOCyclic UserModelingItem DomainAdversarial LearningSocial Domain Representations forGeneratorItem Domain Representations forGeneratorUser Representations on ItemDomain after Mapping (S- I)UIItem Domain Representations sSocial DomainAdversarial LearningRewardVUhI- ShS- IUser Representations on SocialDomain after Mapping (I- S)UISSocial Domain Representations DiscriminatorRewardLoss/RewardGeneratedSamplesp(v u)p(uk u)GeneratorGeneratorDeep Adversarial Social Recommendation, IJCAI, r18

DASOCyclic UserModelingItem DomainAdversarial LearningSocial Domain Representations forGeneratorItem Domain Representations forGeneratorUser Representations on ItemDomain after Mapping (S- I)UIItem Domain Representations sSocial DomainAdversarial LearningRewardVUhI- ShS- IUser Representations on SocialDomain after Mapping (I- S)UISSocial Domain Representations DiscriminatorRewardLoss/RewardGeneratedSamplesp(v u)p(uk u)GeneratorGeneratorDeep Adversarial Social Recommendation, IJCAI, r19

Item Domain Discriminator Modelp DiscriminatorGoal: distinguish real user-item pairs (i.e., real samples) and thegenerated “fake” samples (relevant)(Sigmoid)Score function:Deep Adversarial Social Recommendation, IJCAI, 201920

Item Domain Generator Modelp Generator ModelGoal:1. Approximate the underlying real conditional distribution pIreal(v ui)2. Generate (select/sample) the most relevant items for any given user ui.the transferred user representationfrom social domainOptimization withPolicy Gradient21

Sequential (Session-based) Recommendationuser’s sequential behavior0.80.60.1Next ItemSession-based Recommendations with Recurrent Neural Networks, ICLR, 2016.BERT4Rec: Sequential Recommendation with Bidirectional Encoder Representations from Transformer, CIKM, 2019.22

Sequential (Session-based) Recommendationuser’s sequential behavior0.80.60.1Next ItemGRU based sequential recommendation method(GRU4Rec)Session-based Recommendations with Recurrent Neural Networks, ICLR, 2016.BERT4Rec: Sequential Recommendation with Bidirectional Encoder Representations from Transformer, CIKM, 2019.23

Sequential (Session-based) Recommendationuser’s sequential behavior0.80.60.1Next ItemTransformerLayerGRU based sequential recommendation method(GRU4Rec)BERT4RecSession-based Recommendations with Recurrent Neural Networks, ICLR, 2016.BERT4Rec: Sequential Recommendation with Bidirectional Encoder Representations from Transformer, CIKM, 2019.24

Shortcomings of Existing Deep Recommender SystemsRecommendation Policies§ Offline optimization§ Short-term reward

Shortcomings of Existing Deep Recommender SystemsUsers Items𝑢!𝑣!𝑢!𝑣"𝑢!𝑣#𝑢"𝑣! Recommendation Policies§ Offline optimization§ Short-term rewardGraph-structured Data§ Information Isolated IslandIssue: ignore implicit/explicitrelationships among instances26

Shortcomings of Existing Deep Recommender SystemsUsers Items𝑢!𝑣!𝑢!𝑣"𝑢!𝑣#𝑢"𝑣! 00Field 1Recommendation Policies§ Offline optimization§ Short-term rewardGraph-structured Data§ Information Isolated IslandIssue: ignore implicit/explicitrelationships among instances1100Field m010Field MManually Deisgned Architectures§ Expert knowledge§ Time and engineering efforts27

Wide & Deep Learning for Recommender Systems, 1st DLRS, 2016, Standard MLP Embedding Concatenation , pThe wide linear models can memorize seen feature interactions using cross-product feature transformations. pThe deep models can generalize to previously unseen feature interactions through low-dimensional embeddings. 5, Neural FM,