SlideShare a Scribd company logo
CASER : Personalized Top-N
Sequential Recommendation via
Convolutional Sequence Embedding
Ju-Hee SHIM
Network Science Lab
Dept. of AI
The Catholic University of Korea
E-mail: shim020616@catholic.ac.kr
Jiaxi Tang, Ke Wang
WSDM 2018
2
 INTRODUCTION
• Motivation
 Architecture
• CASER
• Method
 Evaluation
• Datasets
• State-of-the-art methods
• Experimental setup
• Results
 CONCLUSION
Q/A
3
INTRODUCTION
Motivation
 Problems of Existing top-N Recommendation Models
• User’s general preferences are used as the basis for recommendations
• User’s general preferences : Reflect only static behavioral information of the user (ex. A person who likes
Samsung products is only recommended Samsung products, and a person who likes Apple products is
only recommended Apple products)
• However, these unidirectional models have the following limitations:
 Simply recommend items related IPHONE
: losing the opportunity to recommend phone accessories
4
INTRODUCTION
Motivation
 Limitations of Traditional Markov Chain-Based Models
a) Point-Level : The probability of purchasing a specific item often increases when multiple past items are
combined -> fail to capture this effect (ex. A user who buys milk and butter it likely to purchase flour, but this
is not reflected)
b) Skip Behaviors : Unable to account for skipped behaviors -> Traditional models assume continuous influence,
but in reall-world data, “skip” frequently occur
5
Architecture
 Transforming User Sequences into a Matrix “IMAGE” :
• Applying CNN :
• Convert the traditional 1D item sequence into an L x d matrix.
• L : the most recent L items
• d : Embedding dimension
• Horizontal Filters : Learning Union-Level Sequential Patterns
• Capturing patterns where multiple item combinations influence behavior
• Vertical Filters : Learning Point-Level Sequential Patterns
• Similar to traditional Markov Chain approaches
• Adding User embedding :
• Incorporate User Embedding to model Long-term user preferences effectively
CASER
6
Architecture
 Transformer Layer:
• Consists of L bidirectional Transformer layers.
• Each layer refines the user behavior sequence
received from the previous layer to enhance
representation power.
• In each layer, all item representations influence and
update each other.
• Unlike RNN-based models, which pass information
only from past to future, Self-Attention enables global
interaction across all items in the sequence.
Method
7
Architecture
Method
 Embedding Look-up:
• Retrieving Past Item Embeddings :
• Locate L past item embeddings of user U in the latent space
• Stack these embedding to construct the final Embedding
matrix (E) for training
• Create an embedding table using d-dimensional latent factors
• Q(item), P(User)
8
Architecture
Method
 Convolutional Layers :
• Treat the embedding matrix (E) as an "image" and apply
convolutional layers to capture sequential patterns in
user behavior
• Consider sequential patterns as local features within the
image
• Utilize two types of convolutional filters:
• 1) Vertical Convolutional Layer :
• Captures point-level sequential patterns
• Computes a weighted sum over the latent
representations of the past L items
9
Architecture
Method
 Convolutional Layers :
• 2) Horizontal Convolutional Layer :
• Captures union-level patterns.
• Varies the filter height (h) to extract diverse
sequential features
• To extracted most Significant feature, using
max-pooling
10
Architecture
Method
 Fully-connected Layers :
• Concatenate the outputs from the horizontal and
vertical filters
• Feed the concatenated features into a fully-connected
layer to extract high-level abstract features
• Concatenate user embedding with the extracted
features to capture general user preferences -> Pass the
final representation to the output layer for prediction
11
Architecture
Method
 Network Training & Recommendation:
• Apply the sigmoid activation function to the output layer to transform the output value y into a
probability.
• Compute the likelihood across all sequences in the dataset for training
• Use the user’s last L item embeddings to compute y-values for all items
• Select the top-N items with the highest y-values for recommendatio
12
Datasets
Evaluation
 MovieLens
 Gowalla
 Foursquare
 Tmall
13
Evaluation
State-of-the-art methods
 Compared Methods
• POP
• BPR
• FMC
• FPMC
• Fossil
• GRU4Rec
14
Evaluation
Experimental setup
•Evaluation Metrics
• Precision@N
• Recall@N
• MAP
•Optimizer: Adam
•Learning Rate: {1,10^-1,…,10^-4} grid search
•Batch Size : 100
•L2 Regularization
•Dropout : 50%
•Latent Dimensions d : {5,10,20,30,50,100}
•Markov Order L : {1,2,3,…,9}
•Target Number T : {1,2,3}
•Activation Functions : {identity, sigmoid, tanh, ReLU}
•Number of Horizontal Filters : {4,8,16,32,64}
•Number of Vertical Filters : {1,2,4,8,16}
•Loss : BCE loss
•Negative Sampling : random item 3
15
Evaluation
Results
16
Evaluation
Results
 Ablation Study Results
• Caser model outperforms Fossil, GRU4Rec in terms of MAP, with the best performance observed at T =2,3.
• As the Markov Order L increases, performance improves and then plateaus; in sparse datasets,
excessively large L can lead to performance degradation.
• Markov Targets T contributes to performance improvement = Predicting multiple future items
simultaneously is more effective than predicting just one
17
Evaluation
Results
 Ablation Study Results
• Performance results based on the usage of each compontent
• P : personalization(user embedding), h: horizontal convolutional layer, v : vertical convolutional
layer
• The best performance is achieved when all three components are used together
18
Conclusion
Conclusion
 The author of this paper was proposing CASER, a novel approach to top-N sequential recommendation.
CASER captures information from point-level and union-level sequential patterns, skip behaviors, and long-
term user preferences.
 A unique aspect of CASER is it’s attempt to interpret a user’s 1D sequence as a 2D image representation.
This approach could be particularly meaningful in industries where the sequential dependency of user
behavior is weak.
19
Q & A
Q / A

More Related Content

PPTX
250203_JH_labseminar[BERT4Rec : Sequential Recommendation with Bidirectional ...
thanhdowork
 
PPTX
250203_JH_labseminar[BERT4Rec : Sequential Recommendation with Bidirectional ...
thanhdowork
 
PPTX
250113_JH_labseminar[An Attentive Inductive Bias for Sequential Recommendatio...
thanhdowork
 
PDF
Tutorial: Context In Recommender Systems
YONG ZHENG
 
PDF
Two strategies for large-scale multi-label classification on the YouTube-8M d...
Dalei Li
 
PPTX
Recommender Systems: Advances in Collaborative Filtering
Changsung Moon
 
PDF
Gaussian Ranking by Matrix Factorization, ACM RecSys Conference 2015
Harald Steck
 
PPTX
A Combination of Simple Models by Forward Predictor Selection for Job Recomme...
David Zibriczky
 
250203_JH_labseminar[BERT4Rec : Sequential Recommendation with Bidirectional ...
thanhdowork
 
250203_JH_labseminar[BERT4Rec : Sequential Recommendation with Bidirectional ...
thanhdowork
 
250113_JH_labseminar[An Attentive Inductive Bias for Sequential Recommendatio...
thanhdowork
 
Tutorial: Context In Recommender Systems
YONG ZHENG
 
Two strategies for large-scale multi-label classification on the YouTube-8M d...
Dalei Li
 
Recommender Systems: Advances in Collaborative Filtering
Changsung Moon
 
Gaussian Ranking by Matrix Factorization, ACM RecSys Conference 2015
Harald Steck
 
A Combination of Simple Models by Forward Predictor Selection for Job Recomme...
David Zibriczky
 

Similar to 250310_JH_labseminar[CASER : Personalized Top-N Sequential Recommendation via Convolutional Sequence Embedding].pptx (20)

PDF
An introduction to variable and feature selection
Marco Meoni
 
PDF
Where Search Meets Machine Learning: Presented by Diana Hu & Joaquin Delgado,...
Lucidworks
 
PDF
Lucene/Solr Revolution 2015: Where Search Meets Machine Learning
Joaquin Delgado PhD.
 
PDF
Lucene/Solr Revolution 2015: Where Search Meets Machine Learning
S. Diana Hu
 
PPTX
Building High Available and Scalable Machine Learning Applications
Yalçın Yenigün
 
PPTX
BMDSE v1 - Data Scientist Deck
Sasha Lazarevic
 
PDF
DutchMLSchool 2022 - History and Developments in ML
BigML, Inc
 
PDF
1710 track3 zhu
Rising Media, Inc.
 
PPTX
230727_HB_JointJournalClub.pptx
Network Science Lab, The Catholic University of Korea
 
PDF
[IUI 2017] Criteria Chains: A Novel Multi-Criteria Recommendation Approach
YONG ZHENG
 
PPT
Nbvtalkonfeatureselection
Nagasuri Bala Venkateswarlu
 
PPTX
An Introduction to Deep Learning
milad abbasi
 
PPTX
Introduction to Deep Learning
Mehrnaz Faraz
 
PPTX
Machine Learning
Girish Khanzode
 
PPTX
Collaborative Metric Learning (WWW'17)
承剛 謝
 
PPTX
Machine Learning Deep Learning Deep Learning
ssuserd89c50
 
PPTX
How can pre-training help to solve the cold start problem?
Lokesh Vadlamudi
 
PDF
General Tips for participating Kaggle Competitions
Mark Peng
 
PPTX
data reduction techniques-data minig.pptx
farheengul004
 
PPTX
Utilizing Marginal Net Utility for Recommendation in E-commerce
Liangjie Hong
 
An introduction to variable and feature selection
Marco Meoni
 
Where Search Meets Machine Learning: Presented by Diana Hu & Joaquin Delgado,...
Lucidworks
 
Lucene/Solr Revolution 2015: Where Search Meets Machine Learning
Joaquin Delgado PhD.
 
Lucene/Solr Revolution 2015: Where Search Meets Machine Learning
S. Diana Hu
 
Building High Available and Scalable Machine Learning Applications
Yalçın Yenigün
 
BMDSE v1 - Data Scientist Deck
Sasha Lazarevic
 
DutchMLSchool 2022 - History and Developments in ML
BigML, Inc
 
1710 track3 zhu
Rising Media, Inc.
 
[IUI 2017] Criteria Chains: A Novel Multi-Criteria Recommendation Approach
YONG ZHENG
 
Nbvtalkonfeatureselection
Nagasuri Bala Venkateswarlu
 
An Introduction to Deep Learning
milad abbasi
 
Introduction to Deep Learning
Mehrnaz Faraz
 
Machine Learning
Girish Khanzode
 
Collaborative Metric Learning (WWW'17)
承剛 謝
 
Machine Learning Deep Learning Deep Learning
ssuserd89c50
 
How can pre-training help to solve the cold start problem?
Lokesh Vadlamudi
 
General Tips for participating Kaggle Competitions
Mark Peng
 
data reduction techniques-data minig.pptx
farheengul004
 
Utilizing Marginal Net Utility for Recommendation in E-commerce
Liangjie Hong
 
Ad

More from thanhdowork (20)

PPTX
250728_Thuy_Labseminar[Predictive Chemistry Augmented with Text Retrieval].pptx
thanhdowork
 
PPTX
[NS][Lab_Seminar_250728]NeuralWalker.pptx
thanhdowork
 
PPTX
A Novel Shape-Aware Topological Representation for GPR Data with DNN Integrat...
thanhdowork
 
PPTX
250721_Thien_Labseminar[Variational Graph Auto-Encoders].pptx
thanhdowork
 
PPTX
250721_HW_LabSeminar[RingFormer: A Ring-Enhanced Graph Transformer for Organi...
thanhdowork
 
PPTX
250721_Thuy_Labseminar[Thought Propagation: An Analogical Approach to Complex...
thanhdowork
 
PPTX
[NS][Lab_Seminar_250721]On Measuring Long-Range Interactions in Graph Neural ...
thanhdowork
 
PPTX
250714_HW_LabSeminar[Structural Reasoning Improves Molecular Understanding of...
thanhdowork
 
PPTX
[NS][Lab_Seminar_250714]Candidate Set Re-ranking for Composed Image Retrieval...
thanhdowork
 
PPTX
250714_Thuy_Labseminar[BioT5: Enriching Cross-modal Integration in Biology wi...
thanhdowork
 
PPTX
250707_HW_LabSeminar[CHEMICAL-REACTION-AWARE MOLECULE REPRESENTATION LEARNING...
thanhdowork
 
PPTX
[NS][Lab_Seminar_250707]Learning with Noisy Triplet Correspondence for Compos...
thanhdowork
 
PPTX
250707_JW_labseminar[CBAM: Convolutional Block Attention Module].pptx
thanhdowork
 
PPTX
250707_Thuy_Labseminar[3D-MOLM: TOWARDS 3D MOLECULE-TEXT INTERPRETATION IN LA...
thanhdowork
 
PPTX
[NS][Lab_Seminar_250623]ConText-CIR.pptx
thanhdowork
 
PPTX
250629_HW_LabSeminar[ReactGPT: Understanding of Chemical Reactions via In-Con...
thanhdowork
 
PPTX
250630_JW_labseminar[Does GNN Pretraining Help Molecular].pptx
thanhdowork
 
PPTX
[NS][Lab_Seminar_250623]UniFashion: A Unified Vision-Language Model for Multi...
thanhdowork
 
PPTX
250623_JW_labseminar[STRATEGIES FOR PRE-TRAINING GRAPH NEURAL NETWORKS].pptx
thanhdowork
 
PPTX
[NS][Lab_Seminar_250616]FashionERN: Enhance-and-Refine Network for Composed F...
thanhdowork
 
250728_Thuy_Labseminar[Predictive Chemistry Augmented with Text Retrieval].pptx
thanhdowork
 
[NS][Lab_Seminar_250728]NeuralWalker.pptx
thanhdowork
 
A Novel Shape-Aware Topological Representation for GPR Data with DNN Integrat...
thanhdowork
 
250721_Thien_Labseminar[Variational Graph Auto-Encoders].pptx
thanhdowork
 
250721_HW_LabSeminar[RingFormer: A Ring-Enhanced Graph Transformer for Organi...
thanhdowork
 
250721_Thuy_Labseminar[Thought Propagation: An Analogical Approach to Complex...
thanhdowork
 
[NS][Lab_Seminar_250721]On Measuring Long-Range Interactions in Graph Neural ...
thanhdowork
 
250714_HW_LabSeminar[Structural Reasoning Improves Molecular Understanding of...
thanhdowork
 
[NS][Lab_Seminar_250714]Candidate Set Re-ranking for Composed Image Retrieval...
thanhdowork
 
250714_Thuy_Labseminar[BioT5: Enriching Cross-modal Integration in Biology wi...
thanhdowork
 
250707_HW_LabSeminar[CHEMICAL-REACTION-AWARE MOLECULE REPRESENTATION LEARNING...
thanhdowork
 
[NS][Lab_Seminar_250707]Learning with Noisy Triplet Correspondence for Compos...
thanhdowork
 
250707_JW_labseminar[CBAM: Convolutional Block Attention Module].pptx
thanhdowork
 
250707_Thuy_Labseminar[3D-MOLM: TOWARDS 3D MOLECULE-TEXT INTERPRETATION IN LA...
thanhdowork
 
[NS][Lab_Seminar_250623]ConText-CIR.pptx
thanhdowork
 
250629_HW_LabSeminar[ReactGPT: Understanding of Chemical Reactions via In-Con...
thanhdowork
 
250630_JW_labseminar[Does GNN Pretraining Help Molecular].pptx
thanhdowork
 
[NS][Lab_Seminar_250623]UniFashion: A Unified Vision-Language Model for Multi...
thanhdowork
 
250623_JW_labseminar[STRATEGIES FOR PRE-TRAINING GRAPH NEURAL NETWORKS].pptx
thanhdowork
 
[NS][Lab_Seminar_250616]FashionERN: Enhance-and-Refine Network for Composed F...
thanhdowork
 
Ad

Recently uploaded (20)

PPTX
Virus sequence retrieval from NCBI database
yamunaK13
 
PDF
Antianginal agents, Definition, Classification, MOA.pdf
Prerana Jadhav
 
PPTX
Measures_of_location_-_Averages_and__percentiles_by_DR SURYA K.pptx
Surya Ganesh
 
PPTX
How to Close Subscription in Odoo 18 - Odoo Slides
Celine George
 
PPTX
Dakar Framework Education For All- 2000(Act)
santoshmohalik1
 
PPTX
CDH. pptx
AneetaSharma15
 
PPTX
Applications of matrices In Real Life_20250724_091307_0000.pptx
gehlotkrish03
 
PDF
Module 2: Public Health History [Tutorial Slides]
JonathanHallett4
 
PDF
Review of Related Literature & Studies.pdf
Thelma Villaflores
 
PPTX
An introduction to Dialogue writing.pptx
drsiddhantnagine
 
PDF
BÀI TẬP TEST BỔ TRỢ THEO TỪNG CHỦ ĐỀ CỦA TỪNG UNIT KÈM BÀI TẬP NGHE - TIẾNG A...
Nguyen Thanh Tu Collection
 
PPTX
HISTORY COLLECTION FOR PSYCHIATRIC PATIENTS.pptx
PoojaSen20
 
PPTX
Care of patients with elImination deviation.pptx
AneetaSharma15
 
PPTX
family health care settings home visit - unit 6 - chn 1 - gnm 1st year.pptx
Priyanshu Anand
 
PPTX
An introduction to Prepositions for beginners.pptx
drsiddhantnagine
 
PPTX
Gupta Art & Architecture Temple and Sculptures.pptx
Virag Sontakke
 
PPTX
Python-Application-in-Drug-Design by R D Jawarkar.pptx
Rahul Jawarkar
 
PPTX
Command Palatte in Odoo 18.1 Spreadsheet - Odoo Slides
Celine George
 
PPTX
HEALTH CARE DELIVERY SYSTEM - UNIT 2 - GNM 3RD YEAR.pptx
Priyanshu Anand
 
PPTX
Introduction to pediatric nursing in 5th Sem..pptx
AneetaSharma15
 
Virus sequence retrieval from NCBI database
yamunaK13
 
Antianginal agents, Definition, Classification, MOA.pdf
Prerana Jadhav
 
Measures_of_location_-_Averages_and__percentiles_by_DR SURYA K.pptx
Surya Ganesh
 
How to Close Subscription in Odoo 18 - Odoo Slides
Celine George
 
Dakar Framework Education For All- 2000(Act)
santoshmohalik1
 
CDH. pptx
AneetaSharma15
 
Applications of matrices In Real Life_20250724_091307_0000.pptx
gehlotkrish03
 
Module 2: Public Health History [Tutorial Slides]
JonathanHallett4
 
Review of Related Literature & Studies.pdf
Thelma Villaflores
 
An introduction to Dialogue writing.pptx
drsiddhantnagine
 
BÀI TẬP TEST BỔ TRỢ THEO TỪNG CHỦ ĐỀ CỦA TỪNG UNIT KÈM BÀI TẬP NGHE - TIẾNG A...
Nguyen Thanh Tu Collection
 
HISTORY COLLECTION FOR PSYCHIATRIC PATIENTS.pptx
PoojaSen20
 
Care of patients with elImination deviation.pptx
AneetaSharma15
 
family health care settings home visit - unit 6 - chn 1 - gnm 1st year.pptx
Priyanshu Anand
 
An introduction to Prepositions for beginners.pptx
drsiddhantnagine
 
Gupta Art & Architecture Temple and Sculptures.pptx
Virag Sontakke
 
Python-Application-in-Drug-Design by R D Jawarkar.pptx
Rahul Jawarkar
 
Command Palatte in Odoo 18.1 Spreadsheet - Odoo Slides
Celine George
 
HEALTH CARE DELIVERY SYSTEM - UNIT 2 - GNM 3RD YEAR.pptx
Priyanshu Anand
 
Introduction to pediatric nursing in 5th Sem..pptx
AneetaSharma15
 

250310_JH_labseminar[CASER : Personalized Top-N Sequential Recommendation via Convolutional Sequence Embedding].pptx

  • 1. CASER : Personalized Top-N Sequential Recommendation via Convolutional Sequence Embedding Ju-Hee SHIM Network Science Lab Dept. of AI The Catholic University of Korea E-mail: [email protected] Jiaxi Tang, Ke Wang WSDM 2018
  • 2. 2  INTRODUCTION • Motivation  Architecture • CASER • Method  Evaluation • Datasets • State-of-the-art methods • Experimental setup • Results  CONCLUSION Q/A
  • 3. 3 INTRODUCTION Motivation  Problems of Existing top-N Recommendation Models • User’s general preferences are used as the basis for recommendations • User’s general preferences : Reflect only static behavioral information of the user (ex. A person who likes Samsung products is only recommended Samsung products, and a person who likes Apple products is only recommended Apple products) • However, these unidirectional models have the following limitations:  Simply recommend items related IPHONE : losing the opportunity to recommend phone accessories
  • 4. 4 INTRODUCTION Motivation  Limitations of Traditional Markov Chain-Based Models a) Point-Level : The probability of purchasing a specific item often increases when multiple past items are combined -> fail to capture this effect (ex. A user who buys milk and butter it likely to purchase flour, but this is not reflected) b) Skip Behaviors : Unable to account for skipped behaviors -> Traditional models assume continuous influence, but in reall-world data, “skip” frequently occur
  • 5. 5 Architecture  Transforming User Sequences into a Matrix “IMAGE” : • Applying CNN : • Convert the traditional 1D item sequence into an L x d matrix. • L : the most recent L items • d : Embedding dimension • Horizontal Filters : Learning Union-Level Sequential Patterns • Capturing patterns where multiple item combinations influence behavior • Vertical Filters : Learning Point-Level Sequential Patterns • Similar to traditional Markov Chain approaches • Adding User embedding : • Incorporate User Embedding to model Long-term user preferences effectively CASER
  • 6. 6 Architecture  Transformer Layer: • Consists of L bidirectional Transformer layers. • Each layer refines the user behavior sequence received from the previous layer to enhance representation power. • In each layer, all item representations influence and update each other. • Unlike RNN-based models, which pass information only from past to future, Self-Attention enables global interaction across all items in the sequence. Method
  • 7. 7 Architecture Method  Embedding Look-up: • Retrieving Past Item Embeddings : • Locate L past item embeddings of user U in the latent space • Stack these embedding to construct the final Embedding matrix (E) for training • Create an embedding table using d-dimensional latent factors • Q(item), P(User)
  • 8. 8 Architecture Method  Convolutional Layers : • Treat the embedding matrix (E) as an "image" and apply convolutional layers to capture sequential patterns in user behavior • Consider sequential patterns as local features within the image • Utilize two types of convolutional filters: • 1) Vertical Convolutional Layer : • Captures point-level sequential patterns • Computes a weighted sum over the latent representations of the past L items
  • 9. 9 Architecture Method  Convolutional Layers : • 2) Horizontal Convolutional Layer : • Captures union-level patterns. • Varies the filter height (h) to extract diverse sequential features • To extracted most Significant feature, using max-pooling
  • 10. 10 Architecture Method  Fully-connected Layers : • Concatenate the outputs from the horizontal and vertical filters • Feed the concatenated features into a fully-connected layer to extract high-level abstract features • Concatenate user embedding with the extracted features to capture general user preferences -> Pass the final representation to the output layer for prediction
  • 11. 11 Architecture Method  Network Training & Recommendation: • Apply the sigmoid activation function to the output layer to transform the output value y into a probability. • Compute the likelihood across all sequences in the dataset for training • Use the user’s last L item embeddings to compute y-values for all items • Select the top-N items with the highest y-values for recommendatio
  • 13. 13 Evaluation State-of-the-art methods  Compared Methods • POP • BPR • FMC • FPMC • Fossil • GRU4Rec
  • 14. 14 Evaluation Experimental setup •Evaluation Metrics • Precision@N • Recall@N • MAP •Optimizer: Adam •Learning Rate: {1,10^-1,…,10^-4} grid search •Batch Size : 100 •L2 Regularization •Dropout : 50% •Latent Dimensions d : {5,10,20,30,50,100} •Markov Order L : {1,2,3,…,9} •Target Number T : {1,2,3} •Activation Functions : {identity, sigmoid, tanh, ReLU} •Number of Horizontal Filters : {4,8,16,32,64} •Number of Vertical Filters : {1,2,4,8,16} •Loss : BCE loss •Negative Sampling : random item 3
  • 16. 16 Evaluation Results  Ablation Study Results • Caser model outperforms Fossil, GRU4Rec in terms of MAP, with the best performance observed at T =2,3. • As the Markov Order L increases, performance improves and then plateaus; in sparse datasets, excessively large L can lead to performance degradation. • Markov Targets T contributes to performance improvement = Predicting multiple future items simultaneously is more effective than predicting just one
  • 17. 17 Evaluation Results  Ablation Study Results • Performance results based on the usage of each compontent • P : personalization(user embedding), h: horizontal convolutional layer, v : vertical convolutional layer • The best performance is achieved when all three components are used together
  • 18. 18 Conclusion Conclusion  The author of this paper was proposing CASER, a novel approach to top-N sequential recommendation. CASER captures information from point-level and union-level sequential patterns, skip behaviors, and long- term user preferences.  A unique aspect of CASER is it’s attempt to interpret a user’s 1D sequence as a 2D image representation. This approach could be particularly meaningful in industries where the sequential dependency of user behavior is weak.
  • 19. 19 Q & A Q / A

Editor's Notes

  • #19: thank you, the presentation is concluded