SkillAgentSearch skills...

PERIS

Official code of Beyond Learning from Next Item: Sequential Recommendation via Personalized Interest Sustainability [CIKM'22]

Install / Use

/learn @dmhyun/PERIS
About this skill

Quality Score

0/100

Supported Platforms

Zed

README

<h1 align="center"> <span> Beyond Learning from Next Item: Sequential Recommendation via Personalized Interest Sustainability </span> </h1> <p align="center"> <a href="https://www.cikm2022.org/" alt="Conference"> <img src="https://img.shields.io/badge/CIKM'22-Full%20paper-brightgreen" /></a> <a href="https://pytorch.org/" alt="PyTorch"> <img src="https://img.shields.io/badge/PyTorch-%23EE4C2C.svg?e&logo=PyTorch&logoColor=white" /></a> </p> <p align="center"> <span>Official implementation of </span> <a href="https://arxiv.org/pdf/2209.06644.pdf">CIKM'22 paper</a> </p>

Overview

Sequential Recommender Systems

There have been two groups of existing sequential models: user- and item-centric models. This work proposes a recommender system taking advantages of the models in both categories.

<p align="center"><img src="images/intro_comparison.png" alt="graph" width="45%"></p>

Personalized Interest Sustainability with supplementaion schemes

We formulate a task to predict which items each user will consume in the recent period of the training time based on users' consumption history. We then devise simple yet effective schemes to supplement users’ sparse consumption history in both intrinsic and extrinsic manners.

<p align="center"><img src="images/model.png" alt="graph" width="85%"></p>

Recommendation Performance

PERIS significantly outperforms baseline models including general, user-centric, and item-centric sequential models on 11 real-world datasets. This result indicates the effectiveness of the personalized interest sustainability.

<p align="center"><img src="images/performance.png" alt="graph" width="65%"></p>

Major Requirements

  • Python
  • Pytorch
  • Numpy

Preprocessing Data

  1. Download user-item consumption data (and extract the compressed file) into ./data/.

    • Amazon <pre>[Example] <code>wget http://snap.stanford.edu/data/amazon/productGraph/categoryFiles/ratings_Cell_Phones_and_Accessories.csv</code></pre>
    • Yelp
    • Google Maps
    • Other data you want

    :exclamation: Please make sure your data in the same csv format of Amazon data.

    • For Yelp and Google datasets, run the following code to transform the data format.
    <pre><code>python preprocess_yelp.py YELP_DATA.json</code></pre> <pre><code>python preprocess_google.py reviews.clean.json</code></pre>
  2. Split your data into training/validation/test data in ./data/.

<pre><code>python split_data.py your_decompressed_file.csv</code></pre>
  1. Build a dataset for training a recommender syetem with using the splitted data.
<pre><code>python build_recdata.py generated_directory </code></pre>

Training

Train a recommender system with a model name.

<pre><code>python train.py --model_name peris --dataset your_dataset --learning_rate 1e-3 --lamb 0.5 --mu 0.3 --K 128 </code></pre>

Available models

| Model | Paper | Conference | | ------------- | ------------- | ------------- | | peris | Beyond Learning from Next Item: Sequential Recommendation via Personalized Interest Sustainability | CIKM'22 | | lsan | Lightweight Self-Attentive Sequential Recommendation | CIKM'21 | | simplex | SimpleX: A Simple and Strong Baseline for Collaborative Filtering | CIKM'21 |

Citation

If you use this repository for your work, please consider citing our paper:

<pre><code>@inproceedings{hyun2022beyond, title={Beyond Learning from Next Item: Sequential Recommendation via Personalized Interest Sustainability}, author={Hyun, Dongmin and Park, Chanyoung and Cho, Junsu and Yu, Hwanjo}, booktitle={Proceedings of the 31st ACM International Conference on Information \& Knowledge Management}, pages={812--821}, year={2022} } </code></pre>

Related Skills

View on GitHub
GitHub Stars36
CategoryEducation
Updated1y ago
Forks1

Languages

Python

Security Score

65/100

Audited on Jan 21, 2025

No findings