By linear mapping the representations of language models (LMs), we can get a homomorphic item representation space for recommendation. We find that:
- Homomorphic spaces generated by advanced LMs yield excellent recommendation performance.
- Semantic similarities in language representations may imply user preference similarities.
- The complicated user preference similarity may be implicitly encoded in language spaces, with a naive linear mapping matrix to activate.
Exploring the potential of advanced language representations with leading collaborative filtering (CF) components (i.e., nonlinear projection, graph convolution, and contrastive learning objective).
"AlphaRec introduces a new language-representation-based CF paradigm with several desirable advantages: being easy to implement, lightweight, rapid convergence, superior zero-shot recommendation abilities in new domains, and being aware of user intention."
AlphaRec_demo.mp4
- Release user intention capture datasets.
- Upload zero-shot evaluation scripts.
- ...
Our experiments have been tested on Python 3.9.12 with PyTorch 1.13.1+cu117. ποΈ Python version over 3.10 may lead to some bugs in the package 'reckit'.
- Set up a virtualenv and install the pytorch manually. After that, install all the dependencies listed in the
requirements.txtfile by running the following command:
pip install -r requirements.txt- Before using the general recommendation, run the following command to install the evaluator:
pushd models/General/base
python setup.py build_ext --inplace
popdPlease download the datasets from the following anonymous link and put the unzipped dataset in the data folder:
https://drive.google.com/drive/folders/1iGKeTx3vqCtbeVdWkHOwgpbY3-s7QDy_?usp=sharing
Example of the file structure:
βββ assets/
βββ models/
βββ data/
βββ General/
βββ amazon_movie/ # target datasets
βββ cf_data/
βββ item_info/
Books
nohup python main.py --rs_type General --clear_checkpoints --saveID tau_0.15_v3_mlp_ --dataset amazon_book_2014 --model_name AlphaRec --n_layers 2 --patience 20 --cuda 0 --no_wandb --train_norm --pred_norm --neg_sample 256 --lm_model v3 --model_version mlp --tau 0.15 --infonce 1 &>logs/amazon_book_2014_tau_0.15_v3_mlp__2.log &Movies & TV
nohup python main.py --rs_type General --clear_checkpoints --saveID tau_0.15_v3_mlp_ --dataset amazon_movie --model_name AlphaRec --n_layers 2 --patience 20 --cuda 1 --no_wandb --train_norm --pred_norm --neg_sample 256 --lm_model v3 --model_version mlp --tau 0.15 --infonce 1 &>logs/amazon_movie_tau_0.15_v3_mlp__2.log &Games
nohup python main.py --rs_type General --clear_checkpoints --saveID tau_0.2_v3_mlp_ --dataset amazon_game --model_name AlphaRec --n_layers 2 --patience 20 --cuda 2 --no_wandb --train_norm --pred_norm --neg_sample 256 --lm_model v3 --model_version mlp --tau 0.2 --infonce 1 &>logs/amazon_game_tau_0.2_v3_mlp__2.log &Please contact the first author of this paper for queries.
- Leheng Sheng, leheng.sheng@u.nus.edu
You can cite this paper as follows if you find our work helpful:
@inproceedings{AlphaRec,
author = {Leheng Sheng and
An Zhang and
Yi Zhang and
Yuxin Chen and
Xiang Wang and
Tat{-}Seng Chua},
title = {Language Representations Can be What Recommenders Need: Findings and
Potentials},
booktitle = {ICLR},
year = {2025}
}

