LLM for recommendation tasks, based on the P5 paradigm
Top 82.4% on sourcepulse
This repository implements P5, a unified pre-training paradigm for recommendation systems that treats all data as natural language sequences. It aims to advance recommender systems towards a universal engine capable of zero-shot and few-shot predictions via personalized prompts, benefiting researchers and practitioners in recommendation and natural language processing.
How It Works
P5 converts diverse recommendation data (user-item interactions, metadata, reviews) into natural language sequences. It leverages a unified language modeling objective for pre-training, enabling it to serve as a foundation model for various downstream recommendation tasks. This approach facilitates multi-modal integration and instruction-based recommendation, reducing the need for extensive fine-tuning through adaptive personalized prompts.
Quick Start & Requirements
pip install -r requirements.txt
(requirements.txt not provided, inferred from README).data
folder. Raw data can be downloaded and placed in raw_data
.snap
folder.scripts
folder (e.g., bash scripts/pretrain_P5_base_beauty.sh 4
for 4 GPUs).notebooks
folder.Highlighted Details
Maintenance & Community
The project cites VL-T5, PETER, and S3-Rec as acknowledgements. No specific community links (Discord, Slack) or roadmap are provided in the README.
Licensing & Compatibility
The README does not explicitly state a license. The code is released for research purposes. Compatibility for commercial use or closed-source linking is not specified.
Limitations & Caveats
The project relies on specific older versions of PyTorch (1.10.1) and transformers (4.2.1), which may pose compatibility challenges with newer libraries. The README does not detail specific performance benchmarks or known limitations beyond the scope of the research paper.
1 year ago
Inactive