This is the official Pytorch implementation for the paper:
ROMA: Recommendation-Oriented Language Model Adaptation Using Multi-Modal Multi-Domain Item Sequences
python>=3.9.13
cudatoolkit>=12.0
torch>=1.13.1
pytorch-lightning>=2.0.2
transformers>=4.36.2
tqdm>=4.64.1
numpy>=1.23.1
Our experiments are conducted on one assembled upstream pre-training datasets and six downstream fine-tuning datasets.
Datasets | #Users | #Items | #Img.(Cover./%) | #Inters | Avg.SL. |
---|---|---|---|---|---|
Pre-training | 3,608,532 | 1,022,309 | 724,562(70.88%) | 33,572,032 | 9.30 |
Scientific | 11,041 | 5,327 | 3,490(65.52%) | 76,896 | 6.96 |
Instruments | 27,530 | 10,611 | 6,289(59.27%) | 231,312 | 8.40 |
Pet | 47,569 | 37,970 | 30,611(80.62%) | 420,662 | 8.84 |
Arts | 56,210 | 22,855 | 13,418(58.71%) | 492,492 | 8.76 |
Games | 55,223 | 17,389 | 14,967(86.07%) | 496,315 | 8.99 |
Office | 101,501 | 27,932 | 20,542(73.54%) | 798,914 | 7.87 |
Considering the requirement of anonymity and the size limitation, we provide the data of the Scientific domain and a ROMA checkpoint fine-tuned on it for review.
Our supplementary materials include a directory named Scientific
and a checkpoint file named Scientific.ckpt
, please unzip them and put them in the same directory as test.sh
, then you can run
bash test.sh
to check our experimental result on Scientific
domain.
Further validation and open-source implementation will be available after peer review.
- If you have any questions, please feel free to give me your advice.
- Thank you for your reading and guidance.