[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
Skip to main content

Leveraging Large Language Models Knowledge Enhancement Dual-Stage Fine-Tuning Framework for Recommendation

  • Conference paper
  • First Online:
Natural Language Processing and Chinese Computing (NLPCC 2024)

Part of the book series: Lecture Notes in Computer Science ((LNAI,volume 15360))

  • 114 Accesses

Abstract

Large language models(LLMs) have exhibited notable general-purpose task-solving abilities in language understanding and generation, including processing recommendation tasks. The majority of existing research relies on training-free recommendation models that treat LLMs as reasoning engines and directly given the recommended task response. This approach heavily relies on pre-trained knowledge and may lead to excessive costs. As such, we propose a two-stage fine-tuning framework leveraging LLaMA2 and GPT-4 Knowledge Enhancement for recommendation. In particular, we use GPT-4 Instruction-Following data to tune the LLM in first-stage instruction tuning process, achieving lower training costs and better inference performance. In the second stage, through a elaborately designed prompt template, we fine-tune LLM from the first stage in a few-shot setting by interactive sequences based on user ratings. To validate the effectiveness of our framework, we compare against state-of-the-art baseline methods on benchmark datasets. The results demonstrate that our framework has promising recommendation capabilities. Our experiments are executed on a single RTX4090 with LLaMA2-7B.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Subscribe and save

Springer+ Basic
£29.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Chapter
GBP 19.95
Price includes VAT (United Kingdom)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
GBP 49.99
Price includes VAT (United Kingdom)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
GBP 64.99
Price includes VAT (United Kingdom)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Similar content being viewed by others

Notes

  1. 1.

    https://huggingface.co/meta-llama/Llama-2--7b-hf.

  2. 2.

    https://crfm.stanford.edu/2023/03/13/alpaca.html.

References

  1. Bao, K., Zhang, J., Zhang, Y., Wang, W., Feng, F., He, X.: TALLRec: an effective and efficient tuning framework to align large language model with recommendation. In: Proceedings of the 17th ACM Conference on Recommender Systems, pp. 1007–1014 (2023)

    Google Scholar 

  2. Brown, T., et al.: Language models are few-shot learners. Adv. Neural. Inf. Process. Syst. 33, 1877–1901 (2020)

    Google Scholar 

  3. Chen, Z.: PALR: Personalization aware LLMs for recommendation. arXiv preprint arXiv:2305.07622 (2023)

  4. Dettmers, T., Pagnoni, A., Holtzman, A., Zettlemoyer, L.: QLORA: efficient finetuning of quantized LLMs. In: Proceedings of NeurIPS (2023)

    Google Scholar 

  5. Gao, Y., Sheng, T., Xiang, Y., Xiong, Y., Wang, H., Zhang, J.: Chat-REC: Towards interactive and explainable LLMs-augmented recommender system. arXiv preprint arXiv:2303.14524 (2023)

  6. He, R., McAuley, J.: Ups and Downs: modeling the visual evolution of fashion trends with one-class collaborative filtering. In: Proceedings of the 25th International Conference on World Wide Web, pp. 507–517 (2016)

    Google Scholar 

  7. Hidasi, B., Karatzoglou, A., Baltrunas, L., Tikk, D.: Session-based recommendations with recurrent neural networks. arXiv preprint arXiv:1511.06939 (2015)

  8. Hou, Y., et al.: Large language models are zero-shot rankers for recommender systems. In: European Conference on Information Retrieval, pp. 364–381 (2024)

    Google Scholar 

  9. Hu, E.J., et al.: LORA: low-rank adaptation of large language models. arXiv preprint arXiv:2106.09685 (2021)

  10. Kang, W.C., McAuley, J.: Self-attentive sequential recommendation. In: Proceeding of ICDM, pp. 197–206 (2018)

    Google Scholar 

  11. Li, J., et al.: Coarse-to-fine sparse sequential recommendation. In: Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval, pp. 2082–2086 (2022)

    Google Scholar 

  12. Li, J., Ren, P., Chen, Z., Ren, Z., Lian, T., Ma, J.: Neural attentive session-based recommendation. In: Proceedings of the 2017 ACM on Conference on Information and Knowledge Management, pp. 1419–1428 (2017)

    Google Scholar 

  13. Maxwell, F., Konstan, J.A.: The MovieLens datasets: history and context. Proceedings of ACM Transactions on Interactive Intelligent Systems (2015)

    Google Scholar 

  14. Peng, B., Li, C., He, P., Galley, M., Gao, J.: Instruction tuning with GPT-4.arXiv preprint arXiv:2304.03277 (2023)

  15. Sileo, D., Vossen, W., Raymaekers, R.: Zero-shot recommendation as language modeling. In: Proceedings of the 44th European Conference on IR Research, pp. 223–230 (2022)

    Google Scholar 

  16. Steck, H., Baltrunas, L., Elahi, E., Liang, D., Raimond, Y., Basilico, J.: Deep learning for recommender systems: a netflix case study. AI Mag. 42(3), 7–18 (2021)

    Google Scholar 

  17. Tang, J., Wang, K.: Personalized top-n sequential recommendation via convolutional sequence embedding. In: Proceedings of WSDM, pp. 565–573 (2018)

    Google Scholar 

  18. Touvron, H., et al.: LLaMA: open and efficient foundation language models. arXiv preprint arXiv:2307.09288 (2023)

  19. Wang, L., Lim, E.P.: Zero-shot next-item recommendation using large pretrained language models. arXiv preprint arXiv:2304.03153 (2023)

  20. Wang, Z., Wei, W., Cong, G., Li, X., Mao, X., Qiu, M.: Global context enhanced graph neural networks for session-based recommendation. In: Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval, pp. 169–178 (2020)

    Google Scholar 

  21. Wu, L., et al.: A survey on large language models for recommendation. arXiv preprint arXiv:2305.19860 (2023)

  22. Wu, S., Tang, Y., Zhu, Y., Wang, L., Xie, X., Tan, T.: Session-based recommendation with graph neural networks. In: Proceedings of the AAAI Conference on Artificial Intelligence, pp. 346–353 (2019)

    Google Scholar 

  23. Wu, Z., Geiger, A., Icard, T., Potts, C., Goodman, N.D.: Interpretability at scale: identifying causal mechanisms in alpaca. In: Proceedings of NeurIPS (2023)

    Google Scholar 

  24. Yang, Z., et al.: A generic learning framework for sequential recommendation with distribution shifts. arXiv preprint arXiv:2310.20487 (2023)

  25. Yue, Z., Rabhi, S., Moreira, G.d.S.P., Wang, D., Oldridge, E.: LlamaRec: two-stage recommendation using large language models for ranking. In: Proceedings of CIKM (2023)

    Google Scholar 

  26. Ziegler, C.N., McNee, S.M., Konstan, J.A., Lausen, G.: Improving recommendation lists through topic diversification. In: Proceedings of the 14th International Conference on World Wide Web (WWW ’05), pp. 22–32 (2005)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Biqing Zeng .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2025 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Zeng, B., Shi, H., Li, Y., Li, R., Deng, H. (2025). Leveraging Large Language Models Knowledge Enhancement Dual-Stage Fine-Tuning Framework for Recommendation. In: Wong, D.F., Wei, Z., Yang, M. (eds) Natural Language Processing and Chinese Computing. NLPCC 2024. Lecture Notes in Computer Science(), vol 15360. Springer, Singapore. https://doi.org/10.1007/978-981-97-9434-8_26

Download citation

  • DOI: https://doi.org/10.1007/978-981-97-9434-8_26

  • Published:

  • Publisher Name: Springer, Singapore

  • Print ISBN: 978-981-97-9433-1

  • Online ISBN: 978-981-97-9434-8

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics