Extracting Implicit User Preferences in Conversational Recommender Systems Using Large Language Models
<p>Overview of the proposed approach.</p> "> Figure 2
<p>Framework of the proposed method (<b>a</b>): extraction of implicit preferences within conversation; (<b>b</b>): quantifying extracted preferences.</p> "> Figure 3
<p>Prompt for the extraction of implicit preferences within conversation.</p> "> Figure 4
<p>Architecture for multi-label classification model.</p> "> Figure 5
<p>Prompt for movie recommendations.</p> "> Figure 6
<p>Example for the overview of experiments with <math display="inline"><semantics> <mrow> <msup> <mrow> <mi>C</mi> <mi>o</mi> <mi>n</mi> <mi>v</mi> </mrow> <mrow> <mo>+</mo> <mi>U</mi> <mi>i</mi> </mrow> </msup> </mrow> </semantics></math>.</p> ">
Abstract
:1. Introduction
- We propose a method to explicitly extract implicit preferences within conversations.
- We further suggest using the extracted preferences to design a multi-label model that quantifies categorical data.
- We conduct comparative experiments using GPT models with a large number of parameters and open-source LLM models with relatively fewer parameters to evaluate the effectiveness of the proposed approach.
- We demonstrate, through experimental results, that our proposed approach significantly enhances the performance of CRSs.
2. Related Work
3. Proposed Approach
3.1. Extraction of Implicit Preferences Within a Conversation
3.2. Quantifying Extracted Preferences
3.3. Recommendation Process
Algorithm 1 Extracting Implicit Preferences |
Require: |
Conv = Conversations |
Prm1 = Prompt for extracting implicit preferences |
Prm2 = Prompt for recommending item |
LLM(input, Prm) = The output produced by the prompt and input through the LLM |
Train(D, L) = Train a multi-label classifier using D as input and L as labels |
Pred(D) = Predict user preferences using a trained multi-label classifier with D as input |
1: function extractImplicitPreferences(Conv, Prm1) |
2: return LLM(Conv, Prm1) |
3: end function |
4: function trainMultiLabelClassifier(Conv, Prm1) |
5: Uc = list |
6: for c in Conv do |
7: Uc.append(extractImplicitPreferences(c, Prm1)) |
8: Train(Conv, Uc) |
9: end for |
10: end function |
11: function recommedItem(Conv, Prm2) |
12: Reclist = list |
13: for c in Conv do |
14: Conv2 = c + Pred(c) |
15: Reclist.append(LLM(Conv2, Prm2)) |
16: end for |
17: return Reclist |
18: end function |
- Extracting Implicit Preferences: The function extractImplicitPreferences (lines 1–3) uses a prompt-based approach to infer user preferences implicitly embedded in conversations. By passing conversation data (Conv) and a predefined prompt (Prm1) to the LLM, the algorithm extracts nuanced user preferences that are not explicitly stated. This step enables the system to effectively handle implicit information that is often present in real-world dialogues.
- Training the Multi-Label Classifier: The function trainMultiLabelClassifier (lines 4–10) builds a multi-label classifier to predict user preferences. For each conversation in Conv, implicit preferences are extracted using extractImplicitPreferences, and the resulting user preference labels (Uc) are collected. These labelled preferences are then used to train the classifier (Train), creating a model that can predict user preferences based on conversational data. This step enhances the ability of the system to generalize across various conversations.
- Generating Recommendations: The function recommendItem (lines 11–18) generates personalized recommendations by combining the predicted user preferences with the original conversations. For each conversation, the algorithm appends the predicted preferences (Pred(c)) to the conversation data (c) to form an enriched input (Conv2). This augmented input is then passed through the LLM, along with a recommendation prompt (Prm2), to produce a list of recommendations. The final recommendation list (Reclist) is returned as output.
4. Experiments
4.1. Experimental Setup
4.1.1. Datasets and Evaluation Metrics
- Recall@K: Recall@K is a metric used to evaluate the fraction of relevant items successfully retrieved from the top K recommendations. |Relevant Items| represents the total number of relevant items for a given query or user. Retrieved Items@K| denotes the number of items retrieved from the top K positions. The formula for recall @K is given by
- NDCG@K is a measure of the ranking quality that considers the position of relevant items on the recommended list, where higher-ranked relevant items contribute more to the score. indicates the relevance score of an item at position i in the retrieved list. This score is often binary (e.g., 1 if relevant and 0 otherwise). Moreover, i indicates the position index of the ranked list. K refers to the cutoff rank position, up to which the evaluation is performed. IDCG@K is the ideal DCG@K, which represents the maximum possible DCG score up to position K and is used for normalization. The formula for NDCG@K is as follows:
- MRR@K is the average reciprocal rank of the first relevant items in the top K recommendations. This provides an insight into how quickly the first relevant item is retrieved. |Q| denotes the total number of queries and users. indicates the rank position of the first relevant item in the top K results for the i-th query. The formula for MRR@K is as follows:
4.1.2. Baselines
4.1.3. Implementation Details
4.2. Experimental Results and Analysis
4.2.1. Comparison with Existing Studies
4.2.2. Impact of Dataset Characteristics on Performance
4.2.3. Strengths of LLMs in Limited-Data Environments
4.2.4. Implications for Real-World Applications
5. Conclusions
- Key Contributions
- a.
- Implicit preference modelling: Our study introduces a methodology for systematically extracting implicit preferences from conversational data using LLMs. By tailoring prompts to capture nuanced user intentions, our approach identifies preferences that are often overlooked by traditional CRS methods.
- b.
- Quantification of preferences: We propose a multi-label classification framework that quantifies these preferences and transforms categorical data into numerical representations. This quantification enables precise modelling of user preferences and enhances the ability to generate personalized recommendations.
- c.
- Comprehensive evaluation: Through rigorous experiments with GPT-3.5-turbo, GPT-4, LLaMA, and Mistral, our approach achieved a 23.3% improvement in Recall@20 on the ReDial dataset and consistent performance gains across all datasets. This evaluation highlights the scalability and adaptability of the proposed methodology to datasets with diverse user preferences.
- d.
- Dataset-specific insights: Analysis of the dataset characteristics reveals that larger, more diverse datasets, such as ReDial, benefit most from our approach, whereas smaller datasets, such as Inspired, demonstrate modest gains, emphasizing the adaptability of our framework to varying data conditions.
- e.
- LLM effectiveness: Our findings confirm the superior performance of GPT-4 over GPT-3.5-turbo, particularly in scenarios requiring deeper contextual understanding and preference modelling, underscoring the critical role of advanced LLMs in CRS. Furthermore, the advantage of larger-parameter GPT models is evident as they outperform open-source models with fewer parameters.
- Key Findings
- a.
- Performance gains: The integration of quantified user preferences consistently outperformed the reliance on raw conversational data and improved the ranking metrics across diverse datasets.
- b.
- Dataset-dependent performance: Larger datasets with richer user interactions exhibited significant performance gains, whereas smaller datasets benefited less, highlighting the need for dataset-specific adaptations.
- c.
- Scalability challenges: Although effective, the computational overhead of integrating LLMs and classification frameworks poses challenges for real-time applications, particularly in resource-constrained environments.
- d.
- Versatility of advanced LLMs: GPT-4 outperformed GPT-3.5-turbo and open-source models in all experiments, showing its capability to understand nuanced user intents and improve recommendation outcomes.
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
References
- Christakopoulou, K.; Radlinski, F.; Hofmann, K. Towards conversational recommender systems. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (SIGKDD’16), San Francisco, CA, USA, 13–17 August 2016. [Google Scholar]
- Sun, Y.; Zhang, Y. Conversational recommender system. In Proceedings of the 41st International ACM SIGIR Conference on Research & Development in Information Retrieval (SIGIR’18), Ann Arbor, MI, USA, 8–12 July 2018; pp. 235–244. [Google Scholar]
- Zhang, G. User-Centric Conversational Recommendation: Adapting the Need of User with Large Language Models. In Proceedings of the 17th ACM Conference on Recommender Systems (RecSys’23), New York, NY, USA, 18–22 September 2023; pp. 1349–1354. [Google Scholar]
- Chang, Y.; Wang, X.; Wang, J.; Wu, Y.; Yang, L.; Zhu, K.; Chen, H.; Yi, X.; Wang, C.; Wang, Y.; et al. A Survey on Evaluation of Large Language Models. ACM Trans. Intell. Syst. Technol. 2023, 15, 1–45. [Google Scholar] [CrossRef]
- Wu, L.; Zheng, Z.; Qiu, Z.; Wang, H.; Gu, H.; Shen, T.; Qin, C.; Zhu, C.; Zhu, H.; Liu, Q.; et al. A Survey on Large Language Models for Recommendation. arXiv 2023, arXiv:2305.19860. [Google Scholar] [CrossRef]
- Wei, W.; Ren, X.; Tang, J.; Wang, Q.; Su, L.; Cheng, S.; Wang, J.; Yin, D.; Huang, C. LLMRec: Large Language Models with Graph Augmentation for Recommendation. In Proceedings of the 17th ACM International Conference on Web Search and Data Mining (WSDM’24), Merida, Mexico, 4–8 March 2024; pp. 806–815. [Google Scholar]
- He, Z.; Xie, Z.; Jha, R.; Steck, H.; Liang, D.; Feng, Y.; Majumder, B.P.; Kallus, N.; McAuley, J. Large Language Models as Zero-Shot Conversational Recommenders. In Proceedings of the 32nd ACM International Conference on Information and Knowledge Management (CIKM’23), New York, NY, USA, 21–25 October 2023; pp. 720–730. [Google Scholar]
- Farshidi, S.; Rezaee, K.; Mazaheri, S.; Rahimi, A.H.; Dadashzadeh, A.; Ziabakhsh, M.; Eskandari, S.; Jansen, S. Understanding User Intent Modeling for Conversational Recommender Systems: A Systematic Literature Review. User Model. User-Adapt. Interact. 2024, 34, 1643–1706. [Google Scholar] [CrossRef]
- Zou, L.; Xia, L.; Du, P.; Zhang, Z.; Bai, T.; Liu, W.; Nie, J.; Yin, D. Pseudo Dyna-Q: A reinforcement learning framework for interactive recommendation. In Proceedings of the 13th International Conference on Web Search and Data Mining (WSDM’20), Houston, TX, USA, 3–7 January 2020; pp. 816–824. [Google Scholar]
- Deng, Y.; Li, Y.; Sun, F.; Ding, B.; Lam, W. Unified Conversational Recommendation Policy Learning via Graph-Based Reinforcement Learning. In Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR’21), New York, NY, USA, 11–15 July 2021; pp. 1431–1441. [Google Scholar]
- Yang, F.; Chen, Z.; Jiang, Z.; Cho, E.; Huang, X.; Lu, Y. PALR: Personalization Aware LLMs for Recommendation. arXiv 2023, arXiv:2305.07622. [Google Scholar]
- Sanner, S.; Balog, K.; Radlinski, F.; Wedin, B.; Dixon, L. Large Language Models are Competitive Near Cold-start Recommenders for Language-and Item-based Preferences. In Proceedings of the 17th ACM Conference on Recommender Systems, Singapore, 18–22 September 2023; pp. 890–896. [Google Scholar]
- Agrawal, S.; Trenkle, J.; Kawale, J. Beyond Labels: Leveraging Deep Learning and LLMs for Content Metadata. In Proceedings of the 17th ACM Conference on Recommender Systems (RecSys’23), New York, NY, USA, 18–22 September 2023; pp. 74–77. [Google Scholar]
- Zhao, Z.; Fan, W.; Li, J.; Liu, Y.; Mei, X.; Wang, Y.; Li, Q. Recommender Systems in the Era of Large Language Models (LLMs). arXiv 2023, arXiv:2307.02046. [Google Scholar] [CrossRef]
- Xiao, Y.; Zhang, H.; Liu, Y.; Sun, F.; Zhou, K. Enhancing Conversational Recommendation Systems with User and Item Embeddings. IEEE Trans. Knowl. Data Eng. 2023, 35, 3412–3426. [Google Scholar]
- Chen, Y.; Yang, F.; Wang, Y.; Wu, L.; Sun, Q. An Investigation into the Role of Context in CRS: A Multi-Contextual Approach. Inf. Sci. 2024, 658, 256–274. [Google Scholar]
- Mei, Z.; Huang, R.; Zhao, W.; Gao, Z. Towards Multi-turn Dialogue Consistency: Leveraging Contextual Cues in CRS. Expert Syst. Appl. 2023, 207, 117935. [Google Scholar]
- Vinyals, O.; Le, Q. A Neural Conversational Model. In Proceedings of the 32nd International Conference on Machine Learning (ICML), Lille, France, 6–11 July 2015; pp. 174–182. [Google Scholar]
- Ghazvininejad, M.; Brockett, C.; Chang, M.W.; Dolan, B.; Gao, J.; Yih, W.; Galley, M. Knowledge-Grounded Neural Conversation Models. In Proceedings of the 32nd AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018. [Google Scholar]
- Zhou, K.; Zhou, Y.; Song, Y.; Zhang, W. Towards Conversational Recommender Systems via User Feedback-Aware Reinforcement Learning. In Proceedings of the 29th ACM International Conference on Information and Knowledge Management (CIKM), Virtual, 19–23 October 2020; pp. 1445–1454. [Google Scholar]
- Hayati, S.A.; Kang, D.; Zhu, Q.; Shi, W.; Yu, Z. INSPIRED: Toward Sociable Recommendation Dialog Systems. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP 2020), Online, 16–20 November 2020; pp. 3618–3632. [Google Scholar]
- Li, R.; Kahou, S.E.; Schulz, H.; Michalski, V.; Charlin, L.; Pal, C.J. Towards Deep Conversational Recommendations. In Proceedings of the 32nd International Conference on Neural Information Processing Systems (NIPS’18), Montréal, QC, Canada, 3–8 December 2018. [Google Scholar]
- Radford, A.; Narasimhan, K.; Salimans, T.; Sutskever, I. Improving Language Understanding by Generative Pretraining; OpenAI Technical Report. 2018. Available online: https://cdn.openai.com/research-covers/language-unsupervised/language_understanding_paper.pdf (accessed on 8 January 2025).
- Joulin, A.; Grave, E.; Bojanowski, P.; Mikolov, T. Bag of Tricks for Efficient Text Classification. In Proceedings of the 15th Conference of the European Chapter of the Association for Computational Linguistics, Valencia, Spain, 3–7 April 2017; pp. 427–431. [Google Scholar]
- Mikolov, T.; Chen, K.; Corrado, G.; Dean, J. Efficient Estimation of Word Representations in Vector Space. In Proceedings of the International Conference on Learning Representations, Scottsdale, AZ, USA, 2–4 May 2013. [Google Scholar]
- Silver, D.; Huang, A.; Maddison, C.J.; Guez, A.; Sifre, L.; van den Driessche, G.; Schrittwieser, J.; Antonoglou, I.; Panneershelvam, V.; Lanctot, M.; et al. Mastering the game of Go with deep neural networks and tree search. Nature 2016, 529, 484–489. [Google Scholar] [CrossRef] [PubMed]
- Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is All You Need. In Proceedings of the 31st Annual Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; pp. 5998–6008.1. [Google Scholar]
- Smith, A.; Brown, J.; Taylor, P. Enhancing Conversational Recommendation Systems with Sequence-to-Sequence Learning. J. Artif. Intell. Res. 2024, 73, 1045–1062. [Google Scholar]
- Kim, H.; Park, S.; Lee, J. A Transformer-Based Approach for Context-Aware Conversational Recommenders. Expert Syst. Appl. 2024, 215, 119678. [Google Scholar]
- Nguyen, M.; Tran, D.; Pham, T. Multi-modal Learning for CRS: Leveraging Visual and Textual Data. IEEE Trans. Multimed. 2023, 26, 455–470. [Google Scholar]
- Park, J.; Lee, H.; Choi, S. Domain-Specific Conversational Recommender Systems with LLMs: Case Studies in Healthcare and Education. Inf. Process. Manag. 2023, 61, 107489. [Google Scholar]
- Lee, K.; Kim, J.; Yoon, S. Real-Time Recommendation Optimization in CRS Using Reinforcement Learning and LLMs. ACM Trans. Recomm. Syst. 2024, 12, 567–583. [Google Scholar]
- Gao, Y.; Sheng, T.; Xiang, Y.; Xiong, Y.; Wang, H.; Zhang, J. Chat-Rec: Towards Interactive and Explainable LLMs-Augmented Recommender Systems. arXiv 2023, arXiv:2303.14524. [Google Scholar]
- Hou, Y.; Zhang, J.; Lin, Z.; Lu, H.; Xie, R.; McAuley, J.; Zhao, W.X. Large Language Models are Zero-Shot Rankers for Recommender Systems. In Proceedings of the Advances in Information Retrieval: 46th European Conference on Information Retrieval, ECIR 2024, Glasgow, UK, 24–28 March 2024; Proceedings, Part II. Springer: Berlin/Heidelberg, Germany, 2024; pp. 364–381. [Google Scholar]
- Wang, L.; Lim, E.-P. Zero-Shot Next-Item Recommendation using Large Pretrained Language Models. arXiv 2023, arXiv:2304.03153. [Google Scholar]
- Bao, K.; Zhang, J.; Zhang, Y.; Wang, W.; Feng, F.; He, X. TALLRec: An Effective and Efficient Tuning Framework to Align Large Language Model with Recommendation. In Proceedings of the 17th ACM Conference on Recommender Systems (RecSys’23), Singapore, 18–22 September 2023; Association for Computing Machinery: New York, NY, USA, 2023; pp. 1007–1014. [Google Scholar]
- Kang, W.-C.; Ni, J.; Mehta, N.; Sathiamoorthy, M.; Hong, L.; Chi, E.; Cheng, D.Z. Do LLMs Understand User Preferences? Evaluating LLMs On User Rating Prediction. arXiv 2023, arXiv:2305.06474. [Google Scholar]
- Zhang, J.; Xie, R.; Hou, Y.; Zhao, W.X.; Lin, L.; Wen, J.-R. Recommendation as Instruction Following: A Large Language Model Empowered Recommendation Approach. arXiv 2023, arXiv:2305.07001. [Google Scholar] [CrossRef]
Research Type | Related Research |
---|---|
Transformer-based models | Wei et al. [6] introduced LLMRec, which integrates transformer models with graph-based embeddings to better represent user–item interactions, considerably enhancing recommendation relevance and scalability. |
He et al. [7] proposed a zero-shot conversational recommender system that utilizes pretrained LLMs to provide recommendations without requiring domain-specific training. Their approach demonstrated the adaptability of LLMs in diverse scenarios. | |
Zhao et al. [14] explored the application of LLMs in cold-start scenarios by leveraging cross-domain knowledge transfer and attention mechanisms to improve system robustness and user satisfaction. | |
Ghazvininejad et al. [19] introduced knowledge-grounded neural models to improve the ability of CRSs to handle open-domain queries, thereby demonstrating superior dialogue coherence and relevance in recommendations. | |
Vaswani et al. [27] introduced transformer architectures, namely “Attention is All You Need”, which laid the foundation for many neural CRS advancements, enabling improved handling of long-term dependencies and contextual understanding. | |
Smith et al. [28] explored sequence-to-sequence approaches to optimize conversational understanding and real-time recommendation generation and demonstrated improved user satisfaction. | |
Feature-driven advancements | Deng et al. [10] proposed a framework that combines conversational context with graph-based reinforcement learning, enabling systems to deliver personalized and contextually relevant recommendations. |
Mei et al. [17] focused on leveraging contextual cues from user interactions to achieve consistency in multi-turn dialogues, ensuring more coherent and satisfying user experiences. | |
Li et al. [28] explored the integration of temporal and spatial signals into the CRS for dynamic context adaptation, improving the recommendation precision over extended user interactions. | |
Nguyen et al. [30] demonstrated the utility of combining visual and textual data to enhance recommendation diversity and richness, leveraging the ability of CRSs to utilize multimodal information. | |
Kim et al. [29] explored transformer models to integrate temporal and contextual user signals to enhance the precision and relevance of conversational recommendations. | |
CRS with LLMs | Advanced context understanding: LLMs enable CRSs to process complex conversations and identify subtle emotional shifts and dynamic user needs [28]. |
Personalized interaction modelling: Studies have shown that integrating real-time user feedback with LLMs can significantly enhance recommendation accuracy and user satisfaction [29]. | |
Cold-start solutions: LLMs effectively address the cold-start problem by analyzing limited textual metadata and generating recommendations for new users or items [30]. |
Dataset | Inspired | Description | Comprising 1001 human-to-human dialogues annotated with sociable recommendation strategies, this dataset emphasizes social science-informed conversational strategies. |
Preprocessing Steps | -Verified the completeness of annotated fields and excluded dialogues with missing or ambiguous sociable strategy labels. -Balanced the dataset by augmenting underrepresented sociable strategies using data augmentation techniques (e.g., paraphrasing. -Removed duplicated dialogues to avoid bias during training and evaluation. | ||
Purpose | These steps were required to ensure fair representation of all sociable strategies and maintain the focus of the dataset on engaging conversational styles. | ||
ReDial | Description | This dataset contains over 10,000 human-to-human dialogues collected via Amazon Mechanical Turk, where participants were tasked with recommending movies to each other. | |
Preprocessing Steps | -Identified and corrected incomplete dialogues or those with placeholder text (e.g., “N/A”). -Handled missing metadata (e.g., movie titles, genres) by cross-referencing external movie databases (e.g., IMDb). -Standardized dialogue formatting to facilitate consistent parsing during training. | ||
Purpose | Minor inconsistencies were addressed to improve the quality of the dataset and maintain its integrity for model training and evaluation. | ||
Reddit-movie | Description | Derived from real user conversations on Reddit, the dataset captures users’ natural expressions of preferences and personalized tendencies. | |
Preprocessing Steps | -Removed irrelevant or off-topic threads that did not pertain to movie recommendations. -Filtered out conversations with less than three turns to ensure meaningful dialogue structure. -Normalized text by removing URLs, emojis, and special characters to maintain focus on the conversational content. | ||
Purpose | These steps were essential to clean noise and enhance the utility of the dataset for conversational recommendation tasks. |
Dataset | Dialogs | Turns | Items |
---|---|---|---|
Inspired | 825 | 2051 | 1548 |
ReDial | 2311 | 9913 | 4216 |
Reddit-Movie | 8413 | 9410 | 6504 |
Dataset | Models | Methods | recall@1 | ndcg@1 | mrr@1 | recall@5 | ndcg@5 | mrr@5 | recall@10 | ndcg@10 | mrr@10 | recall@20 | ndcg@20 | mrr@20 |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
Reddit-movie | GPT-3.5-turbo | Conv | 0.019955 | 0.019955 | 0.019955 | 0.070381 | 0.045299 | 0.037097 | 0.10662 | 0.05702 | 0.041931 | 0.13744 | 0.06493 | 0.044169 |
0.02079 | 0.02079 | 0.02079 | 0.07248 | 0.04678 | 0.03837 | 0.103248 | 0.056663 | 0.04241 | 0.129659 | 0.06345 | 0.04433 | |||
0.019906 | 0.019906 | 0.019906 | 0.069305 | 0.044663 | 0.03661 | 0.105497 | 0.056364 | 0.041437 | 0.135968 | 0.064197 | 0.043653 | |||
GPT-4.0 | Conv | 0.019906 | 0.019906 | 0.019906 | 0.069305 | 0.044646 | 0.03659 | 0.105497 | 0.056349 | 0.041417 | 0.136115 | 0.064219 | 0.043644 | |
0.019906 | 0.019906 | 0.019906 | 0.06994 | 0.045015 | 0.036868 | 0.106035 | 0.056685 | 0.041681 | 0.136604 | 0.064541 | 0.043904 | |||
0.01996 | 0.01996 | 0.01996 | 0.06994 | 0.04502 | 0.03688 | 0.10608 | 0.05671 | 0.0417 | 0.1369 | 0.06463 | 0.04394 | |||
LLaMa3 | Conv | 0.009195 | 0.009195 | 0.009195 | 0.031155 | 0.020187 | 0.016602 | 0.047736 | 0.025502 | 0.018769 | 0.0583 | 0.028241 | 0.019556 | |
0.01614 | 0.01614 | 0.01614 | 0.053067 | 0.034484 | 0.028423 | 0.07625 | 0.041993 | 0.031527 | 0.106622 | 0.049613 | 0.033587 | |||
0.017656 | 0.017656 | 0.017656 | 0.06011 | 0.039 | 0.03209 | 0.087499 | 0.047832 | 0.035719 | 0.109117 | 0.053388 | 0.037292 | |||
Mistral | Conv | 0.010369 | 0.010369 | 0.010369 | 0.03541 | 0.022933 | 0.018851 | 0.05248 | 0.028437 | 0.021114 | 0.065783 | 0.031893 | 0.02211 | |
0.01433 | 0.01433 | 0.01433 | 0.045437 | 0.029834 | 0.024743 | 0.070625 | 0.037933 | 0.028058 | 0.095569 | 0.044258 | 0.029805 | |||
0.015064 | 0.015064 | 0.015064 | 0.048763 | 0.031875 | 0.026361 | 0.073706 | 0.039938 | 0.029687 | 0.090482 | 0.044245 | 0.030902 | |||
Inspired | GPT-3.5-turbo | Conv | 0.03968 | 0.03968 | 0.03968 | 0.09898 | 0.07007 | 0.06057 | 0.13012 | 0.08018 | 0.06477 | 0.14377 | 0.08375 | 0.0658 |
0.033703 | 0.033703 | 0.033703 | 0.081911 | 0.058672 | 0.051017 | 0.107935 | 0.067122 | 0.054525 | 0.118601 | 0.069945 | 0.055367 | |||
0.034556 | 0.034556 | 0.034556 | 0.090444 | 0.06344 | 0.054536 | 0.122014 | 0.073664 | 0.058766 | 0.134812 | 0.077067 | 0.059788 | |||
GPT-4.0 | Conv | 0.05247 | 0.05247 | 0.05247 | 0.1186 | 0.08676 | 0.07627 | 0.15102 | 0.09742 | 0.08077 | 0.17065 | 0.10265 | 0.08235 | |
0.045648 | 0.045648 | 0.045648 | 0.102816 | 0.075083 | 0.06597 | 0.133532 | 0.08509 | 0.070145 | 0.146331 | 0.088477 | 0.071154 | |||
0.047782 | 0.047782 | 0.047782 | 0.108788 | 0.079623 | 0.069994 | 0.138652 | 0.089288 | 0.07399 | 0.153157 | 0.093116 | 0.075125 | |||
LLaMa3 | Conv | 0.020478 | 0.020478 | 0.020478 | 0.063567 | 0.043722 | 0.037102 | 0.085751 | 0.050805 | 0.039972 | 0.119454 | 0.059316 | 0.042305 | |
0.025597 | 0.025597 | 0.025597 | 0.069113 | 0.048126 | 0.041197 | 0.091724 | 0.055468 | 0.044249 | 0.099829 | 0.057622 | 0.044894 | |||
0.022611 | 0.022611 | 0.022611 | 0.062713 | 0.042895 | 0.036391 | 0.086604 | 0.050717 | 0.039676 | 0.09215 | 0.052195 | 0.040123 | |||
Mistral | Conv | 0.030717 | 0.030717 | 0.030717 | 0.066126 | 0.048447 | 0.042662 | 0.09215 | 0.056922 | 0.0462 | 0.115188 | 0.062892 | 0.047917 | |
0.018771 | 0.018771 | 0.018771 | 0.056314 | 0.037565 | 0.031421 | 0.075085 | 0.043517 | 0.03380 | 0.091297 | 0.047621 | 0.034938 | |||
0.02901 | 0.02901 | 0.02901 | 0.069539 | 0.049799 | 0.043309 | 0.102816 | 0.060441 | 0.047633 | 0.133959 | 0.068247 | 0.049741 | |||
ReDial | GPT-3.5-turbo | Conv | 0.035445 | 0.035445 | 0.035445 | 0.099548 | 0.068206 | 0.057902 | 0.139744 | 0.081122 | 0.063185 | 0.176923 | 0.090731 | 0.065931 |
0.028431 | 0.028431 | 0.028431 | 0.085143 | 0.056976 | 0.047754 | 0.122851 | 0.069061 | 0.052679 | 0.160483 | 0.078741 | 0.05542 | |||
0.04261 | 0.04261 | 0.04261 | 0.11908 | 0.0814 | 0.06903 | 0.1687 | 0.09737 | 0.07558 | 0.21456 | 0.10908 | 0.07885 | |||
GPT-4.0 | Conv | 0.038688 | 0.038688 | 0.038688 | 0.11546 | 0.077825 | 0.065476 | 0.161011 | 0.092539 | 0.071537 | 0.213876 | 0.106065 | 0.075329 | |
0.041101 | 0.041101 | 0.041101 | 0.108899 | 0.07555 | 0.064613 | 0.154827 | 0.090354 | 0.070692 | 0.204299 | 0.102952 | 0.074195 | |||
0.04796 | 0.04796 | 0.04796 | 0.12624 | 0.08813 | 0.07559 | 0.181 | 0.10568 | 0.08273 | 0.23341 | 0.119 | 0.08643 | |||
LlaMA3 | Conv | 0.024133 | 0.024133 | 0.024133 | 0.066139 | 0.045536 | 0.038766 | 0.094872 | 0.054811 | 0.042583 | 0.121041 | 0.061549 | 0.044497 | |
0.029412 | 0.029412 | 0.029412 | 0.103469 | 0.066754 | 0.05473 | 0.152036 | 0.082543 | 0.061295 | 0.192006 | 0.092717 | 0.064121 | |||
0.031222 | 0.031222 | 0.031222 | 0.094646 | 0.063353 | 0.053089 | 0.139216 | 0.077736 | 0.059009 | 0.186652 | 0.089775 | 0.062338 | |||
Mistral | Conv | 0.027602 | 0.027602 | 0.027602 | 0.081297 | 0.054711 | 0.046008 | 0.112293 | 0.064774 | 0.050184 | 0.146983 | 0.073663 | 0.052685 | |
0.029035 | 0.029035 | 0.029035 | 0.083635 | 0.056307 | 0.047382 | 0.116817 | 0.067022 | 0.051797 | 0.147059 | 0.074707 | 0.053928 | |||
0.032881 | 0.032881 | 0.032881 | 0.097662 | 0.06513 | 0.054516 | 0.138537 | 0.078418 | 0.060041 | 0.175415 | 0.087761 | 0.062615 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Kim, W.-S.; Lim, S.; Kim, G.-W.; Choi, S.-M. Extracting Implicit User Preferences in Conversational Recommender Systems Using Large Language Models. Mathematics 2025, 13, 221. https://doi.org/10.3390/math13020221
Kim W-S, Lim S, Kim G-W, Choi S-M. Extracting Implicit User Preferences in Conversational Recommender Systems Using Large Language Models. Mathematics. 2025; 13(2):221. https://doi.org/10.3390/math13020221
Chicago/Turabian StyleKim, Woo-Seok, Seongho Lim, Gun-Woo Kim, and Sang-Min Choi. 2025. "Extracting Implicit User Preferences in Conversational Recommender Systems Using Large Language Models" Mathematics 13, no. 2: 221. https://doi.org/10.3390/math13020221
APA StyleKim, W.-S., Lim, S., Kim, G.-W., & Choi, S.-M. (2025). Extracting Implicit User Preferences in Conversational Recommender Systems Using Large Language Models. Mathematics, 13(2), 221. https://doi.org/10.3390/math13020221