Authors:

Bohao WangFeng LiuJiawei ChenYudi WuXingyu LouJun WangYan FengChun ChenCan Wang

Paper:

https://arxiv.org/abs/2408.08208

Leveraging Large Language Models for Denoising Sequential Recommendations: An In-Depth Look at LLM4DSR

Introduction

Large Language Models (LLMs) have revolutionized artificial intelligence with their remarkable capabilities in content comprehension, generation, and semantic reasoning. This paper explores the innovative application of LLMs in denoising sequential recommendations, a task that relies heavily on the accuracy of users’ historical interaction sequences. These sequences often contain noise due to various factors, such as clickbait or accidental interactions, which can significantly degrade the performance of recommendation models. The proposed method, LLM4DSR, aims to leverage LLMs to identify and replace noisy interactions, thereby enhancing the accuracy and effectiveness of sequential recommendations.

Preliminary

Sequential Recommendation

Sequential recommendation systems predict the next item a user is likely to interact with based on their historical interaction sequence. Formally, let ( V = {v_1, v_2, \ldots, v_{|V|}} ) be the set of items, and ( S_i = {s_{i1}, s_{i2}, \ldots, s_{i|S_i|}} ) denote the (i)-th historical interaction sequence. The goal is to predict the next item ( s_{i(t+1)} ) given the sequence ( S_{i1:t} ).

Denoise Task Formulation

The objective of the denoising task is to identify and remove noise within a sequence ( S ), generating a noiseless sequence ( S’ = {s’1, s’_2, \ldots, s’{|S’|}} ). Unlike previous methods that simply delete noisy items, LLM4DSR replaces identified noise with more appropriate items, maintaining the sequence length and potentially introducing useful information.

Method

Training LLM as a Denoiser

To adapt pre-trained LLMs for denoising tasks, a self-supervised fine-tuning process is employed. An instruction dataset is constructed by replacing a proportion of items in sequences with randomly selected alternatives. The LLM is then fine-tuned to identify noisy items and suggest appropriate replacements.

Prompt for Self-Supervised Task:

  • Instruction: Identify items that do not align with the main interests reflected by the majority of the items and suggest alternatives.
  • Input: A list of items the user has interacted with.
  • Output: Noise items and suggested replacements.

Uncertainty Estimation

An uncertainty estimation module is introduced to assess the reliability of the identified noisy items. The probability of each item being noise is calculated using the LLM, and items with probabilities exceeding a threshold ( \eta ) are classified as noise.

Replace Noise with Suggested Items

Instead of simply removing noise, LLM4DSR replaces it with items that better match the user’s preferences. The LLM generates a suggested item, which is then aligned with the item set using grounding techniques to ensure it is a valid replacement.

Experiments

Experimental Settings

Three real-world datasets (Amazon Video Games, Amazon Toy and Games, and Amazon Movies) were used, with sequences segmented using a sliding window of length 11. Two noise settings were tested: natural noise in real-world datasets and artificial noise introduced by randomly replacing items in sequences.

Performance Comparison (RQ1)

LLM4DSR consistently outperformed existing state-of-the-art denoising methods across various datasets and backbones. Explicit denoising methods like STEAM and HSD showed diminished effectiveness when retrained on different backbones, while implicit methods like FMLP and CL4SRec exhibited unstable performance.

Ablation Study (RQ2)

An ablation study was conducted to verify the contributions of various components of LLM4DSR. The results highlighted the effectiveness of each component, particularly the replacement operation, which significantly improved performance by providing items that better align with user interests.

Case Study (RQ3)

A case study demonstrated how LLM4DSR leverages open knowledge to identify and replace noise. For example, in a user’s viewing history predominantly consisting of drama, romance, and comedy films, the LLM identified a science fiction film as noise and suggested a more appropriate replacement.

Related Work

Sequential Recommendation

Sequential recommendation systems predict the next item based on historical interactions, considering the temporal order to capture dynamic changes in user interests. Various deep learning models, such as RNNs, CNNs, and self-attention mechanisms, have been employed to enhance these systems.

Denoising Recommendation

Denoising methods aim to mitigate the impact of noisy data on recommendation systems. Implicit methods reduce noise influence without explicitly removing it, while explicit methods directly modify sequence content. LLM4DSR leverages the extensive open knowledge of LLMs to enhance denoising effectiveness.

LLMs for Recommendation

LLMs have been utilized in various recommendation tasks, including collaborative filtering, sequential recommendation, and CTR tasks. They have also been employed for explaining item embeddings, recommendation results, and conversational recommendations. LLM4DSR is the first to explore LLMs for sequence recommendation denoising.

Conclusion

LLM4DSR is a novel method for sequence recommendation denoising using LLMs. It leverages the extensive open knowledge embedded in LLMs to identify and replace noisy interactions, significantly enhancing recommendation performance. Future research could focus on transferring the denoising capabilities of LLMs to smaller models to reduce computational overhead.

Ethical Considerations

Denoising methods transfer the model’s prior knowledge to the data, potentially amplifying biases. It is crucial to ensure that the outputs of pre-trained LLMs are inherently fair and to conduct bias testing when selecting the backbone model.

Share.

Comments are closed.

Exit mobile version