Abstract
Sequential recommender models should capture evolving user preferences over time, but there is a risk of obtaining biased results such as false positives and false negatives due to noisy interactions. Generative models effectively learn the underlying distribution and uncertainty of the given data to generate new data, and they exhibit robustness against noise. In particular, utilizing the Diffusion model, which generates data through a multi-step process of adding and removing noise, enables stable and effective recommendations. The Diffusion model typically leverages a Gaussian distribution with a mean fixed at zero, but there is potential for performance improvement in generative models by employing distributions with higher degrees of freedom. Therefore, we propose a Diffusion model-based sequential recommender model that uses a new noise distribution. The proposed model improves performance through a Weibull distribution with two parameters determining shape and scale, a modified Transformer architecture based on Macaron Net, normalized loss, and a learning rate warmup strategy. Experimental results on four types of real-world e-commerce data show that the proposed model achieved performance gains ranging from a minimum of 2.53% to a maximum of 13.52% across HR@K and NDCG@K metrics compared to the existing Diffusion model-based sequential recommender model.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.