Abstract. The multi-armed bandit problem has been extensively studied and applied in various fields such as recommendation systems and resource allocation. However, there is still a lack of research in the movie recommendation domain, such as the comparison of algorithm performance and the impact of movie text information. This paper uses the Movielens 100K movie review data set to compare and analyze the performance of classic algorithms-Greedy, UCB, and TS-in the multi-armed bandit problem. Additionally, to explore the influence of movie text information (such as movie summaries and user reviews) on system algorithms, this experiment utilizes web scraping technology through HTTP requests and HTML parsing, to extract data from the Douban website, obtaining the information and forming a new data set. In this movie review data set, the Greedy algorithm outperforms both UCB and TS algorithms. This study finds that incorporating movie text information improves the accuracy of all three algorithms, thus contributing to the enhancement of movie recommendation systems. This experiment did not consider larger datasets or movie image information, which could be addressed in future improvements.
Read full abstract