Abstract

With more than 2 billion viewers per month, YouTube is the most widely used video-sharing website worldwide. On this website, users can watch, upload, and share videos covering a wide range of subjects. YouTube comments include facts, opinions, and responses to videos in addition to starting discussions. The number of YouTube comments makes it difficult to manually analyze them all. The study of reading, understanding, and creating text in human languages encompasses a broad range of methods and techniques under the umbrella of natural language processing or NLP. The primary goal of the research is to find and analyze YouTube comments, which, when used with natural language processing algorithms, might be beneficial for the channels' continued development. One of the NLP methods used for this research was tokenization, which is used to break down text into individual words or tokens. Stemming and lemmatization are used to reduce the root words and normalize the variation. Categorization is performed by identifying the named entities, such as people, organizations, and locations. The machine translation was used to convert the comments from one language to another.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call