Abstract

The volume of generated and stored data from social media has increased in the last decade. Therefore, analyzing and understanding this kind of data can offer relevant information in different contexts and can assist researchers and companies in the decision-making process. However, the data are scattered in a large volume, come from different sources, with different formats and are rapidly created. Such facts make the knowledge extraction difficult, turning it in a complex and high costly process. The scientific contribution of this paper is the development of a social media data integration model based on a data warehouse to reduce the computational costs related to data analysis, as well as support the application of techniques to discover useful knowledge. Differently from the literature, we focus on both social media Facebook and Twitter. Also, we contribute with the proposition of a model for the acquisition, transformation and loading data, which can enable the extraction of useful knowledge in a context where the human capability of understanding is exceeded. The results showed that the proposed data warehouse improves the quality of data mining algorithms compared to related works, while being able to reduce the execution time.

Highlights

  • In the last few years, the amount of data produced in the internet with the advent of web 2.0 technology has increased, especially the data from social media environment (Ghani et al, 2018)

  • This paper describes its definitions for building the schema and the opinion analysis, which can be positive, negative or neutral (Balazs and Velásquez, 2016)

  • The big data is defined as large data set with no pattern that exceeds the human capacity for understanding

Read more

Summary

Introduction

In the last few years, the amount of data produced in the internet with the advent of web 2.0 technology has increased, especially the data from social media environment (Ghani et al, 2018). This had a significant matter in contemporary society due to the ease of sharing and helping communication among people. The big data is defined as large data set with no pattern that exceeds the human capacity for understanding Such property is referenced in computer science nowadays, especially due to its potential in decision-making process and discovering trends and associations (Sivarajah et al, 2017). A brief explanation of principals is: Volume - is given by its magnitude and its size that can be terabytes, petabytes or exabytes

Objectives
Methods
Results
Discussion
Conclusion
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call