Time-synchronized sentiment labeling via autonomous online comments data mining: A multimodal information fusion on large-scale multimedia data

  • Abstract
  • References
  • Similar Papers
Abstract
Translate article icon Translate Article Star icon
Take notes icon Take Notes

Time-synchronized sentiment labeling via autonomous online comments data mining: A multimodal information fusion on large-scale multimedia data

ReferencesShowing 10 of 36 papers
  • Cite Count Icon 503
  • 10.2307/2185261
What an Emotion is: A Sketch
  • Apr 1, 1988
  • The Philosophical Review
  • Robert C Roberts

  • Cite Count Icon 26
  • 10.1016/j.bdr.2021.100189
Social Media Data and Users' Preferences: A Statistical Analysis to Support Marketing Communication
  • Jan 7, 2021
  • Big Data Research
  • Elisa Arrigo + 2 more

  • Open Access Icon
  • Cite Count Icon 149
  • 10.1109/access.2018.2851311
Sentiment Analysis of Big Data: Methods, Applications, and Open Challenges
  • Jan 1, 2018
  • IEEE Access
  • Shahid Shayaa + 7 more

  • Open Access Icon
  • Cite Count Icon 728
  • 10.1007/s10462-022-10144-1
A survey on sentiment analysis methods, applications, and challenges
  • Feb 7, 2022
  • Artificial Intelligence Review
  • Mayur Wankhade + 2 more

  • Open Access Icon
  • Cite Count Icon 1162
  • 10.1016/j.inffus.2017.02.003
A review of affective computing: From unimodal analysis to multimodal fusion
  • Feb 3, 2017
  • Information Fusion
  • Soujanya Poria + 3 more

  • Open Access Icon
  • PDF Download Icon
  • Cite Count Icon 57
  • 10.1109/access.2020.3002215
The Applications of Sentiment Analysis for Russian Language Texts: Current Challenges and Future Perspectives
  • Jan 1, 2020
  • IEEE Access
  • Sergey Smetanin

  • Cite Count Icon 6
  • 10.1016/j.bdr.2020.100180
Search History Visualization for Collaborative Web Searching
  • Dec 17, 2020
  • Big Data Research
  • Luyan Xu + 2 more

  • Cite Count Icon 114
  • 10.1108/jrim-05-2017-0030
Online sentiment analysis in marketing research: a review
  • Jan 31, 2018
  • Journal of Research in Interactive Marketing
  • Meena Rambocas + 1 more

  • Cite Count Icon 1470
  • 10.1016/b978-0-12-558701-3.50007-7
Chapter 1 - A GENERAL PSYCHOEVOLUTIONARY THEORY OF EMOTION
  • Jan 1, 1980
  • Theories of Emotion
  • Robert Plutchik

  • Cite Count Icon 1488
  • 10.4324/9781315798868
Imagery and Verbal Processes
  • Nov 26, 2013
  • A Paivio

Similar Papers
  • Research Article
  • Cite Count Icon 1
  • 10.1016/j.measurement.2024.115728
A Brain-Inspired Decision-Making method for upper limb exoskeleton based on Multi-Brain-Region structure and multimodal information fusion
  • Sep 11, 2024
  • Measurement
  • Wendong Wang + 4 more

A Brain-Inspired Decision-Making method for upper limb exoskeleton based on Multi-Brain-Region structure and multimodal information fusion

  • Research Article
  • Cite Count Icon 15
  • 10.1016/j.cmpb.2021.106451
Information fusion and multi-classifier system for miner fatigue recognition in plateau environments based on electrocardiography and electromyography signals
  • Oct 2, 2021
  • Computer Methods and Programs in Biomedicine
  • Shoukun Chen + 6 more

Information fusion and multi-classifier system for miner fatigue recognition in plateau environments based on electrocardiography and electromyography signals

  • Research Article
  • Cite Count Icon 1
  • 10.3233/jcm-247565
Evaluation on algorithms and models for multi-modal information fusion and evaluation in new media art and film and television cultural creation
  • Aug 14, 2024
  • Journal of Computational Methods in Sciences and Engineering
  • Junli Shao + 1 more

This paper promoted the development of new media art and film and television culture creation through multi-modal information fusion and analysis, and discussed the existing problems of new media art and film and television culture creation at present, including piracy, management problems and lack of innovation ability. The network structure of RNN neural network can cycle information among neurons, retain the memory of previous user information in the progressive learning sequence, analyze user behavior data through previous memory, accurately recommend users, and provide artists with a basis for user preferences. The viewing experience scores for works 1 to 5 created using traditional creative methods were 6.23, 6.02, 6.56, 6.64, and 6.88, respectively. The viewing experience scores for works 1 to 5 created through multi-modal information fusion and analysis were 9.41, 9.08, 9.11, 9.61, and 8.44, respectively. Movies created through multi-modal information fusion and analysis had higher viewing experience ratings. The results of this article emphasize that multi-modal information fusion and analysis can overcome the limitations of traditional single creative methods, provide rich and diverse expressions, and enable creators to more flexibly respond to complex creative needs, thereby achieving better creative effects.

  • Research Article
  • Cite Count Icon 8
  • 10.1360/n112017-00211
Intelligence methods of multi-modal information fusion in human-computer interaction
  • Apr 1, 2018
  • SCIENTIA SINICA Informationis
  • Minghao Yang + 1 more

We first introduce the concepts of single-modal information processing and multi-modal information fusion in cognitive science. Some classical multi-modal information fusion models and their computer implementations in history are also explained. Under the conditions that each channels information can be obtained, and their features could be unified representation synchronously, the fusion of multi-modal information can be transformed into classification or regression problems. For practical human-computer interaction systems, the performance of multi-modal information fusion largely relies on the accuracy of the single-modal information identification and the design of the interactive system. We present a practical example of multi-modal information fusion system, and discuss its performances on human computer interaction. Finally, the possible and important development trends for multi-modal human-computer interaction techniques and systems are discussed.

  • PDF Download Icon
  • Research Article
  • Cite Count Icon 2
  • 10.1155/2022/1534440
Low-Carbon Awareness Information Technology of Enterprise Executives Based on Big Data and Multimodal Information Fusion
  • Jun 28, 2022
  • Mobile Information Systems
  • Guimei Yang

The so-called multimodal information refers to the information from different information sources on different or the same side of the same description target. These pieces of information are different in terms of storage structure, representation, semantic connotation, credibility, and emphasis, but there is a certain inevitable connection between them. This paper aims to study how to analyze and study the low carbon of enterprises with the help of multimodal information fusion based on the background of big data and construct the evolutionary neural network of the improved adaptive genetic algorithm. This paper puts forward the problem of low carbon in enterprises, which is based on multimodal information fusion, and then elaborates on the concept and related algorithms of multimodal information fusion. Information fusion has carried out case design and analysis. Through the research and analysis of enterprise low carbon and self-adaptive algorithm, it can be obtained that the neural network has reached the threshold of 3.95 after iterating for nearly 60 generations, and stopped iterating to obtain the best individual. The evolutionary neural network in this paper reaches stability after a small number of iterations and can basically achieve a certain low carbonization.

  • Research Article
  • Cite Count Icon 27
  • 10.1007/s11227-013-1050-4
Towards a framework for large-scale multimedia data storage and processing on Hadoop platform
  • Dec 5, 2013
  • The Journal of Supercomputing
  • Wei Kuang Lai + 3 more

Cloud computing techniques take the form of distributed computing by utilizing multiple computers to execute computing simultaneously on the service side. To process the increasing quantity of multimedia data, numerous large-scale multimedia data storage computing techniques in the cloud computing have been developed. Of all the techniques, Hadoop plays a key role in the cloud computing. Hadoop, a computing cluster formed by low-priced hardware, can conduct the parallel computing of petabytes of multimedia data. Hadoop features high-reliability, high-efficiency, and high-scalability. The numerous large-scale multimedia data computing techniques include not only the key core techniques, Hadoop and MapReduce, but also the data collection techniques, such as File Transfer Protocol and Flume. In addition, distributed system configuration allocation, automatic installation, and monitoring platform building and management techniques are all included. As a result, only with the integration of all the techniques, a reliable large-scale multimedia data platform can be offered. In this paper, we introduce how cloud computing can make a breakthrough by proposing a multimedia social network dataset on Hadoop platform and implementing a prototype version. Detailed specifications and design issues are discussed as well. An important finding of this article is that we can save more time if we conduct the multimedia social networking analysis using Cloud Hadoop Platform rather than using a single computer. The advantages of cloud computing over the traditional data processing practices are fully demonstrated in this article. The applicable framework designs and the tools available for the large-scale data processing are also proposed. We show the experimental multimedia data including data sizes and processing time.

  • Research Article
  • 10.54097/c6jmgg90
Research on the Application of Multimodal Information Fusion in Financial Data Prediction
  • Sep 27, 2025
  • Journal of Computer Science and Artificial Intelligence
  • Qingying Zhi

The high complexity and uncertainty of financial markets make it difficult for predictive models that rely on a single data source to effectively address these challenges. Multimodal information fusion, by integrating heterogeneous data sources such as text, numerical data, images, and time series, reveals hidden connections and opens new avenues for improving the accuracy and robustness of financial forecasts. This paper systematically analyzes the theoretical foundations, key technical approaches, and typical scenarios for applying multimodal information fusion to financial forecasting. The study first analyzes the characteristics and categories of multimodal financial data, encompassing structured market data, unstructured text, time series data streams, and visual information. Secondly, it focuses on the core technologies of multimodal fusion, covering the data layer, feature layer, decision layer, and hybrid fusion strategies. It also analyzes the advantages of deep learning models in processing and fusing heterogeneous data. Through specific application cases such as stock price trend forecasting, credit risk assessment, and macroeconomic indicator forecasting, this paper demonstrates the significant effectiveness of multimodal fusion in capturing market sentiment, identifying potential risks, and improving forecasting accuracy. The study also identifies key challenges currently faced, including data heterogeneity, model interpretability, computational efficiency, and privacy protection. Finally, we outline future research directions, emphasizing the importance of cross-modal alignment, adaptive fusion mechanisms, interpretability enhancement, and privacy-preserving fusion frameworks. This study provides theoretical support and practical references for deepening the application of multimodal fusion in intelligent financial decision-making.

  • Research Article
  • Cite Count Icon 1
  • 10.3724/sp.j.1087.2008.00199
Video retrieval model based on multimodal information fusion
  • Jul 10, 2008
  • Journal of Computer Applications
  • Jing Zhang

In allusion to the complex requirement of query, a new video retrieval model based on multimodal information fusion was brought forward in this paper. It included multi-models like text retrieval, image query, semantic features extraction, and used relational algebra expression to fuse these multimodal information. Experimental results demonstrate that our method could fully utilize the advantages of multimodal information fusion based on relational expression in video retrieval, and achieve good performance on complex semantic video retrieval.

  • Research Article
  • Cite Count Icon 164
  • 10.1016/j.inffus.2019.06.019
A snapshot research and implementation of multimodal information fusion for data-driven emotion recognition
  • Jun 13, 2019
  • Information Fusion
  • Yingying Jiang + 5 more

A snapshot research and implementation of multimodal information fusion for data-driven emotion recognition

  • Research Article
  • Cite Count Icon 3
  • 10.1080/10447318.2023.2254644
Research on Emotion Recognition Method of Flight Training Based on Multimodal Fusion
  • Sep 16, 2023
  • International Journal of Human–Computer Interaction
  • Wendong Wang + 2 more

The emotional activities of the human body are mainly regulated by the autonomic nervous system, the central nervous system, and the advanced cognition of the human brain. This paper proposes an emotional state recognition method in pilot training tasks based on multimodal information fusion. A set of emotion perception recognition systems, a two-dimensional valence-arousal emotional model, and a multimodal information intelligent perception model were established based on the human-computer interaction mode during flight training; an intelligent perception system was designed to collect and intelligently perceive four kinds of peripheral physiological signals of pilots in real-time. And based on traditional machine learning models, a binary tree support vector machine was designed to optimize and improve the multimodal information co-integration decision model, which increased the accuracy of emotional state recognition in flight training by 37.58% on average. The experimental result showed that it realizes accurate monitoring and identification of real-time emotional state, helps to improve the effect of flight training and flight safety, maintains the efficiency of operation, and has important research significance and application prospects in the field of pilot training.

  • Preprint Article
  • 10.32920/ryerson.14668260.v1
Audiovisual Emotion Recognition Using Entropy-estimation-based Multimodal Information Fusion
  • May 24, 2021
  • Zhibing Xie

Understanding human emotional states is indispensable for our daily interaction, and we can enjoy more natural and friendly human computer interaction (HCI) experience by fully utilizing human’s affective states. In the application of emotion recognition, multimodal information fusion is widely used to discover the relationships of multiple information sources and make joint use of a number of channels, such as speech, facial expression, gesture and physiological processes. This thesis proposes a new framework of emotion recognition using information fusion based on the estimation of information entropy. The novel techniques of information theoretic learning are applied to feature level fusion and score level fusion. The most critical issues for feature level fusion are feature transformation and dimensionality reduction. The existing methods depend on the second order statistics, which is only optimal for Gaussian-like distributions. By incorporating information theoretic tools, a new feature level fusion method based on kernel entropy component analysis is proposed. For score level fusion, most previous methods focus on predefined rule based approaches, which are usually heuristic. In this thesis, a connection between information fusion and maximum correntropy criterion is established for effective score level fusion. Feature level fusion and score level fusion methods are then combined to introduce a two-stage fusion platform. The proposed methods are applied to audiovisual emotion recognition, and their effectiveness is evaluated by experiments on two publicly available audiovisual emotion databases. The experimental results demonstrate that the proposed algorithms achieve improved performance in comparison with the existing methods. The work of this thesis offers a promising direction to design more advanced emotion recognition systems based on multimodal information fusion and has great significance to the development of intelligent human computer interaction systems.

  • Preprint Article
  • 10.32920/ryerson.14668260
Audiovisual Emotion Recognition Using Entropy-estimation-based Multimodal Information Fusion
  • May 24, 2021
  • Zhibing Xie

Understanding human emotional states is indispensable for our daily interaction, and we can enjoy more natural and friendly human computer interaction (HCI) experience by fully utilizing human’s affective states. In the application of emotion recognition, multimodal information fusion is widely used to discover the relationships of multiple information sources and make joint use of a number of channels, such as speech, facial expression, gesture and physiological processes. This thesis proposes a new framework of emotion recognition using information fusion based on the estimation of information entropy. The novel techniques of information theoretic learning are applied to feature level fusion and score level fusion. The most critical issues for feature level fusion are feature transformation and dimensionality reduction. The existing methods depend on the second order statistics, which is only optimal for Gaussian-like distributions. By incorporating information theoretic tools, a new feature level fusion method based on kernel entropy component analysis is proposed. For score level fusion, most previous methods focus on predefined rule based approaches, which are usually heuristic. In this thesis, a connection between information fusion and maximum correntropy criterion is established for effective score level fusion. Feature level fusion and score level fusion methods are then combined to introduce a two-stage fusion platform. The proposed methods are applied to audiovisual emotion recognition, and their effectiveness is evaluated by experiments on two publicly available audiovisual emotion databases. The experimental results demonstrate that the proposed algorithms achieve improved performance in comparison with the existing methods. The work of this thesis offers a promising direction to design more advanced emotion recognition systems based on multimodal information fusion and has great significance to the development of intelligent human computer interaction systems.

  • Research Article
  • 10.3934/era.2024292
Resnet-1DCNN-REA bearing fault diagnosis method based on multi-source and multi-modal information fusion
  • Jan 1, 2024
  • Electronic Research Archive
  • Xu Chen + 5 more

<p>In order to address the issue of multi-information fusion, this paper proposed a method for bearing fault diagnosis based on multisource and multimodal information fusion. Existing bearing fault diagnosis methods mainly rely on single sensor information. Nevertheless, mechanical faults in bearings are intricate and subject to countless excitation disturbances, which poses a great challenge for accurate identification if only relying on feature extraction from single sensor input. In this paper, a multisource information fusion model based on auto-encoder was first established to achieve the fusion of multi-sensor signals. Based on the fused signals, multimodal feature extraction was realized by integrating image features and time-frequency statistical information. The one-dimensional vibration signals were converted into two-dimensional time-frequency images by continuous wavelet transform (CWT), and then they were fed into the Resnet network for fault diagnosis. At the same time, the time-frequency statistical features of the fused 1D signal were extracted from the integrated perspective of time and frequency domains and inputted into the improved 1D convolutional neural network model based on the residual block and attention mechanism (1DCNN-REA) model to realize fault diagnosis. Finally, the tree-structured parzen estimator (TPE) algorithm was utilized to realize the integration of two models in order to improve the diagnostic effect of a single model and obtain the final bearing fault diagnosis results. The proposed model was validated using real experimental data, and the results of the comparison and ablation experiments showed that compared with other models, the proposed model can precisely diagnosis the fault type with an accuracy rate of 98.93%.</p>

  • Research Article
  • Cite Count Icon 3
  • 10.1155/2022/2447263
Design of Multimodal Neural Network Control System for Mechanically Driven Reconfigurable Robot.
  • May 25, 2022
  • Computational Intelligence and Neuroscience
  • Zhang Youchun + 1 more

According to the characteristics and division rules of the modules, this paper divides the rotary module, the swing module, and the mobile module. In order to realize the rapid identification of modules, according to the basic principles of module design, the above modules are put into the module library established by Access. According to the modular modeling method, kinematic models are established, respectively. In order to automatically establish the kinematic model of the robot, a unified expression of modules is established. According to the unified expression of the modules, the kinematics of the reconfigurable robot is analyzed. According to the characteristics of the configuration plane, the configuration plane is divided, and the expression form of the position and attitude of the configuration plane is given. Combined with the principle of neural network and multimodal information fusion, a multimodal information fusion model based on long- and short-term memory neural network is established. Aiming at the control problem of mechanically driven reconfigurable robots, a specific long- and short-term memory neural network model is designed, and the long- and short-term memory neural network algorithm is applied to the robot control problem based on multimodal information fusion. The design of the controller and driver of each joint is the basis of the distributed control system. This paper discusses the hardware design of the joint controller and driver and the realization of the position control system and discusses the method of realizing distributed control based on the Modbus protocol of RS485 communication. Through the comprehensive experiment of the configuration, the point control and the continuous path control are carried out to verify the correctness of the theoretical analysis of the system and the reliability of the system hardware and software operation.

  • Research Article
  • 10.1177/14759217251333761
Fault diagnosis of wind turbine blades under wide-weather multi-operating conditions based on multi-modal information fusion and deep learning
  • May 5, 2025
  • Structural Health Monitoring
  • Ying Han + 2 more

Continuous health monitoring of wind turbine blades under all-weather and multi-operating conditions represents a significant challenge in the renewable energy sector. In this article, we present a fault diagnosis approach leveraging multi-modal information fusion and deep learning with continuous state division, thereby overcoming the limitations of traditional methods in complex and noisy environments. The operational conditions of wind turbine blades are categorized into two primary states: the wind operation state and the sudden shutdown state. Additionally, various climate types, including sunny, foggy, windy conditions, differing lighting levels, and others, are considered in the analysis. During wind operation, sound and vibration signals exhibit higher efficacy for fault detection; however, high noise levels may introduce interferences. To address the issue of indistinct fault characteristics after deep convolution due to multiple noise factors, which could result in reduced diagnostic accuracy, we propose a robust fast Fourier transform-ResTransNet model. In the shutdown state, vibration and sound data features become less prominent, making image processing techniques advantageous. Nevertheless, diverse climate types can lead to challenges such as low visibility, high noise, and other interferences. Consequently, we design a Swin-Transformer model that integrates infrared thermal imaging and visible light imagery. This model resolves the problem of non-homologous data representation and ensures accurate information interaction under multi-source data fusion. Simulation results confirm that the proposed fault diagnosis method achieves substantial improvements over existing approaches. To validate the practical applicability of our method, we construct a real-world wind turbine operational environment, simulate several common blade fault scenarios, and collect actual vibration, sound, and image data under varying weather conditions. Based on these simulations, we establish a multi-modal information fusion model tailored for different weather types. Furthermore, to facilitate the integration of our research into real-world wind farm operations, we develop a human–computer interface that enables seamless deployment. The corresponding source code is publicly available at https://github.com/midfigher/Humancomputer-interaction .

More from: Big Data Research
  • Research Article
  • 10.1016/j.bdr.2025.100570
Tangible progress: Employing visual metaphors and physical interfaces in AI-based English language learning
  • Nov 1, 2025
  • Big Data Research
  • Mei Wang + 4 more

  • Research Article
  • 10.1016/j.bdr.2025.100569
Exogenous Variable Driven Cotton Prices Prediction: Comparison of Statistical Model with Sequence Based Deep Learning Models
  • Oct 1, 2025
  • Big Data Research
  • G.Y Chandan + 1 more

  • Research Article
  • 10.1016/j.bdr.2025.100534
Big data analytics for smart home energy management system based on IOMT using AHP and WASPAS
  • Aug 1, 2025
  • Big Data Research
  • Jingze Zhou + 3 more

  • Research Article
  • 10.1016/j.bdr.2025.100540
The influence of China's exchange rate market on the Belt and Road trade market: Based on temporal two-layer networks
  • Aug 1, 2025
  • Big Data Research
  • Xiaoyu Zhang + 2 more

  • Research Article
  • 10.1016/j.bdr.2025.100553
Deep neural network modeling for financial time series analysis
  • Aug 1, 2025
  • Big Data Research
  • Zheng Fang + 1 more

  • Research Article
  • 10.1016/j.bdr.2025.100552
Time-synchronized sentiment labeling via autonomous online comments data mining: A multimodal information fusion on large-scale multimedia data
  • Aug 1, 2025
  • Big Data Research
  • Jiachen Ma + 3 more

  • Research Article
  • 10.1016/j.bdr.2025.100539
Exploring the impact of high schools, socioeconomic factors, and degree programs on higher education success in Italy
  • Aug 1, 2025
  • Big Data Research
  • Cristian Usala + 2 more

  • Research Article
  • 10.1016/j.bdr.2025.100557
Research on Modeling of the Imbalanced Fraudulent Transaction Detection Problem Based on Embedding-Aware Conditional GAN
  • Aug 1, 2025
  • Big Data Research
  • Luping Zhi + 1 more

  • Research Article
  • 10.1016/j.bdr.2025.100554
Compression of big data collected in wind farm based on tensor train decomposition
  • Aug 1, 2025
  • Big Data Research
  • Keren Li + 6 more

  • Research Article
  • 10.1016/j.bdr.2025.100551
BETM: A new pre-trained BERT-guided embedding-based topic model
  • Aug 1, 2025
  • Big Data Research
  • Yang Liu + 3 more

Save Icon
Up Arrow
Open/Close
  • Ask R Discovery Star icon
  • Chat PDF Star icon

AI summaries and top papers from 250M+ research sources.

Search IconWhat is the difference between bacteria and viruses?
Open In New Tab Icon
Search IconWhat is the function of the immune system?
Open In New Tab Icon
Search IconCan diabetes be passed down from one generation to the next?
Open In New Tab Icon