Abstract
Sentiment analysis is one of the core tasks in natural language processing, and its main goal is to identify and classify the sentiment tendencies contained in texts. Traditional sentiment analysis methods and shallow models often fail to capture the rich semantic information and contextual relationships in text, while increasing the network depth is prone to problems such as network degradation, which has some limitations in terms of accuracy and performance. Based on this foundation, a sentiment analysis model called BERT-Invos (Bidirectional Encoder Representations from Transformers (BERT)-based stacked involutions) is introduced. The model utilizes the dynamic pre-training language model BERT to encode text, enabling rich semantic features and contextual understanding. In addition, the model employs stacked involutions with varying dimensions to extract features and perceive local information, gradually learning different scales and hierarchical representations of the input text. Furthermore, the proposed method incorporates nonlinear activation functions such as ReLU6 and H-Swish to enhance the model’s expression capability and performance, ultimately delivering classification results. In the experiment, a financial news sentiment dataset was utilized for model validation and comparison against other models. The results revealed that the model achieved an accuracy of 96.34% in sentiment analysis tasks, with precision, recall, and F1 score reaching 96.37%, 96.34%, and 96.34%, respectively. Additionally, the loss value could be minimized to 0.07 with stable convergence, thereby enhancing the accuracy of sentiment classification and reducing loss rates. This improvement facilitates better capturing of local patterns in text and addresses the issue of degradation in deep neural networks. Regarding the deep architecture of the proposed model, future work will focus on further exploring optimization techniques for model compression and deployment.
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.Data availability
The emotional analysis datasets that are used in this paper are available on the following public link: https://github.com/wwwxmu/Dataset-of-financial-news-sentiment-classification.
References
Kalashami MP, Pedram MM, Sadr H (2022) EEG feature extraction and data augmentation in emotion recognition. Comput Intell Neurosci 2022(1):7028517–7028517
McCallum A, Freitag D, Pereira FC et al (2000) Maximum entropy Markov models for information extraction and segmentation. Icml 17(2000):591–598
Wallach HM (2004) Conditional random fields: an introduction. Tech Rep (CIS) 24(2):22–31
Hidayat THJ, Ruldeviyani Y, Aditama AR et al (2022) Sentiment analysis of twitter data related to Rinca Island development using Doc2vec and SVM and logistic regression as classifier. Procedia Comput Sci 197(1):660–667
Rahman H, Tariq J et al (2022) Multi-tier sentiment analysis of social media text using supervised machine learning. Comput Mater Continua 74(3):5527–5543
Palmer M, Roeder J, Muntermann J (2022) Induction of a sentiment dictionary for financial analyst communication: a data-driven approach balancing machine learning and human intuition. J Bus Anal 5(1):8–28
Ojeda-Hernández M, López-Rodríguez D, Mora Á (2023) Lexicon-based sentiment analysis in texts using formal concept analysis. Int J Approx Reason 155(1):104–112
Krizhevsky A, Sutskever I, Hinton GE (2017) Imagenet classification with deep convolutional neural networks. Commun ACM 60(6):84–90
Ce P, Tie B (2020) An analysis method for interpretability of CNN text classification model. Future Internet 12(12):228–242
Liu LW, Yu X (2019) Circulating neural network (rnn) and its application. IEEE J Sci Technol World 32(6):54–55
Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, Kaiser Ł, Polosukhin I (2017) Attention is all you need. Adv Neural Inf Process Syst 30(1):1–11
Girshick R, Donahue J, Darrell T, Malik J (2014) Rich feature hierarchies for accurate object detection and semantic segmentation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 32(1): 580–584
Mohades Deilami F, Sadr H, Tarkhan M (2022) Contextualized Multidimensional Personality Recognition using Combination of Deep Neural Network and Ensemble Learning. Neural Process Lett 54(5):3811–3828
Deng J, Cheng L, Wang Z (2021) Attention-based BiLSTM fused CNN with gating mechanism model for Chinese long text classification. Comput Speech Lang 68(1):101182–101194
Sadr H, Nazari Soleimandarabi M (2022) Acnn-tl: attention-based convolutional neural network coupling with transfer learning and contextualized word representation for enhancing the performance of sentiment classification. J Supercomput 78(7):10149–10175
Chen Y, Dai X, Liu M, Chen D, Yuan L, Liu Z (2020) Dynamic convolution: attention over convolution kernels. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 38(1): 11030–11039
Li D, Hu J, Wang C, Li X, She Q, Zhu L, Zhang T, Chen Q (2021) Involution: inverting the inherence of convolution for visual recognition. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 39(1): 12321–12330
He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 34(1): 770–778
Devlin J, Chang MW, Lee K, Toutanova K (2019) Bert: pre-training of deep bidirectional transformers for language understanding. In: Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. Association for Computational Linguistics, Stroudsburg, PA, 22(1): 4171–4186
Yin D, Meng T, Chang K-W (2020) SentiBERT: a transferable transformer-based architecture for compositional sentiment semantics. In: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, 58(1): 3695–3706
Wan C-X, Li B (2021) Financial causal sentence recognition based on BERT-CNN text classification. J Supercomput 78(5):1–25
Zhang M, Wang J (2021) Automatic extraction of flooding control knowledge from rich literature texts using deep learning. Appl Sci 78(5):1–25
Yuan S, Wang Q (2022) Imbalanced traffic accident text classification based on Bert-RCNN. J Phys Conf Ser 2170(5):1–9
Xu G, Zhang Z, Zhang T, Yu S, Meng Y, Chen S (2022) Aspect-level sentiment classification based on attention-BiLSTM model and transfer learning. Knowl Based Syst 245(5):1–9
Howard A, Sandler M, Chu G et al (2019) Searching for mobilenetv3. CoRR 67(1):1–10
Funding
This work was supported by the National Natural Science Foundation of China (No.62103350) and Shandong Provincial Natural Science Foundation (ZR2020QF046).
Author information
Authors and Affiliations
Contributions
Idea conception, design of the model and computational framework, simulation of experiments and analysis of experimental results were carried out by SL. The first draft of the manuscript was written by SL. A previous version of the manuscript was commented on by QL. Comments on the manuscript were provided by all authors. Qicheng Liu was responsible for the overall direction and planning.
Corresponding author
Ethics declarations
Conflict of interest
The authors declared that they do not have any commercial or associative interest that represents a Conflict of interest in connection with the work submitted.
Ethics approval
This article does not contain any studies with human participants or animals performed by any of the authors.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Liu, S., Liu, Q. A sentiment analysis model based on dynamic pre-training and stacked involutions. J Supercomput 80, 15613–15635 (2024). https://doi.org/10.1007/s11227-024-06052-6
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11227-024-06052-6