TY - JOUR
T1 - Domain-Specific Sentiment Analysis
T2 - An Optimized Deep Learning Approach for the Financial Markets
AU - Yekrangi, Mehdi
AU - Nikolov, Nikola S.
N1 - Publisher Copyright:
© 2013 IEEE.
PY - 2023
Y1 - 2023
N2 - Although different studies are caried out by deep learning models for financial markets sentiment analysis, there is a lack of specific embedding method that regards the domain. Therefore, the goal of this study is to discover what type of embedding techniques along with different classification algorithms work better for the financial markets' sentiment analysis to present an optimized embedding method in the domain. In this paper we present a broad comparative study of multiple classification models trained for sentiment analysis and will improve their performance with an optimized embedding layer. We use a heterogeneous corpus of both formal (news headlines) and informal (tweets) text to increase the robustness and build the models with CBOW, GloVe, and BERT pre-trained embeddings as well as developing an optimized embedding layer to improve the results. The best results reported here are by our LSTM model with the fine-tuned embedding layer, which has an accuracy of 0.84 and a macro-average F1-score of 0.8. Our results give evidence that the fine-tuned embedding is superior to utilising pretrained CBOW, GloVe, and BERT embeddings for financial markets sentiment analysis. We train SVM, MLP, CNN, generic RNN and LSTM models by a comprehensive approach in input data and algorithms. As a result, a sentiment analysis model is presented with a robust performance for different datasets in the domain.
AB - Although different studies are caried out by deep learning models for financial markets sentiment analysis, there is a lack of specific embedding method that regards the domain. Therefore, the goal of this study is to discover what type of embedding techniques along with different classification algorithms work better for the financial markets' sentiment analysis to present an optimized embedding method in the domain. In this paper we present a broad comparative study of multiple classification models trained for sentiment analysis and will improve their performance with an optimized embedding layer. We use a heterogeneous corpus of both formal (news headlines) and informal (tweets) text to increase the robustness and build the models with CBOW, GloVe, and BERT pre-trained embeddings as well as developing an optimized embedding layer to improve the results. The best results reported here are by our LSTM model with the fine-tuned embedding layer, which has an accuracy of 0.84 and a macro-average F1-score of 0.8. Our results give evidence that the fine-tuned embedding is superior to utilising pretrained CBOW, GloVe, and BERT embeddings for financial markets sentiment analysis. We train SVM, MLP, CNN, generic RNN and LSTM models by a comprehensive approach in input data and algorithms. As a result, a sentiment analysis model is presented with a robust performance for different datasets in the domain.
KW - deep learning
KW - financial markets
KW - natural language processing
KW - neural networks
KW - Sentiment analysis
KW - text mining
UR - http://www.scopus.com/inward/record.url?scp=85164697087&partnerID=8YFLogxK
U2 - 10.1109/ACCESS.2023.3293733
DO - 10.1109/ACCESS.2023.3293733
M3 - Article
AN - SCOPUS:85164697087
SN - 2169-3536
VL - 11
SP - 70248
EP - 70262
JO - IEEE Access
JF - IEEE Access
ER -