Abstract
News articles have significant impacts on asset prices in financial markets. A great number of attempts have been conducted to ascertain how news articles influence stock prices. News articles have been reported to contain sentimental and fundamental information that affects stock price fluctuations, and many studies have been conducted to evaluate stock price fluctuations using them as analytical data. However, the limitations in the number of available datasets usually become the hurdle for the model accuracy. This study aims to improve the analytical model’s accuracy by generating news articles using language generation technology. We tested whether the model that used the generated data was better than the trained model with real-world data. The model constructed in this research is a model that evaluates news articles distributed to financial markets based on the price fluctuation rate of stock prices and predicts and evaluates stock price fluctuations. This study labeled based on high-frequency trading data and generated news articles using a large-scale language generation model (GPT-2). Also, we analyzed and verified the effect. In this study, we succeeded in generating news articles using the large-scale language generation model and improving the classification accuracy. Our method proposed in this paper has great potential to improve text analysis accuracy in various areas.
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.Data availability
Data available from Thomson Reuters Real-Time News: Feed and Archive, Nikkei Needs and Refinitiv Eikon. The authors do not have permission to share data.
References
Aishwarya A, Jiasen L, Stanislaw A, Margaret MC, Lawrence Z, Dhruv B, Devi P (2015) VQA: visual question answering. In: Proceedings of the international conference on computer vision
Ashish V, Noam S, Niki P, Jakob U, Llion J, Aidan NG, Lukasz K, Illia P (2017) Attention is all you need. In Advances in neural information processing systems, pp 6000–6010
Bajgar O, Kadlec R, Kleindienst J (2016) Embracing data abundance: book test dataset for reading comprehension. arXivpreprint arXiv:1610.00956
Balducci B, Marinova D (2018) Unstructured data in marketing. J Acad Market Sci 46:557–590
Benjamini Y, Hochberg Y (1995) Controlling the false discovery rate: a practical and powerful approach to multiple testing. J R Stat Soc B 57:289–300
Blei DM, Ng AY, Jordan MI (2003) Latent Dirichlet allocation. J Mach Learn Res 3:993–1022
Bojanowski P, Grave E, Joulin A, Mikolov T (2016) Enriching word vectors with subword information. arXiv preprint arXiv:1607.04606
Fung GPC, Yu JX, Lam W (2002) News sensitive stock trend prediction. In: Proceedings of the 6th Pacific-Asia conference on knowledge discovery and data mining, pp 481-493
Fung GPC, Yu JX, Lam W (2003) Stock prediction: integrating text mining approach using real-time news. In Proceedings of the IEEE international conference on computational intelligence for financial engineering, pp 395-402
Gidófalvi G (2001) Using news articles to predict stock price movements. Technical Report University of California, Department of Computer Science and Engineering
Gong C, He D, Tan X, Qin T, Wang L, Liu T.-Y (2018) Frage: frequency-agnostic word representation. In: Advances in neural information processing systems, pp 1341–1352
Goshima K, Takahashi H (2016) Text analysis system for measuring the influence of news articles on intraday price changes in financial markets. Smart Innov Syst Technol 58:341–348
Grave E, Joulin A, Usunier N (2016) Improving neural language models with a continuous cache. arXiv preprint arXiv: 1612.04426
Hoang L, Wiseman S, Rush AM (2018) Entity tracking improves cloze-style reading comprehension. arXiv preprint arXiv: 1810.02891
Hochreiter S, Schmidhuber J (1997) Long short-term memory. Neural Comput 9(8):1735–1780
Hu Z, Liu W, Bian J, Liu X, Liu T-Y (2018) Listening to chaotic whispers: a deep learning framework for news-oriented stock trend prediction, In Proceedings of the eleventh ACM international conference on web search and data mining, pp 261–269
Jacob D, Ming-Wei C, Kenton L, Kristina T (2018) Bert: pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805
Jordan B, Diego F, Aniko E, Cristiano P, Pedro A (2020) LSTM and GPT-2 synthetic speech transfer learning for speaker recognition to overcome data scarcity. arXiv preprintarXiv:2007.00659
Liu Y, Ott M, Goyal N, Du J, Joshi M, Chen D, Levy O, Lewis M, Zettlemoyer L, Stoyanov V (2019) Roberta: A robustly optimized bert pretraining approach. arXiv preprint arXiv:1907.11692
Mikolov T, Chen K, Corrado G, Dean J (2013a) Efficient estimation of word representations in vector space
Mikolov T, Sutskever I, Chen K, Corrado G, Dean J (2013b) Distributed representations of words and phrases and their compositionality. In Proceedings of the NeurIPS
Mittermayer MA (2004) Forecasting intraday stock price trends with text mining techniques. In Proceedings of the 37th Hawaii international conference on system sciences
Nishi Y, Suge A, Takahashi H (2019) Text analysis on the stock market in the automotive industry through fake news generated by GPT-2. In Proceedings of the JSAI international symposia on AI 2019
Nishi Y, Suge A, Takahashi H (2020) Construction of news article evaluation system using language generation model. Agents Multi-Agent Syst Technol Appl 186:313–320
Nishi Y, Suge A, Takahashi H (2020b) News articles evaluation analysis in automotive industry using GPT-2 and co-occurrence network. New Front Artif Intell 12331:103–114
Radford A, Narasimhan K, Salimans T, Sutskever I (2018) Improving language understanding by generative pre-training. Technical Report Open AI
Radford A, Wu J, Child R, Luan D, Amodei, D, Sutskever I (2019) Language models are un-supervised multitask learners. Technical Report Open AI
Reiter E, Dale R (2000) Building natural language generation systems. Cambridge University Press, Cambridge
Rong X (2014) word2vec parameter learning explained. arXiv preprint arXiv:1411.2738
Schumaker RP, Chen H (2009) Textual analysis of stock market prediction using breaking financial news. ACM Trans Inf Syst 27:1–19
Schuster M, Paliwal KK (1997) Bidirectional recurrent neural networks. IEEE Trans Signal Process 45(11):2673–2681
Subramanian S, Li R, Pilault J, Pal C (2019) On extractive and abstractive neural document summarization with transformer language models. arXiv preprint arXiv:1909.03186
Suge A, Takahashi H (2018) Analyzing the relationship between news and the stock market through high-frequency data. The Securities Analysts Association of Japan (in Japanese)
Swets JA (1996) Signal detection theory and ROC analysis in psychology and diagnostics: collected papers. Lawrence Erlbaum Associates, Mahwah, NJ
Vargas MR, de Lima BSLP, Evsukoff AG (2017) Deep learning for stock market prediction from financial news articles. In 2017 IEEE international conference on computational intelligence and virtual environments for measurement systems and applications (CIVEMSA), pp 60–65
Zhang R, Guo J, Fan Y, Lan Y, Xu J, Cheng X (2018) Learning to control the specificity in neural response generation. In Proceedings of the 56th annual meeting of the association for computational linguistics, pp 1108–1117
Acknowledgements
This study extends a previous study (Nishi et al. 2020b) and builds on a more accurate classification model through high-frequency data and sentence generation models.
Funding
Partial funding by JSPS KAKENHI, JP20K01751.
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
The authors have no conflicts of interest to declare that are relevant to the content of this article.
Rights and permissions
About this article
Cite this article
Nishi, Y., Suge, A. & Takahashi, H. Construction of a news article evaluation model utilizing high-frequency data and a large-scale language generation model. SN Bus Econ 1, 104 (2021). https://doi.org/10.1007/s43546-021-00106-0
Received:
Accepted:
Published:
DOI: https://doi.org/10.1007/s43546-021-00106-0