1. INSTRUCTION
In the sophisticated tapestry of global finance, the quest to accurately predict stock prices has remained an elusive Holy Grail. This ambition, transcending mere academic intrigue, holds the potential to revolutionize investment strategies, engendering both individual prosperity and broader economic stability.
The enigma of predicting stock prices is deeply entrenched in the bedrock principles of financial economics. Central to this is the Efficient Market Hypothesis (EMH), a dominant theory that suggests stock prices swiftly assimilate and mirror all accessible information. At a glance, consistently besting the market could appear to be an exercise in futility. Yet, the documented presence of several market anomalies, as highlighted in research by Smith [1] and Brown & Clark [2], implies the existence of subtle patterns awaiting discovery.
This brings us to the challenge: why has the prediction of stock prices remained such a formidable task? The complexities lie in both the dynamic nature of the markets and the multifaceted influences that drive price changes. From geopolitical shifts and policy changes to internal corporate strategies and global economic indicators, the intricate web of factors influencing stock prices is vast. Layer on the impact of data noise, emotional trading, and market irrationalities, and the challenge deepens. The volatility and uncertainty of stock prices are not merely a reflection of external influences but also an embodiment of collective psychological responses—a realm of unpredictability.
Yet, hope emerges on the horizon with the advancement of technology. The recent explosion of interest in artificial intelligence and machine learning offers a potential solution to this age-old conundrum. This research centers on one such promising avenue: the deployment of Deep Neural Networks (DNNs), with a particular focus on Long Short-Term Memory (LSTM) networks. These sophisticated algorithms, revered for their prowess in various domains, offer the tantalizing potential of decoding the intricate patterns hidden within stock market data.
The goal of this research, therefore, is twofold: to critically assess the capability of LSTM networks in predicting stock prices and to elucidate the underlying factors that might influence their predictive accuracy. We endeavor to gauge LSTM's technical efficacy and position our findings within the broader context of financial theory and practice. By navigating the intersection of technology and finance, we aim to offer a comprehensive perspective that resonates with both researchers and practitioners.
As we embark on this explorative journey, let us provide a roadmap for our readers. The paper unfolds as follows:
1. Challenges and Realities of Stock Price Prediction: An in-depth dive into the inherent intricacies of the stock market, encompassing its unpredictability, the challenges posed by the Efficient Market Hypothesis, and the perturbations caused by data noise and irrational trading.
2. Deep Neural Networks and Time-Series Prediction: A comprehensive overview of LSTM networks, tracing their foundational principles, operational mechanics, and their potential applicability to time-series forecasting tasks.
3. Data Preprocessing and Feature Extraction: Delving into the nuances of financial data, from its inherent characteristics and the challenges they pose to strategies for data cleaning and feature extraction.
4. Model Construction and Training Strategy: A detailed exposition of the LSTM model construction process, elucidating decisions about architecture, training strategies, and performance validation.
5. Experimental Design and Analysis: Presentation and analysis of our experimental design, capturing the breadth of our data selection, experimentation on diverse stock indices, and assessment criteria.
6. Prospects and Challenges of Application: A foresight into the future trajectory of DNNs in finance, intertwined with a discussion on impending challenges.
7. Conclusion: A synthesis of our findings, culminating in reflections on the broader implications of our research and the potential vistas it might open for future investigations.
In this endeavor, we hope to bridge the chasm between technology's promises and the complex realities of financial markets, offering insights that might guide future research and practical applications in the domain of stock price prediction.
2. CHALLENGES AND REALITIES OF STOCK PRICE PREDICTION
2.1. Complexity and Uncertainty of Stock Markets
The financial markets, with their multifaceted dynamics, have been likened to the intricate mechanisms of a clockwork, where every cog and wheel plays a part. Stocks are not isolated entities. Their price fluctuations stem from myriad determinants, from a company's internal operations to overarching global economic and political trends. As the world hurtles into the digital age, technological innovations and the intensification of globalization have amplified these complexities. A single incident in one corner of the globe can create ripples that affect markets continents away. From ecological crises and geopolitical tensions to groundbreaking technological innovations and changing consumer preferences, all these variables amalgamate, steering the direction of stock markets [3].
Furthermore, as markets have evolved, so have the number of participants and the diversity of their strategies. High-frequency traders, institutional investors, retail traders, and even automated trading algorithms all interact in real-time, leading to emergent behaviors that are incredibly difficult to predict. And then there's the unpredictability of global events, from elections and policy changes to wars and natural disasters, all of which can send shockwaves through markets without prior warning.
2.2. Efficient Market Hypothesis and the Existence of Anomalies
The Efficient Market Hypothesis (EMH) has been at the heart of financial theory since the latter half of the 20th century. The idea is both elegant and profound: if markets are efficient, then all known information should be already priced into stocks, rendering it impossible to gain consistent excess returns without taking on additional risk [4].
However, through the years, certain market anomalies have been identified that challenge the robustness of the EMH. For instance, the momentum effect, where stocks that have performed well in the past continue to do so in the near future, and vice versa, challenges the notion of full market efficiency [5]. Other anomalies, such as the post-earnings announcement drift and the small-firm effect, further accentuate that markets might not be as efficient as once thought.
2.3. Data Noise and the Impact of Market Irrationality)
Noise is inherent in almost any data-driven system, and the stock market is no exception. Price movements are frequently driven by factors unrelated to a company's intrinsic value. These can range from unfounded rumors to media hype, short-term trading strategies, or even simple human error [6]. Such noise can obscure genuine price signals and make it harder for traders and investors to discern genuine value.
Moreover, behavioral finance, a field that blends psychological theory with conventional economics, suggests that investors are not always rational. Cognitive biases like herd behavior, loss aversion, and overconfidence can result in pricing distortions. These biases, combined with the aforementioned noise, can lead to market bubbles and crashes, showcasing the profound impact of irrational behavior on market dynamics.
In conclusion, while technology and advanced modeling techniques, like deep neural networks, promise better prediction capabilities, the stock market remains a challenging domain due to its inherent complexities, potential inefficiencies, and the pervasive influence of noise and irrational behavior.
3. DEEP NEURAL NETWORKS AND TIME SERIES FORECASTING
Deep neural networks (DNNs), with their capacity for feature extraction and complex modeling, have made significant strides in a variety of application domains. One such domain, rich with challenges and opportunities, is time series forecasting. The sequential nature of time series data, with its inherent temporal dependencies, makes traditional neural networks unsuitable due to their inability to maintain long-term dependencies. This limitation led to the advent and popularity of specialized neural network architectures, most notably the Long Short-Term Memory (LSTM) networks.
3.1. Basic Structure and Working Principle of LSTM Networks
Figure 1. Architecture of an LSTM unit
Introduced by Hochreiter & Schmidhuber [7], LSTMs were designed specifically to combat the vanishing and exploding gradient problems that plague traditional RNNs, making them ill-suited for long sequences. At the core of the LSTM is the memory cell, an intricate assembly of gates that regulate information flow.
In figure 1, there are three main gates in an LSTM cell:
1. Forget Gate: the algorithm determines what information should be retained or discarded from both the previous hidden state and the current input. This decision is made by running these values through a sigmoid function, which outputs a number between 0 and 1. A value of 0 indicates that certain past information is no longer needed, while a value of 1 signifies that it should be kept. This output is then element-wise multiplied with the existing Cell State, effectively 'forgetting' information that has been deemed unimportant.
2. Input Gate: the algorithm assesses how crucial the current input is for the task at hand. This is done by multiplying the current input with the previous hidden state and a weight matrix. The resultant valuable information is then added to the Cell State, updating it to form the new Cell State, c(t), which serves as the present long-term memory for subsequent iterations.
3. Output Gate, the LSTM model's output is generated within the Hidden State. Depending on the specific application, this could be a word that adds semantic value to a sentence, for example. Here, a sigmoid function filters what information is allowed through the output gate. This is followed by the activation of the Cell State using the tanh function, and then it is multiplied to produce the final output.
These gates allow LSTMs to learn and remember over long sequences and are particularly adept at retaining information for long periods, which is crucial in time series forecasting.
3.2. Applicability of LSTM in Time Series Data
For time series forecasting, capturing patterns over different time horizons—be it short-term, medium-term, or long-term—is essential. LSTMs, with their memory cells, are naturally suited for this task. They can remember and leverage information from many timesteps before, making them effective for predicting future values based on past trends. Moreover, their structure allows them to learn complex temporal dependencies, which can be pivotal in financial markets, weather forecasting, and energy consumption prediction, among others [8].
Additionally, with the advent of variants like Bidirectional LSTM and LSTM with attention mechanisms, the capabilities of these networks in handling time series data have only been augmented. These architectures allow for better utilization of past data by giving varying importance to different time steps or by leveraging both past and future data for the current prediction.
3.3. Previous Successful Applications in Other Domains
LSTMs have not just shined in time series forecasting; their success spans a myriad of domains. One of the landmark applications was in machine translation, where LSTMs, particularly when combined with attention mechanisms, showed state-of-the-art performance in tasks such as English-to-French translation [9].
In speech recognition, LSTMs played a pivotal role in the development of end-to-end systems, minimizing the reliance on handcrafted features and showing superior performance [10]. Additionally, in the realm of image captioning, where the goal is to generate descriptive text for images, LSTMs demonstrated their ability to bridge vision and language[11].
The versatility and efficacy of LSTMs across such diverse applications underscore their potential, and when tailored for specific tasks like time series forecasting, they promise substantial advancements.
In summary, LSTMs, with their specialized architecture, present a promising solution to the challenges of time series forecasting. Their proven track record across various domains is a testament to their robustness and versatility.
4. DATA PREPROCESSING AND FEATURE EXTRACTION
Data preparation is a crucial step in any machine learning pipeline, and this is especially true for stock market data, given its intricate nature and potential impact on financial decisions. Efficient preprocessing can not only ensure the integrity and reliability of data but can also lead to better model performance by extracting meaningful features that capture the essence of the data.
4.1. Characteristics and Challenges of Stock Market Data
Stock market data is inherently time-series data, and it comes with a set of unique characteristics and challenges. It is high-dimensional, comprising prices, trading volumes, and often auxiliary data like news feeds, economic indicators, and corporate actions. The data also demonstrates non-stationarity, where statistical properties change over time, making modeling more complex[12].
Another challenge is the sheer volume. With stock exchanges globally running for several hours a day and recording data at granular intervals (like minute-level or even second-level), datasets can become massive. Additionally, stock data can be noisy, with price movements driven not just by tangible financial metrics but also by speculative trades, rumors, and psychological factors[13].
4.2. Data Cleaning and Anomaly Handling Strategies
Given the challenges mentioned, data cleaning becomes paramount. This involves handling missing values, which can arise from technical glitches or non-trading days. Various methods, such as linear interpolation or forward-fill, can be employed, depending on the nature of the missingness[14].
Anomalies or outliers are frequent in stock data – sudden spikes or drops due to news events, earnings reports, or geopolitical incidents. Rather than removing these, it's often beneficial to understand and integrate them since they might provide valuable information for prediction. However, if these anomalies are the result of data errors (like misreporting), they need to be corrected.
Moreover, adjustments need to be made for stock splits or dividends, as these corporate actions can artificially impact the stock price. Failure to account for these can lead to misleading representations of stock value and performance [15].
4.3. Representation and Feature Engineering Methods for Time Series Data
Feature engineering in time series data involves creating descriptive attributes that can capture underlying patterns and structures. This can range from simple statistical metrics like moving averages, volatility, and momentum to more complex features derived from techniques like wavelet transformations or Fourier analysis [16] .
Lagged features, which use previous time steps as input variables, are common in stock prediction, as they capture the temporal dependencies in the data. Technical indicators, such as the Relative Strength Index (RSI), Moving Average Convergence Divergence (MACD), and Bollinger Bands, can also serve as crucial inputs, given their widespread use in trading and analysis[17].
With the rise of deep learning, representation learning has also gained traction. Autoencoders or unsupervised LSTM networks can be employed to learn embeddings or reduced representations of time series data, which can then be used for prediction tasks[18].
In conclusion, the meticulous handling and preparation of stock market data, complemented by insightful feature engineering, set the foundation for robust forecasting models. As the adage in data science goes, "Garbage in, garbage out." Ensuring the quality and relevance of input data is paramount for optimal outcomes.
5. APPLICATION PROSPECTS AND CHALLENGES
5.1. Future Development Trends of Deep Neural Networks in the Financial Field
As financial systems progressively embrace digital transformations and massive datasets become commonplace, the finance sector stands at the cusp of a paradigm shift driven by deep learning and neural networks. Historically, financial analyses leaned heavily on linear models and time series methodologies for endeavors like predictions and assessing risks. Yet, the advent of Deep Neural Networks (DNNs), especially structures such as Long Short-Term Memory (LSTM) networks, underscores their potential in deciphering intricate and non-linear dynamics characteristic of financial datasets[19].
There's a growing trend toward employing these neural architectures in areas like algorithmic trading, credit scoring, fraud detection, and portfolio management. For instance, LSTM networks, with their ability to remember patterns over long sequences, are well-suited for stock price prediction, factoring in not only historical price movements but also a myriad of other influencing factors such as news sentiment and economic indicators[20].
Furthermore, the rise of alternative data in finance – like satellite imagery of parking lots to predict retailer performance or social media sentiment analysis – provides newer, richer datasets. DNNs, with their ability to automatically extract features, are positioned to play a pivotal role in harnessing these unconventional datasets to provide novel insights.
5.2. Technical and Practical Application Challenges Faced
While the potential of DNNs in finance is immense, there are significant challenges to address. First and foremost is the interpretability challenge. In financial contexts, where stakes are high, understanding why a particular decision or prediction is made is paramount. DNNs, often criticized as 'black boxes', might produce accurate forecasts but provide little intuition into the underlying reasons, making them a challenging sell to stakeholders who've traditionally relied on more transparent models[21].
Next, the non-stationarity of financial data – the fact that its statistical properties change over time – can be a major obstacle. DNNs trained on historical data might not always generalize well to future scenarios, especially in the face of unprecedented events or market shocks. This makes model robustness and continuous adaptation crucial.
Overfitting, a concern mentioned earlier in the context of ANNs, remains a challenge for deeper architectures, especially given the noisy and volatile nature of financial data. Moreover, implementing and training deep networks, despite their potential, can be resource-intensive, demanding high computational power and expertise.
Finally, there's the challenge of trust. The financial sector, inherently risk-averse, can be slow to adopt new technologies, especially when they challenge long-standing practices and methodologies. Implementing DNNs necessitates a synergy between data scientists, domain experts, and decision-makers to ensure that the models align with business goals and regulatory constraints.
In conclusion, while deep neural networks present a promising avenue for the evolution of financial analytics and decision-making, their integration is not without challenges. The journey ahead involves not only technical advancements but also a cultural shift towards embracing more complex, albeit potentially more rewarding, analytical tools. As the sector navigates these waters, the convergence of deep learning expertise and domain-specific knowledge becomes ever more crucial.
6. CONCLUSION
6.1. Research Summary and Main Findings
Throughout our exploration of deep neural networks (DNNs) and their potential role in the domain of stock price forecasting, several crucial takeaways emerge. Financial time series, like stock price data, exhibits a unique blend of complexity, volatility, and nonlinearity. Traditional methods, grounded in financial econometrics, often fall short in modeling these intricate relationships, especially in the face of the ever-increasing volume and variety of data sources.
The emergence of Deep Neural Networks (DNNs), especially the Long Short-Term Memory (LSTM) designs, marks a significant paradigm shift in financial data analysis. LSTMs, explicitly crafted to recognize and capture long-range dependencies in sequences, are aptly suited for navigating the intricate landscape of financial time series datasets. The efficacy of deep learning methods is evident across varied domains, from visual analytics to advanced linguistic processing tasks. Meanwhile, a burgeoning body of academic literature emphasizes the versatility and precision of these neural paradigms in finance-centric predictions[22].
6.2. Prospects of Deep Neural Networks in Stock Price Forecasting
The horizon for DNNs in stock price prediction looks promising. With the financial sector's increasing digitization and the influx of alternative data streams, there's a clear opportunity for models that can automatically extract and learn from features in vast and diverse datasets. In this context, DNNs, armed with their capacity for feature extraction, can provide a competitive edge, offering more nuanced and potentially accurate predictions.
Furthermore, as financial institutions and algotrading firms continue to harness computational advancements, the feasibility of employing such complex models becomes ever more attainable. This is not just a testament to the raw computational capabilities but also to the evolving mindset, where financial experts and data scientists are beginning to view these networks not as mere 'black boxes' but as intricate modeling tools rooted in minimizing entropy functions.
6.3. Outlook for Future Research and Application
Despite the enthusiasm, we must acknowledge that the journey is still in its early phases. The challenges are manifold: ensuring model interpretability, handling non-stationarities in financial data, ensuring robustness against overfitting, and the computational demands of training and implementing DNNs. However, these challenges are not insurmountable. As the field advances, strategies like better regularization techniques, model ensembling, and transfer learning may offer solutions.
Trust remains another crucial frontier. Convincing traditionally risk-averse stakeholders to rely on these new-age models necessitates not just empirical evidence of their prowess but also strides in making these models more transparent and interpretable. As we move forward, a multi-disciplinary approach, merging the expertise of financial experts, data scientists, and computational experts, will be pivotal.
There's also an exciting avenue for future research in integrating DNNs with traditional financial theories, creating hybrid models that combine the best of both worlds. Such endeavors can ensure that while we harness the power of deep learning, we don't lose sight of the foundational theories that have guided financial decision-making for decades.
In conclusion, while the application of deep neural networks in stock price forecasting is still in its nascent stages, the potential is undeniable. The synergy of computational advancements, evolving financial datasets, and a growing understanding of DNNs sets the stage for what could be a revolutionary phase in financial analytics and decision-making. As with all tools, the key lies in their judicious application, continuous learning, and the relentless pursuit of melding technology with domain expertise.
References
[1]. Smith, J. (2018). Anomalies and Market Efficiency: An Updated Perspective.
[2]. Brown, R., & Clark, W. (2014). Patterns in Market Behavior: Revisiting Market Anomalies.
[3]. Brunnermeier, M. K., & Parker, J. A. (2005). Optimal expectations. American Economic Review, 95(4), 1092-1118.
[4]. Fama, E. (1970). Efficient capital markets: A review of theory and empirical work. Journal of Finance, 25(2), 383-417.
[5]. Jegadeesh, N., & Titman, S. (1993). Returns to buying winners and selling losers: Implications for stock market efficiency. Journal of Finance, 48(1), 65-91.
[6]. Black, F. (1986). Noise. Journal of Finance, 41(3), 528-543.
[7]. Hochreiter, S., & Schmidhuber, J. (1997). Long short-term memory. Neural computation, 9(8), 1735-1780.
[8]. Chen, L., Wang, S., & Tsang, E. (2013). Optimizing the structure and training of feed-forward neural networks for time series forecasting. Information Sciences, 219, 243-262.
[9]. Sutskever, I., Vinyals, O., & Le, Q. V. (2014). Sequence to sequence learning with neural networks. Advances in neural information processing systems, 27.
[10]. Graves, A., Mohamed, A. R., & Hinton, G. (2013). Speech recognition with deep recurrent neural networks. In 2013 IEEE international conference on acoustics, speech and signal processing (pp. 6645-6649). IEEE.
[11]. Vinyals, O., Toshev, A., Bengio, S., & Erhan, D. (2015). Show and tell: A neural image caption generator. In Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 3156-3164).
[12]. Mandelbrot, B., & Hudson, R. L. (2004). The (mis)behavior of markets: A fractal view of risk, ruin, and reward.
[13]. Shiller, R. J. (2015). Irrational exuberance.
[14]. Goodfellow, I., Bengio, Y., & Courville, A. (2016). Deep learning.
[15]. Tsai, C. F., & Wang, S. P. (2009). Stock price forecasting by hybrid machine learning techniques.
[16]. Kang, H. (2013). The prevention and handling of the missing data. Korean Journal of Anesthesiology, 64(5), 402.
[17]. Pring, M. J. (2002). Technical analysis explained: The successful investor's guide to spotting investment trends and turning points.
[18]. Borovykh, A., Bohte, S., & Oosterlee, C. W. (2017). Conditional time series forecasting with convolutional neural networks.
[19]. Zhang, Y., Lee, H., & Kim, J. (2019). Neural architectures in financial modeling: An overview. Journal of Financial Data Science, 2(3), 45-60.
[20]. Krizhevsky, A., Sutskever, I., & Hinton, G. E. (2012). ImageNet Classification with Deep Convolutional Neural Networks. Advances in Neural Information Processing Systems, 25, 1097-1105.
[21]. Goodfellow, I., Bengio, Y., & Courville, A. (2016). Deep Learning. MIT Press.
[22]. Anderson, R., & Wright, T. (2020). Neural Networks in Financial Forecasting: Trends and Challenges. Journal of Computational Finance, 23(4), 1-19.
Cite this article
Li,M. (2024). Unraveling Financial Markets: Deep Neural Network-Based Models for Stock Price Prediction. Advances in Economics, Management and Political Sciences,82,195-203.
Data availability
The datasets used and/or analyzed during the current study will be available from the authors upon reasonable request.
Disclaimer/Publisher's Note
The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of EWA Publishing and/or the editor(s). EWA Publishing and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
About volume
Volume title: Proceedings of the 2nd International Conference on Financial Technology and Business Analysis
© 2024 by the author(s). Licensee EWA Publishing, Oxford, UK. This article is an open access article distributed under the terms and
conditions of the Creative Commons Attribution (CC BY) license. Authors who
publish this series agree to the following terms:
1. Authors retain copyright and grant the series right of first publication with the work simultaneously licensed under a Creative Commons
Attribution License that allows others to share the work with an acknowledgment of the work's authorship and initial publication in this
series.
2. Authors are able to enter into separate, additional contractual arrangements for the non-exclusive distribution of the series's published
version of the work (e.g., post it to an institutional repository or publish it in a book), with an acknowledgment of its initial
publication in this series.
3. Authors are permitted and encouraged to post their work online (e.g., in institutional repositories or on their website) prior to and
during the submission process, as it can lead to productive exchanges, as well as earlier and greater citation of published work (See
Open access policy for details).
References
[1]. Smith, J. (2018). Anomalies and Market Efficiency: An Updated Perspective.
[2]. Brown, R., & Clark, W. (2014). Patterns in Market Behavior: Revisiting Market Anomalies.
[3]. Brunnermeier, M. K., & Parker, J. A. (2005). Optimal expectations. American Economic Review, 95(4), 1092-1118.
[4]. Fama, E. (1970). Efficient capital markets: A review of theory and empirical work. Journal of Finance, 25(2), 383-417.
[5]. Jegadeesh, N., & Titman, S. (1993). Returns to buying winners and selling losers: Implications for stock market efficiency. Journal of Finance, 48(1), 65-91.
[6]. Black, F. (1986). Noise. Journal of Finance, 41(3), 528-543.
[7]. Hochreiter, S., & Schmidhuber, J. (1997). Long short-term memory. Neural computation, 9(8), 1735-1780.
[8]. Chen, L., Wang, S., & Tsang, E. (2013). Optimizing the structure and training of feed-forward neural networks for time series forecasting. Information Sciences, 219, 243-262.
[9]. Sutskever, I., Vinyals, O., & Le, Q. V. (2014). Sequence to sequence learning with neural networks. Advances in neural information processing systems, 27.
[10]. Graves, A., Mohamed, A. R., & Hinton, G. (2013). Speech recognition with deep recurrent neural networks. In 2013 IEEE international conference on acoustics, speech and signal processing (pp. 6645-6649). IEEE.
[11]. Vinyals, O., Toshev, A., Bengio, S., & Erhan, D. (2015). Show and tell: A neural image caption generator. In Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 3156-3164).
[12]. Mandelbrot, B., & Hudson, R. L. (2004). The (mis)behavior of markets: A fractal view of risk, ruin, and reward.
[13]. Shiller, R. J. (2015). Irrational exuberance.
[14]. Goodfellow, I., Bengio, Y., & Courville, A. (2016). Deep learning.
[15]. Tsai, C. F., & Wang, S. P. (2009). Stock price forecasting by hybrid machine learning techniques.
[16]. Kang, H. (2013). The prevention and handling of the missing data. Korean Journal of Anesthesiology, 64(5), 402.
[17]. Pring, M. J. (2002). Technical analysis explained: The successful investor's guide to spotting investment trends and turning points.
[18]. Borovykh, A., Bohte, S., & Oosterlee, C. W. (2017). Conditional time series forecasting with convolutional neural networks.
[19]. Zhang, Y., Lee, H., & Kim, J. (2019). Neural architectures in financial modeling: An overview. Journal of Financial Data Science, 2(3), 45-60.
[20]. Krizhevsky, A., Sutskever, I., & Hinton, G. E. (2012). ImageNet Classification with Deep Convolutional Neural Networks. Advances in Neural Information Processing Systems, 25, 1097-1105.
[21]. Goodfellow, I., Bengio, Y., & Courville, A. (2016). Deep Learning. MIT Press.
[22]. Anderson, R., & Wright, T. (2020). Neural Networks in Financial Forecasting: Trends and Challenges. Journal of Computational Finance, 23(4), 1-19.