Deep Learning For Financial Applications - A Survey
Deep Learning For Financial Applications - A Survey
Abstract
Computational intelligence in finance has been a very popular topic for both academia and
financial industry in the last few decades. Numerous studies have been published resulting
arXiv:2002.05786v1 [q-fin.ST] 9 Feb 2020
in various models. Meanwhile, within the Machine Learning (ML) field, Deep Learning
(DL) started getting a lot of attention recently, mostly due to its outperformance over the
classical models. Lots of different implementations of DL exist today, and the broad interest
is continuing. Finance is one particular area where DL models started getting traction,
however, the playfield is wide open, a lot of research opportunities still exist. In this paper,
we tried to provide a state-of-the-art snapshot of the developed DL models for financial
applications, as of today. We not only categorized the works according to their intended
subfield in finance but also analyzed them based on their DL models. In addition, we also
aimed at identifying possible future implementations and highlighted the pathway for the
ongoing research within the field.
Keywords: deep learning, finance, computational intelligence, machine learning, financial
applications, algorithmic trading, portfolio management, risk assesment, fraud detection
1. Introduction
Stock market forecasting, algorithmic trading, credit risk assessment, portfolio allocation,
asset pricing and derivatives market are among the areas where ML researchers focused on
developing models that can provide real-time working solutions for the financial industry.
Hence, a lot of publications and implementations exist in the literature.
However, within the ML field, DL is an emerging area with a rising interest every year.
As a result, an increasing number of DL models for finance started appearing in conferences
and journals. Our focus in this paper is to present different implementations of the developed
financial DL models in such a way that the researchers and practitioners that are interested
in the topic can decide which path they should take.
In this paper, we tried to provide answers to the following research questions:
• What are the areas that have promising potentials from an academic/industrial re-
search perspective?
Preprint submitted to Applied Soft Computing February 17, 2020
• Which DL models are preferred (and more successful) in different applications?
3. Deep Learning
Deep Learning is a particular type of ML that consists of multiple ANN layers. It provides
high-level abstraction for data modelling [21]. In the literature, different DL models exist:
Deep Multilayer Perceptron (DMLP), CNN, RNN, LSTM, Restricted Boltzmann Machines
(RBMs), Deep Belief Networks (DBNs), and Autoencoders (AEs).
With multi-layer deep ANNs, more efficient classification and regression performances
are achieved when compared against shallow nets. DMLPs’ learning process is implemented
through backpropagation. The amount of the output error in the output layer neurons is
also reflected back to the neurons in the previous layers. In DMLP, Stochastic Gradient
Descent (SGD) method is (mostly) used for the optimization of learning (to update the
weights of the connections between the layers). In Figure 1, a DMLP model, the layers, the
neurons in layers, the weights between the neurons are shown.
4
Forward Pass Backpropagation
l Output Units l
Wkl
Wkl
k Hidden Units H2 k
Wjk Wjk
j Hidden Units H1 j
Wij Wij
i Input Units i
Figure 1: Deep Multi Layer Neural Network Forward Pass and Backpropagation [21]
0.25 0.50
3x3
3x3
3x3 D D
Fully
Input Convolutional Subsampling Convolutional Subsampling Connected
Layer Convolutional Convolutional
Layer Max
Layer Layer Fully Layer
Layer Dropout
Layer Layer Pooling Connected
5
3.3. Recurrent Neural Network (RNN)
In the literature, RNN has been mostly used on sequential data such as time-series data,
audio and speech data, language. It consists of RNN units that are structured consecu-
tively. Unlike feed-forward networks, RNNs use internal memory to process the incoming
inputs. RNNs are used in the analysis of the time series data in various fields (handwriting
recognition, speech recognition, etc).
There are different types of RNN structures: one to many, many to one, many to many.
Generally, RNN processes the input sequence series one by one at a time, during its op-
eration. Units in the hidden layer hold information about the history of the input in the
"state vector" [21]. RNNs can be trained using the Backpropagation Through Time (BPTT)
method. Using BPTT, the differentiation of the loss at any time t has reflected the weights
of the network at the previous time. Training of RNNs are more difficult than Feedforward
Neural Networks (FFNNs) and the training period of RNNs takes longer.
In Figure 3, the information flow in the RNN’s hidden layer is divided into discrete times.
The status of the node S at different times of t is shown as st , the input value x at different
times is xt , and the output value o at different times is shown as ot . The parameter values
(U, W, V ) are always used in the same step.
Output: o
ot-1 ot ot+1
V V V V
s st-1 st st+1
W W W W W
U U U U
6
Output of
Current Block
(Ht)
X X
σ σ tanh σ
Output of Output of
Previous Block Current Block
(Ht-1) (Ht)
Input (Xt)
7
c c1 c2 c3 cm
h1 h2 h3 hm
Wmxn
v1 v2 vn
b b1 b2 bn
v1 v2 vn
Encoder Decoder
Input Output
Code
4. Financial Applications
There are a lot of financial applications of soft computing in the literature. DL has been
studied in most of them, although, some opportunities still exist in a number of fields.
Throughout this section, we categorized the implementation areas and presented them
in separate subsections. Besides, in each subsection we tabulated the representative features
of the relevant studies in order to provide as much information as possible in the limited
space.
9
Also, the readers should note that there were some overlaps between different imple-
mentation areas for some papers. There were two main reasons for that: In some papers,
multiple problems were addressed separately, for e.g. text mining was studied for feature
extraction, then algorithmic trading was implemented. For some other cases, the paper
might fit directly into multiple implementation areas due to the survey structure, for e.g.
cryptocurrency portfolio management. In such cases we included the papers in all of the
relevant subsections creating some overlaps.
Some of the existing study areas can be grouped as follows:
Art. Data Set Period Feature Set Method Performance Cri- Environment
teria
[35] GarantiBank in BIST, 2016 OCHLV, Spread, PLR, Graves MSE, RMSE, MAE, Spark
Turkey Volatility, Turnover, LSTM RSE, Correlation R-
etc. square
[36] CSI300, Nifty50, HSI, 2010-2016 OCHLV, Technical WT, Stacked au- MAPE, Correlation -
Nikkei 225, S&P500, Indicators toencoders, LSTM coefficient, THEIL-
DJIA U
[37] Chinese Stocks 2007-2017 OCHLV CNN + LSTM Annualized Return, Python
Mxm Retracement
[38] 50 stocks from NYSE 2007-2016 Price data SFM MSE -
[39] The LOB of 5 stocks of 2010 FI-2010 dataset: WMTR, MDA Accuracy, Precision, -
Finnish Stock Market bid/ask and volume Recall, F1-Score
[40] 300 stocks from SZSE, 2014-2015 Price data FDDR, DNN+RL Profit, return, SR, Keras
Commodity profit-loss curves
[41] S&P500 Index 1989-2005 Price data, Volume LSTM Return, STD, SR, Python,
Accuracy TensorFlow,
Keras, R,
H2O
[42] Stock of National Bank 2009-2014 FTSE100, DJIA, GASVR, LSTM Return, volatility, Tensorflow
of Greece (ETE). GDAX, NIKKEI225, SR, Accuracy
EUR/USD, Gold
[43] Chinese stock-IF-IH-IC 2016-2017 Decisions for price MODRL+LSTM Profit and loss, SR -
contract change
[44] Singapore Stock Market 2010-2017 OCHL of last 10 days DNN RMSE, MAPE, -
Index of Index Profit, SR
[45] GBP/USD 2017 Price data Reinforcement SR, downside de- Python,
Learning + LSTM viation ratio, total Keras, Ten-
+ NES profit sorflow
[46] Commodity, FX future, 1991-2014 Price Data DNN SR, capability ratio, C++,
ETF return Python
[47] USD/GBP, S&P500, 2016 Price data AE + CNN SR, % volatility, avg H2O
FTSE100, oil, gold return/trans, rate of
return
[48] Bitcoin, Dash, Ripple, 2014-2017 MA, BOLL, the LSTM, RNN, Accuracy, F1- Python, Ten-
Monero, Litecoin, Doge- CRIX returns, Eu- MLP measure sorflow
coin, Nxt, Namecoin ribor interest rates,
OCHLV
[49] S&P500, KOSPI, HSI, 1987-2017 200-days stock price Deep Q-Learning, Total profit, Corre- -
and EuroStoxx50 DNN lation
[50] Stocks in the S&P500 1990-2015 Price data DNN, GBT, RF Mean return, MDD, H2O
Calmar ratio
[51] Fundamental and Tech- - Fundamental , tech- CNN - -
nical Data, Economic nical and market in-
Data formation
Using a different model, Zhang et. al. [38] proposed a novel State Frequency Memory
(SFM) recurrent network for stock price prediction with multiple frequency trading patterns
and achieved better prediction and trading performances. In an HFT trading system, Tran
et al. [39] developed a DL model that implements price change forecasting through mid-
price prediction using high-frequency limit order book data with tensor representation. In
[40], the authors used Fuzzy Deep Direct Reinforcement Learning (FDDR) for stock price
prediction and trading signal generation.
For index prediction, the following studies are noteworthy. In [41], the price prediction
11
of S&P500 index using LSTM was implemented. Mourelatos et al. [42] compared the
performance of LSTM and GA with a SVR (GASVR) for Greek Stock Exchange Index
prediction. Si et al. [43] implemented Chinese intraday futures market trading model with
DRL and LSTM. Yong et al. [44] used feed-forward DNN method and Open,Close,High,
Low (OCHL) of the time series index data to predict Singapore Stock Market index data.
Forex or cryptocurrency trading was implemented in some studies. In [45], agent inspired
trading using deep (recurrent) reinforcement learning and LSTM was implemented and
tested on the trading of GBP/USD. In [46], feedforward deep MLP was implemented for
the prediction of commodities and FX trading prices. Korczak et al. [47] implemented a
forex trading (GBP/PLN) model using several different input parameters on a multi-agent-
based trading environment. One of the agents was using CNN as the prediction model and
outperformed all other models.
On the cryptocurrency side, Spilak et al. [48] used several cryptocurrencies (Bitcoin,
Dash, Ripple, Monero, Litecoin, Dogecoin, Nxt, Namecoin) to construct a dynamic portfolio
using LSTM, RNN, MLP methods.
In a versatile study, Jeong et al. [49] combined deep Q-learning and DNN to implement
price forecasting and they intended to solve three separate problems: Increasing profit in
a market, prediction of the number of shares to trade, and preventing overfitting with
insufficient financial data.
In [52], technical analysis indicator’s (Relative Strength Index (RSI)) buy & sell limits
were optimized with GA which was used for buy-sell signals. After optimization, DMLP was
also used for function approximation. In [53], the authors combined deep Fully Connected
Neural Network (FNN) with a selective trade strategy unit to predict the next price. In
[54], the crossover and Moving Average Convergence and Divergence (MACD) signals were
used to predict the trend of the Dow 30 stocks’ prices. Sirignano et al. [55] proposed a novel
method that used limit order book flow and history information for the determination of
the stock movements using LSTM model. Tsantekidis et al. [56] also used limit order book
time series data and LSTM method for the trend prediction.
Several studies focused on utilizing CNN based models due to their success in image
classification problems. However, in order to do that, the financial input data needed to be
transformed into images which required some creative preprocessing. Gudelek et al. [57]
converted time series of price data to 2-dimensional images using technical analysis and
classified them with deep CNN. Similarly, Sezer et al. [58] also proposed a novel technique
that converts financial time series data that consisted of technical analysis indicator outputs
to 2-dimensional images and classified these images using CNN to determine the trading
signals. In [59], candlestick chart graphs were converted into 2-dimensional images. Then,
unsupervised convolutional AE was fed with the images to implement portfolio construction.
Tsantekidis et al. [60] proposed a novel method that used the last 100 entries from the limit
order book to create a 2-dimensional image for the stock price prediction using CNN method.
In [61], an innovative method was proposed that uses CNN with correlated features combined
together to predict the trend of the stocks prices. Finally, Sezer et al. [62] directly used bar
chart images as inputs to CNN and predicted if the image class was Buy, Hold or Sell, hence
a corresponding Algo-trading model was developed.
12
Table 2: Classification (Buy-sell Signal, or Trend Detection) Based Algo-trading Models
[52] Stocks in Dow30 1997-2017 RSI DMLP with ge- Annualized Spark MLlib, Java
netic algorithm return
[53] SPY ETF, 10 stocks 2014-2016 Price data FFNN Cumulative gain MatConvNet,
from S&P500 Matlab
[54] Dow30 stocks 2012-2016 Close data and LSTM Accuracy Python, Keras,
several technical Tensorflow,
indicators TALIB
[55] High-frequency record 2014-2017 Price data, record LSTM Accuracy -
of all orders of all orders, trans-
actions
[56] Nasdaq Nordic (Kesko 2010 Price and volume LSTM Precision, Re- -
Oyj, Outokumpu Oyj, data in LOB call, F1-score,
Sampo, Rautaruukki, Cohen’s k
Wartsila Oyj)
[57] 17 ETFs 2000-2016 Price data, techni- CNN Accuracy, MSE, Keras, Tensorflow
cal indicators Profit, AUROC
[58] Stocks in Dow30 and 9 1997-2017 Price data, techni- CNN with feature Recall, precision, Python, Keras,
Top Volume ETFs cal indicators imaging F1-score, annual- Tensorflow, Java
ized return
[59] FTSE100 2000-2017 Price data CAE TR, SR, MDD, -
mean return
[60] Nasdaq Nordic (Kesko 2010 Price, Volume CNN Precision, Re- Theano, Scikit
Oyj, Outokumpu Oyj, data, 10 orders of call, F1-score, learn, Python
Sampo, Rautaruukki, the LOB Cohen’s k
Wartsila Oyj)
[61] Borsa Istanbul 100 2011-2015 75 technical indi- CNN Accuracy Keras
Stocks cators and OCHLV
[62] ETFs and Dow30 1997-2007 Price data CNN with feature Annualized Keras, Tensorflow
imaging return
[63] 8 experimental assets - Asset prices data RL, DNN, Genetic Learning and -
from bond/derivative Algorithm genetic algorithm
market error
[64] 10 stocks from S&P500 - Stock Prices TDNN, RNN, Missed oppor- -
PNN tunities, false
alarms ratio
[65] London Stock Exchange 2007-2008 Limit order book CNN Accuracy, kappa Caffe
state, trades,
buy/sell orders,
order deletions
[66] Cryptocurrencies, Bit- 2014-2017 Price data CNN, RNN, Accumulative -
coin LSTM portfolio value,
MDD, SR
Serrano et al. [63] proposed a novel method called “GoldAI Sachs” Asset Banker Rein-
forcement Learning Algorithm for algorithmic trading. The proposed method used a ran-
dom neural network, GP, and Reinforcement Learning (RL) to generate the trading signals.
Saad et al. [64] compared Timedelay Neural Network (TDNN), RNN and Probabilistic
Neural Network (PNN) for trend detection using 10 stocks from S&P500. In [65], HFT
microstructures forecasting with CNN method was performed. In [66], cryptocurrency port-
folio management based on three different proposed models (basic RNN, LSTM and CNN)
was implemented.
Tino et al. [67] used The Deutscher Aktienindex (DAX), London Financial Times Stock
Exchange Index (FTSE)100, call and put options prices to predict the changes with Markov
models and used the financial time series data to predict volatility changes with RNN.
Meanwhile, Chen et al. [68] proposed a method that uses a filterbank CNN Algorithm on
13
15x15 volatility times series converted synthetic images. In the study, the financial domain
knowledge and filterbank mechanism were combined to determine the trading signals. Bari
et al. [69] used text mining to extract information from the tweets and financial news and
used LSTM, RNN, Gated-Recurrent Unit (GRU) for the generation of the trading signals.
Dixon et al. [70] used RNN for the sequence classification of the limit order book to predict
a next event price-flip.
Table 3: Stand-alone and/or Other Algorithmic Models
Chen et al. [71] used 1-dimensional CNN with an agent-based RL algorithm on the
Taiwan stock index futures (TAIFEX) dataset. Wang et al. [72] proposed a Deep Co-
investment Network Learning (DeepCNL) method that used convolutional and RNN layers.
The investment pattern was determined using the extracted Rise-Fall trends. Day et al. [73]
used financial sentiment analysis using text mining and DNN for stock algorithmic trading.
Sirignano et al. [74] proposed a “spatial neural network” model that used limit order book
and spatial features for algorithmic trading. Their model estimates the best bid-ask prices
using bid, ask prices in the limit order book. Gao et al. [75] used GRU, LSTM units, CNN,
and MLP to model Q values for the implementation of the DRL method.
Before going into the details about specific DL implementations, it is worthwhile to men-
tion the existing ML surveys on the topic. Kirkos et al. [84], Ravi et al. [85], Fethi et al. [86]
reviewed the bank performance assessment studies based on Artificial Intelligence (AI) and
ML models. Lahsasna et al. [87], Chen et al.[88] surveyed the credit scoring and credit risk
assessment studies based on soft computing techniques whereas Marques et. al. [89] focused
only on Evolutionary Computation (EC) Models for credit scoring implementations. Mean-
while, Kumar et al. [90], Verikas et al. [91] reviewed ML implementations of bankruptcy
prediction studies. Similarly, Sun et al. [92] provided a comprehensive survey about research
on financial distress and corporate failures. Apart from these reviews, for assessing overall
risk, Lin et al. [93] surveyed the financial crisis prediction studies based on ML models.
Since risk assessment is becoming vital for survival in today’s financial world, a lot of
researchers turned their attention to DL for higher accuracy. Table 4, Table 5 provide
snapshot information about the different risk assessment studies implemented using various
DL models.
For credit score classification (Table 4), Luo et al. [77], used CDS data for Corporate
15
Credit rating and corresponding credit classification (A,B or C). Among the tested models,
DBN with RBM performed the best. This implementation was probably the first study to
implement Credit rating with DBN. Similarly, in [78], a cascaded hybrid model of DBN,
Backpropagation and SVM for credit classification was implemented and good performance
results (the accuracy was above 80-90 %) were achieved. In [79], credit risk classification
was achieved by using an ensemble of deep MLP networks each using subspaces of the
whole space by k-means (using minority class in each, but only a partial subspace of the
majority class). The data imbalance problem was handled by using multiple subspaces for
each classifier, where each of them had all the positive (minor) instances, but a subsample
of negative (majority) instances, finally they used an ensemble of deep MLPs combining
each subspace model. In [80], credit scoring was performed using a SAE network and GP
model to create credit assessment rules in order to generate good or bad credit cases. In
another study, Neagoe et. al. [81] classified credit scores using various DMLP and deep CNN
networks. In a different study [82], consumer credit scoring classification was implemented
with a 2-D representation of the input consumer data through transforming the data into a
2-D pixel matrix. Then the resulting images were used as the training and test data for CNN.
2-D pixel matrix representation of the consumer data was adapted by using CNN for image
classification. This was the first implementation of credit scoring using CNN. Niimi [83]
used UCI credit approval dataset 1 to compare DL, SVM, Logistic Regression (LR), Random
Forest (RF), eXtreme Gradient Boosting (XGBoost) and provided information about credit
fraud and credit approval applications; then experimented with the credit approval problem
with several models. Various models were compared for credit approval classification. Also,
some introduction about credit fraud detection was provided.
Financial distress prediction for banks and corporates are studied extensively (Table 5).
In [94], a hybrid DBN with SVM was used for financial distress prediction to identify whether
the firm was in trouble or not, whereas bank risk classification was studied in [95]. In [96],
news semantics were extracted by the word sequence learning and associated events were
labeled with the bank stress, then from the formed semantic vector representation, the bank
stress was determined and classified against a threshold. Prediction and semantic meaning
extraction were integrated in a neat way. In another study [97], text mining was again used
for identifying the bank distress by extracting the data from financial news and then using
a Deep Feed Forward Network (DFFN) on semantic sentence vectors extracted from word
embeddings to classify if there was an event or not. Similarly, Cerchiello et al. [98] used
text mining from the financial news to classify bank distress. Malik et al. [99] evaluated
the bank stress by first predicting the bank’s performance through an LSTM network, then
Backpropagation network was used for finding the bank stress level.
Table 5: Financial Distress, Bankruptcy, Bank Risk, Mortgage Risk, Crisis Forecasting Studies
1
https://archive.ics.uci.edu/ml/datasets.html
16
Table 5: Financial Distress, Bankruptcy, Bank Risk, Mortgage Risk, Crisis Forecasting Studies
There are also a number of research papers that were focused on bankruptcy or corporate
default prediction. Ribeiro et al. [100] implemented bankruptcy prediction with DBN. The
results of DBN were compared with SVM and RBM. Yeh et al. [101] used the stock returns
of default and solvent companies as inputs to RBM used as SAE, then the output of RBM
was used as input to DBN to predict if the company was solvent or default. The results
were compared with an SVM model and the DBN model outperformed SVM. Hosaka et al.
[102] tried a different approach by converting the financial data to the image to use CNN
for bankruptcy prediction.
The remaining implementations of risk assessment are as follows: Sirignano et al. [103]
used the mortgage application data of 20 years for identifying the mortgage risk using various
parameters. They also performed a lot of analyses relating different factors that affected the
mortgage payment structure. The authors also analyzed the prepayment and delinquency
behavior in their assessment. For another mortgage risk assessment application, Kvamme
et al. [104] used CNN and RF models to predict whether a customer would default on its
mortgage or not. In a different study, Abroyan et al. [105] used CNN and LSTM networks
17
to classify if a transaction performed on the stock market (trade) was risky or not and high
accuracy was achieved. Finally, Chatzis et al. [106] developed several ML and DL models for
detecting events that caused the stock market to crash. DL models had good classification
(detecting crisis or not) performance.
18
Table 6: Fraud Detection Studies
Portfolio selection and smart indexing were the main focuses of [130] and [131] using AE
and LSTM networks. Lin et al. [132] used the Elman network for optimal portfolio selection
by predicting the stock returns for t+1 and then constructing the optimum portfolio ac-
cording to the returns. Meanwhile, Maknickiene et al. [141] used Evolino RNN for portfolio
selection and return prediction accordingly. The selected portfolio components (stocks) were
orthogonal in nature.
In [134], through predicting the next month’s return, top to be performed portfolios were
constructed and good monthly returns were achieved with LSTM and LSTM-MLP combined
DL models. Similarly, Batres et al. [135] combined DBN and MLP for constructing a stock
portfolio by predicting each stock’s monthly log-return and choosing the only stocks that
were expected to perform better than the performance of the median stock. Lee et al.
20
[136] compared 3 RNN models (S-RNN, LSTM, GRU) for stock price prediction and then
constructed a threshold-based portfolio with selecting the stocks according to the predictions.
With a different approach, Iwasaki et al. [137] used the analyst reports for sentiment analyses
through text mining and word embeddings and used the sentiment features as inputs to Deep
Feedforward Neural Network (DFNN) model for the stock price prediction. Then different
portfolio selections were implemented based on the projected stock returns.
DRL was selected as the main DL model for [138]. Liang et al. [138] used DRL for
portfolio allocation by adjusting the stocks weights using various RL models. Chen et al.
[139] compared different ML models (including DFFN) for hedge fund return prediction and
hedge fund selection. DL and RF models had the best performance.
Cryptocurrency portfolio management also started getting attention from DL researchers.
In [140], portfolio management (allocation and adjustment of weights) was implemented
by CNN and DRL on selected cryptocurrencies. Similarly, Jiang et al. [66] implemented
cryptocurrency portfolio management (allocation) based on 3 different proposed models,
namely RNN, LSTM and CNN.
4.5. Asset Pricing and Derivatives Market (options, futures, forward contracts)
Accurate pricing or valuation of an asset is a fundamental study area in finance. There
are a vast number of ML models developed for banks, corporates, real estate, derivative
products, etc. However, DL has not been applied to this particular field and there are
some possible implementation areas that DL models can assist the asset pricing researchers
or valuation experts. There were only a handful of studies that we were able to pinpoint
within the DL and finance community. There are vast opportunities in this field for future
studies and publications.
Meanwhile, financial models based on derivative products is quite common. Options
pricing, hedging strategy development, financial engineering with options, futures, forward
contracts are among some of the studies that can benefit from developing DL models. Some
recent studies indicate that researchers started showing interest in DL models that can
provide solutions to this complex and challenging field. Table 8 summarizes these studies
with their intended purposes.
Table 8: Asset Pricing and Derivatives Market Studies
Art. Der.Type Data Set Period Feature Set Method Performance Env.
Criteria
[137] Stock ex- Analyst reports 2016-2018 Text LSTM, CNN, Bi- Accuracy, R2 R,
change on the TSE and LSTM Python,
Osaka Exchange MeCab
[142] Options Simulated a - Price data, option DNN RMSE, the av- Tensorflow
range of call strike/maturity, erage percentage
option prices dividend/risk free pricing error
rates, volatility
[143] Futures, TAIEX Options 2017 OCHLV, fundamen- MLP, MLP with RMSE, MAE, -
Options tal analysis, option Black scholes MAPE
price
[144] Equity re- Returns in 1975-2017 57 firm characteris- Fama-French R2 ,RMSE Tensorflow
turns NYSE, AMEX, tics n-factor model
NASDAQ DL
21
Iwasaki et al. [137] used a DFNN model and the analyst reports for sentiment analyses
to predict the stock prices. Different portfolio selection approaches were implemented after
the prediction of the stock prices. Culkin et al. [142] proposed a novel method that used
feedforward DNN model to predict option prices by comparing their results with Black
& Scholes option pricing formula. Similarly, Hsu et al. [143] proposed a novel method
that predicted TAIEX option prices using bid-ask spreads and Black & Scholes option price
model parameters with 3-layer DMLP. In [144], characteristic features such as Asset growth,
Industry momentum, Market equity, Market Beta, etc. were used as inputs to a Fama-French
n-factor model DL to predict US equity returns in National Association of Securities Dealers
Automated Quotations (NASDAQ), American Stock Exchange (AMEX), New York Stock
Exchange (NYSE) indices.
22
Table 9: Cryptocurrency and Blockchain Studies
Chen et al. [145] proposed a blockchain transaction traceability algorithm using Takagi-
Sugeno fuzzy cognitive map and 3-layer DMLP. Bitcoin data (Hash value, bitcoin address,
public/private key, digital signature, etc.) was used as the dataset. Nan et al. [146] proposed
a method for bitcoin mixing detection that consisted of different stages: Constructing the
Bitcoin transaction graph, implementing node embedding, detecting outliers through AE.
Lopes et al. [147] combined the opinion market and price prediction for cryptocurrency
trading. Text mining combined with 2 models, CNN and LSTM were used to extract the
opinion. Bitcoin, Litecoin, StockTwits were used as the dataset. Open,Close,High, Low,
Volume (OCHLV) of prices, technical indicators, and sentiment analysis were used as the
feature set.
In another study, Jiang et al. [66] presented a financial-model-free RL framework for the
Cryptocurrency portfolio management that was based on 3 different proposed models, basic
RNN, LSTM and CNN. In [140], portfolio management was implemented by CNN and DRL
on 12 most-volumed cryptocurrencies. Bitcoin, Ethereum, Bitcoin Cash and Digital Cash
were used as the dataset.
In addition, Spilak et al. [48] used 8 cryptocurrencies (Bitcoin, Dash, Ripple, Monero,
Litecoin, Dogecoin, Nxt, Namecoin) to construct a dynamic portfolio using LSTM, RNN,
MLP methods. McNally et al. [148] compared Bayesian optimized RNN, LSTM and Au-
toregressive Integrated Moving Average (ARIMA) to predict the bitcoin price direction.
Sensitivity, specificity, precision, accuracy, Root Mean Square Error (RMSE) were used as
the performance metrics.
24
Table 10: Financial Sentiment Studies coupled with Text Mining for Forecasting
Even though financial sentiment is highly coupled with text mining, we decided to repre-
sent those two topics in different subsections. The main reason for such a choice is not only
the existence of some financial sentiment studies which do not directly depend on financial
textual data (like [158]) but also the existence of some financial text mining studies that are
not automatically used for sentiment analysis which will be covered in the next section.
25
directional movement estimation. Akita et al. [171] implemented a method that used text
mining and price prediction together for forecasting prices. Verma et al. [172] combined
news data with financial data to classify the stock price movement. Bari et al. [69] used
text mining for extracting information from the tweets and news. In the method, time
series models were used for stock trade signal generation. In [173], a method that performed
information fusion from news and social media sources was proposed to predict the trend of
the stocks.
In [174], social media news were used to predict the index price and the index direction
with RNN-Boost through Latent Dirichlet Allocation (LDA) features. Hu et al. [175]
proposed a novel method that used text mining techniques and Hybrid Attention Networks
based on the financial news for forecasting the trend of stocks. Li et al. [176] implemented
intraday stock price direction classification using the financial news and stocks prices. In
[177], financial news data and word embedding with Word2vec were implemented to create
the inputs for Recurrent CNN (RCNN) to predict the stock price.
Table 11: Text Mining Studies without Sentiment Analysis for Forecasting
26
Table 11: Text Mining Studies without Sentiment Analysis for Forecasting
Minami et al. [178] proposed a method that predicted the stock price with corporate
action event information and macro-economic index data using LSTM. In [179], a novel
method that used a combination of RBM, DBN and word embeddings to create word vectors
for RNN-RBM-DBN network was proposed to predict the stock prices. Buczkowski et al.
[180] proposed a novel method that used expert recommendations, ensemble of GRU and
LSTM for prediction of the prices.
In [181] a novel method that used character-based neural language model using financial
news and LSTM was proposed. Liu et al. [182] proposed a method that used word em-
beddings with word2Vec, technical analysis features and stock prices for price prediction.
In [183], Deep Neural Generative Model (DGM) with news articles using Paragraph Vector
algorithm was used for creation of the input vector to predict the stock prices. In [184],
the stock price data and word embeddings were used for stock price prediction. The results
showed that the extracted information from embedding news improves the performance.
Table 12: Financial Sentiment Studies coupled with Text Mining without Forecasting
27
Table 12: Financial Sentiment Studies coupled with Text Mining without Forecasting
Akhtar et al. [185] compared CNN, LSTM and GRU based DL models against MLP for
financial sentiment analysis. Rawte et al. [95] tried to solve three separate problems using
CNN, LSTM, SVM, RF: Bank risk classification, sentiment analysis and Return on Assets
(ROA) regression.
Chang et al. [186] implemented the estimation of information content polarity (nega-
tive/positive effect) with text mining, word vector, lexical, contextual input and various
LSTM models. They used the financial news from Reuters.
Jangid et al. [187] proposed a novel method that is a combination of LSTM and CNN
for word embedding and sentiment analysis using Bidirectional LSTM (Bi-LSTM) for aspect
extraction. The proposed method used multichannel CNN for financial sentiment analysis.
Shijia et al. [188] used an attention-based LSTM for the financial sentiment analysis using
news headlines and microblog messages. Sohangir et al. [189] used LSTM, doc2vec, CNN
and stock market opinions posted in StockTwits for sentiment analysis. Mahmoudi et al.
[190] extracted tweets from StockTwits to identify the user sentiment. In the evaluation
approach, they also used emojis for the sentiment analysis. Kitamori et al. [191] extracted
the sentiments from financial news and used DNN to classify positive and negative news.
In [192], the sentiment/aspect prediction was implemented using an ensemble of LSTM,
CNN and GRU networks. In a different study, Li et al. [193] proposed a DL based sentiment
analysis method using RNN to identify the top sellers in the underground economy. Moore
et al. [194] used text mining techniques for sentiment analysis from the financial news.
Table 13: Other Text Mining Studies
28
Table 13: Other Text Mining Studies
In [195], individual social security payment types (paid, unpaid, repaid, transferred) were
classified and predicted using LSTM, HMM and SVM. Sohangir et al. [196] used two neural
network models (doc2Vec, CNN) to find the top authors in StockTwits messages and to
classify the authors as expert or non-expert for author classification purposes.
In [123], the character sequences in financial transactions and the responses from the
other side was used to detect if the transaction was fraud or not with LSTM. Wang et al.
[121] used text mining and DNN models to detect automobile insurance fraud.
In [96], the news semantics were extracted by the word sequence learning, bank stress
was determined and classified with the associated events. Day et al. [73] used financial
sentiment analysis using text mining and DNN for stock algorithmic trading.
Cerchiello et al. [98] used the fundamental data and text mining from the financial news
(Reuters) to classify the bank distress. In [97], the bank distress was identified by extracting
the data from the financial news through text mining. The proposed method used DFNN
on semantic sentence vectors to classify if there was an event or not.
29
Table 14: Other - Theoretical or Conceptual Studies
Art. Subtopic Data Set Period Feature Set Method Performance Env.
Criteria
[49] Improving trad- S&P500, KOSPI, HSI, 1987-2017 200-days stock price Deep Q- Total profit, -
ing decisions and EuroStoxx50 Learning and Correlation
DNN
[193] Identifying Forums data 2004-2013 Sentences and key- Recursive Precision, -
Top Sellers In words neural tensor recall, f-
Underground networks measure
Economy
[195] Predicting Taiwan’s National Pen- 2008-2014 Insured’s id, area- RNN Accuracy, to- Python
Social Ins. Pay- sion Insurance code, gender, etc. tal error
ment Behavior
[199] Speedup 45 CME listed commod- 1991-2014 Price data DNN - -
ity and FX futures
30
Table 15: Other Financial Applications
Art. Subtopic Data Set Period Feature Set Method Performance Env.
Criteria
[200] Forecasting Stocks in NYSE, NAS- 1970-2017 16 fundamental fea- MLP, LFM MSE, Com- -
Fundamentals DAQ or AMEX ex- tures from balance pount annual
changes sheet return, SR
[201] Predicting Bank Phone calls of bank mar- 2008-2010 16 finance-related CNN Accuracy -
Telemarketing keting data attributes
[202] Corporate 22 pharmaceutical com- 2000-2015 11 financial and 4 RBM, DBN RMSE, profit -
Performance panies data in US stock patent indicator
Prediction market
All-years count
Last-3-years count
Publication Count
40
30
20
10
0
f in
al
r is
f in
po
fra
ot
cr
as
ot
go
he
yp
he
se
k
r tf
an
an
ud
tp
r it
rf
r
as
to
ol
ci
ci
de
cu
th
se
hm
in
r ic
io
al
al
te
eo
sm
an
r
te
i
r
ic
ct
ng
en
en
ry
c
xt
an
e
io
ia
tr a
nt
tim
cy
an
or
ag
m
la
di
d
in
an
en
em
co
pp
ng
de
in
d
ta
nc
lic
e
g
r
bl
n
i
na
ep
at
va
t
oc
io
ly
tu
tiv
kc
ns
sis
al
es
ha
stu
Topic Name
in
dy
First and foremost, we clustered the various topics within the financial applications
research and presented them in Figure 8. A quick glance at the figure shows us financial text
mining and algorithmic trading are the top two fields that the researchers most worked on
followed by risk assessment, sentiment analysis, portfolio management and fraud detection,
respectively. The results indicate most of the papers were published within the last 3 years
implying the domain is very hot and actively studied. We can also observe these phenomena
by analyzing Figure 9. Also, it is worth to mention that the few papers that were published
before 2013 all used RNN based models.
31
Histogram of Publication Count in Years
60
Publication Count in Year
50
40
30
20
10
0
19
19
20
20
20
20
20
20
20
20
20
20
20
20
20
20
20
20
20
20
20
98
99
00
01
02
03
04
05
06
07
08
09
10
11
12
13
14
15
16
17
18
Year
60 All-years count
Last-3-years count
50
Publication Count
40
30
20
10
0
RN
CN
RL
RB
M
th
BN
M
N
er
LP
Model Type
When the papers were clustered by the DL model type as presented in Figure 10, we
observe the dominance of RNN, DMLP and CNN over the remaining models, which might
be expected, since these models are the most commonly preferred ones in general DL im-
plementations. Meanwhile, RNN is a general umbrella model which has several versions
including LSTM, GRU, etc. Within the RNN choice, most of the models actually belonged
32
to LSTM, which is very popular in time series forecasting or regression problems. It is also
used quite often in algorithmic trading. More than 70% of the RNN papers consisted of
LSTM models.
Figure 11 presents the commonly used software and frameworks for DL model implemen-
tations through Wordcloud whereas Figure 12 provides the details about the development
environments. The left chart (Figure 12a) presents the high level view where Python had
the lion’s share with 80% over R (with 10%) and the other languages. The chart on the
right (Figure 12b) provides the details about how the developers are using Python through
different libraries and frameworks.
2.9 9.6
2.
2 25.7
2. 28.4
1.5 2
1.5
Python Based
All in One
0.9
0.9
21.1 0.9
2. 0.9
8
0.9
3.7
80.1 7.3 3.7 0.9
0.9
0.9
33
Meanwhile, DMLP generally fits well for classification problems; hence it is a common
choice for most of the financial application areas. However, since it is a natural extension of
its shallow counterpart MLP, it has a longer history than the other DL models.
Engineering Letters -
Frontiers in Signal Processing -
Future Generation Computer Systems 5.768
IEEE Access 4.098
IEEE Transactions on Industrial Informatics 7.377
IEEE Transactions on Knowledge and Data Engineering 3.857
IEEE Transactions on Neural Networks and Learning Systems 11.683
IEICE Transactions on Information and Systems -
International Journal of Computer Applications 3.12
International Journal of Intelligent Systems and Applications in Engineering -
International Journal of Machine Learning and Computing -
Journal of Big Data -
Journal of Computational Science 2.502
Journal of Mathematical Finance 0.39
Neural Computing and Applications 4.664
Pattern Recognition Letters 2.810
Plos One 2.776
Sustainability 2.592
The Journal of Supercomputing 2.16
0 1 2 3 4 5
Journal Count
Last-3-years count Other-years count
Figure 13: Top Journals - corresponding numbers next to the bar graph are representing the impact factor
of the journals
CNN started getting more attention lately since most of the implementations appeared
within the last 3 years. Careful analysis of CNN papers indicates that a recent trend of
representing financial data with a 2-D image view in order to utilize CNN is growing. Hence
34
CNN based models might overpass the other models in the future. It actually passed DMLP
for the last 3 years.
The top journals are tabulated in Fig 13. The journals with the most published papers in
the last 3 years include Expert Systems with Applications, Decision Support Systems, Ap-
plied Soft Computing, Neurocomputing, Knowledge-based Systems and European Journal
of Operational Research.
36
and provides opportunities. It would be beneficial to further explore the possibilities using
that approach in different problems. The playfield is still wide open.
Graph CNN is another model that is closely related but still showing some discrepan-
cies. It has not been used much, only one study was published that relates graph-CNN with
financial applications. However, versatile transformations of financial data into graphs, inte-
grating sentiment analysis through graph representations and constructing different models
can create opportunities for researchers to build better performing financial applications.
There are also recently developed DL models, like GAN, Capsule networks, etc. that
can also provide viable alternatives to existing implementations. They have started showing
up in various non-financial studies, however to the best of our knowledge, no known imple-
mentation of such kind for financial applications exists. It might open up a new window
of opportunities for financial researchers and practitioners. In addition to such new mod-
els, innovative paradigms like transfer learning, one-shot learning can be tested within the
environment.
Since financial text mining is overtaking the other topics in an accelerated fashion, new
data models like Stock2Vec [168] can be enhanced for better and more representative models.
In addition, Natural Language Processing (NLP) based ensemble models or more integration
of data semantics into the picture can increase the accuracy of the existing models.
Finally, according to our observations, hybrid models are preferred more over the native
or standalone models in most studies. This trend will likely continue, however, researchers
need to introduce more versatile, sometimes unconventional models for better results. Hybrid
models integrating various simple DL layers like cascaded CNN-LSTM blocks can have better
outcomes since ensembling spatial and temporal information together in a novel way might
be an important milestone for researchers seeking for "alpha" in their models.
38
• What financial application areas are of interest to DL community?
Response: Financial text mining, Algo-trading, risk assessments, sentiment analysis,
portfolio management and fraud detection are among the most studied areas of finance
research. (Please check Figure 8)
• What are the areas that have promising potentials from an academic/industrial re-
search perspective?
Response: Cryptocurrencies, blockchain, behavioral finance, HFT and derivatives mar-
ket have promising potentials for research.
7. Conclusions
The financial industry and academia have started realizing the potentials of DL in various
application areas. The number of research work keeps on increasing every year with an
accelerated fashion. However, we are just in the early years of this new era, more studies will
be implemented and new models will keep pouring in. In this survey, we wanted to highlight
the state-of-the-art DL research for the financial applications. We not only provided a
snapshot of the existing research status but also tried to identify the future roadway for
intended researchers. Our findings indicate there are incredible opportunities within the
field and it looks like they will not disappear anytime soon. So, we encourage the researchers
that are interested in the area to start exploring.
39
8. Acknowledgement
This work is supported by the Scientific and Technological Research Council of Turkey
(TUBITAK) grant no 215E248.
Glossary
AE Autoencoder. 4, 8, 9, 11, 12, 15, 18–20, 23, 29, 29, 30
30 DFNN Deep Feedforward Neural Network. 21, 22,
AI Artificial Intelligence. 15 24, 29
AMEX American Stock Exchange. 20–22, 31 DGM Deep Neural Generative Model. 27
ANN Artificial Neural Network. 3, 4, 7, 8, 17, 19 DGP Deep Gaussian Process. 9
ARIMA Autoregressive Integrated Moving Aver- DJI Dow Jones Index. 24
age. 23 DJIA Dow Jones Industrial Average. 11, 24, 25
AUC Area Under the Curve. 20, 26, 27, 29 DL Deep Learning. 1–4, 6, 8–11, 14–16, 18–25, 28,
AUROC Area Under the Receiver Operating 31–39
Characteristics. 13, 15, 17–19, 24, 26, 28, DLR Deep Learning Representation. 26
29 DMI Directional Movement Index. 24
BA Balanced Accuracy. 17 DMLP Deep Multilayer Perceptron. 4, 12, 13, 16,
BELM Basic Extreme Learning Machine. 26 19, 20, 22–24, 32, 34, 35, 39
BHC Bank Holding Companies. 17, 27 DNN Deep Neural Network. 5, 11–15, 17–22, 24,
Bi-GRU Bidirectional Gated Recurrent Unit. 26 25, 28–30
Bi-LSTM Bidirectional LSTM. 20, 21, 24, 28 DP Discriminant Power. 17
BIST Istanbul Stock Exchange Index. 11 DQL Deep Q-Learning. 19
BOLL Bollinger Band. 22 DRL Deep Reinforcement Learning. 9, 12, 14, 21,
BPTT Backpropagation Through Time. 6 23, 29, 30, 35
CAE Convolutional Autoencoder. 13 DRSE Deep Random Subspace Ensembles. 24, 28
CAGR Compound Annual Growth Rate. 20 DTW Dynamic Time Warping. 14
CART Classification and Regression Trees. 17 EA Evolutionary Algorithm. 3, 10, 19
CCI Commodity Channel Index. 24 EC Evolutionary Computation. 15, 38
CDAX German Stock Market Index Calculated by ELM Extreme Learning Machine. 26
Deutsche Börse. 26 EMA Exponential Moving Average. 24
CDS Credit Default Swaps. 15, 16 ETF Exchange-Traded Fund. 11, 13
CGAN Conditional GAN. 17 FDDR Fuzzy Deep Direct Reinforcement Learn-
CME Chicago Mercantile Exchange. 14, 30 ing. 11
CNN Convolutional Neural Network. 2, 4, 5, 10– FE-QAR Fixed Effects Quantile VAR. 17
18, 20–32, 34–37, 39 FFNN Feedforward Neural Network. 6, 8, 13, 18,
CRIX The Cryptocurrency Index. 22 27, 30
CRSP Center for Research in Security Prices. 17 FN False Negative. 15, 17
CSI China Securities Index. 11, 25 FNN Fully Connected Neural Network. 12
DAX The Deutscher Aktienindex. 13, 14 FP False Positive. 15, 17, 19, 29
DBN Deep Belief Network. 4, 8, 15–17, 20, 27, 30, FPGA Field Programmable Gate Array. 38
31 FTSE London Financial Times Stock Exchange In-
DCNL Deep Co-investment Network Learning. 14 dex. 11, 13, 14
DCNN Deep Convolutional Neural Network. 15 G-mean Geometric Mean. 17
DDPG Deep Deterministic Policy Gradient. 20 GA Genetic Algorithm. 3, 12
Deep-FASP The Financial Aspect and Sentiment GAN Generative Adversarial Network. 9, 37
Prediction task with Deep neural net- GASVR GA with a SVR. 11, 12
works. 28 GBDT Gradient-Boosted-DecisionTrees. 19
DFFN Deep Feed Forward Network. 16, 17, 21, GBT Gradient Boosted Trees. 11
40
GP Genetic Programming. 3, 13, 15, 16 31
GPU Graphic Processing Unit. 38 OCHL Open,Close,High, Low. 11, 12, 24
GRU Gated-Recurrent Unit. 14, 19–21, 27, 28, 32 OCHLV Open,Close,High, Low, Volume. 11, 13,
HAN Hybrid Attention Network. 26 14, 20–23, 25, 27
HFT High Frequency Trading. 10, 11, 13, 38, 39 PCA Principal Component Analysis. 4, 19, 26
HMM Hidden Markov Model. 24, 29, 30 PCC Pearson’s Correlation Coefficient. 14
HS China Shanghai Shenzhen Stock Index. 26 PLR Piecewise Linear Representation. 11
HSI Hong Kong Hang Seng Index. 11, 30 PNN Probabilistic Neural Network. 13
IBB iShares Nasdaq Biotechnology ETF. 20, 29, PPO Proximal Policy Optimization. 20
30 PSO Particle Swarm Optimization. 3
KELM Kernel Extreme Learning Machine. 26 R2 Squared correlation, Non-linear regression mul-
KOSPI The Korea Composite Stock Price Index. tiple correlation. 20, 21, 24, 28
11, 30 RBM Restricted Boltzmann Machine. 4, 7, 8, 15–
KS Kolmogorov–Smirnov. 17 17, 19, 20, 27, 30, 31
LAR Linear Auto-regression Predictor. 27 RCNN Recurrent CNN. 26
LDA Latent Dirichlet Allocation. 19, 26, 29 ReLU Rectified Linear Unit. 4
LFM Lookahead Factor Models. 31 RF Random Forest. 11, 16–21, 27, 28
LOB Limit Order Book Data. 13 RL Reinforcement Learning. 11, 13, 14, 18, 20–23
LR Logistic Regression. 16 RMSE Root Mean Square Error. 11, 17, 19, 21,
LSTM Long-Short Term Memory. 2, 4, 6, 10–14, 23, 26, 27, 31
16–30, 32, 33, 35, 37, 39 RNN Recurrent Neural Network. 2, 4, 6, 10–14,
MA Moving Average. 22, 24 20–24, 26–33, 39
MACD Moving Average Convergence and Diver- ROA Return on Assets. 17, 27, 28
gence. 12, 24, 27 ROC Price of Change. 24
MAE Mean Absolute Error. 11, 21, 26 RSE Relative Squared Error. 11
MAPE Mean Absolute Percentage Error. 11, 21, RSI Relative Strength Index. 12, 13, 24
26, 27 S&P500 Standard’s & Poor’s 500 Index. 11, 13,
MCC Matthew Correlation Coefficient. 26–28 14, 20, 24–27, 30
MDA Multilinear Discriminant Analysis. 11 SAE Stacked Autoencoder. 10, 16, 17
MDD Maximum Drawdown. 11, 13, 20, 22 SCI SSE Composite Index. 24
ML Machine Learning. 1–4, 10, 15, 18–21, 23, 25, SFM State Frequency Memory. 11
35, 36, 38, 39 SGD Stochastic Gradient Descent. 4
MLP Multilayer Perceptron. 4, 11, 12, 14–16, 18, SPY SPDR S&P 500 ETF. 13
20–23, 28, 31, 34 SR Sharpe-ratio. 11, 13, 14, 20, 22, 26, 31
MODRL Multi-objective Deep Reinforcement STD Standard Deviation. 11, 20
Learning. 11 SVD Singular Value Decomposition. 29, 30
MOEA Multiobjective Evolutionary Algorithm. 3, SVM Support Vector Machine. 4, 15–17, 24, 27–30
19 SVR Support Vector Regressor. 12, 27, 28, 30
MSE Mean Squared Error. 11, 13, 19, 20, 23, 24, SZSE Shenzhen Stock Exchange Composite Index.
26, 28, 31 11
MV-t Multivariate t Distribution. 17 TAIEX Taiwan Capitalization Weighted Stock In-
MVN Multivariate Normal Distribution. 17 dex. 21
NASDAQ National Association of Securities Deal- TALIB Technical Analysis Library Package. 13,
ers Automated Quotations. 14, 20–22, 25, 26
31 TAQ Trade and Quote. 20
NES Natural Evolution Strategies. 11 TDNN Timedelay Neural Network. 13
NIFTY National Stock Exchange of India. 26 TEMA Triple Exponential Moving Average. 24
NIKKEI Tokyo Nikkei Index. 11 TF-IDF Term Frequency-Inverse Document Fre-
NLP Natural Language Processing. 17, 29, 37–39 quency. 26
NN Neural Network. 3, 17, 24, 26, 29, 30 TGRU Two-stream GRU. 25, 26
NYSE New York Stock Exchange. 11, 20–22, 25, THEIL-U Theil’s inequality coefficient. 11
41
TN True Negative. 15 WBA Weighted Balanced Accuracy. 17
TP True Positive. 15, 19, 29 WMTR Weighted Multichannel Time-series Re-
TR Total Return. 13 gression. 11
TSE Tokyo Stock Exchange. 20, 21, 24
WSURT Wilcoxon Sum-rank Test. 28
TWSE Taiwan Stock Exchange. 26
VAR Vector Auto Regression. 17 WT Wavelet Transforms. 10, 11
VWL WL Kernel-based Method. 14 XGBoost eXtreme Gradient Boosting. 16, 17
42
References
[1] Omer Berat Sezer, Mehmet Ugur Gudelek, and Ahmet Murat Ozbayoglu. Financial time series fore-
casting with deep learning : A systematic literature review: 2005-2019, 2019.
[2] Arash Bahrammirzaee. A comparative survey of artificial intelligence applications in finance: artificial
neural networks, expert system and hybrid intelligent systems. Neural Computing and Applications,
19(8):1165–1195, June 2010.
[3] D. Zhang and L. Zhou. Discovering golden nuggets: Data mining in financial application. IEEE
Transactions on Systems, Man and Cybernetics, Part C (Applications and Reviews), 34(4):513–522,
November 2004.
[4] Asunción Mochón, David Quintana, Yago Sáez, and Pedro Isasi Viñuela. Soft computing techniques
applied to finance. Applied Intelligence, 29:111–115, 2007.
[5] Pulakkazhy. Mining in banking and its applications: A review. Journal of Computer Science, 9(10):
1252–1259, October 2013.
[6] Sendhil Mullainathan and Jann Spiess. Machine learning: An applied econometric approach. Journal
of Economic Perspectives, 31(2):87–106, May 2017.
[7] Keke Gai, Meikang Qiu, and Xiaotong Sun. A survey on fintech. Journal of Network and Computer
Applications, 103:262–273, 2018.
[8] Boris Kovalerchuk and Evgenii Vityaev. Data Mining in Finance: Advances in Relational and Hybrid
Methods. Kluwer Academic Publishers, Norwell, MA, USA, 2000.
[9] Rafik A. Aliev, Bijan Fazlollahi, and Rashad R. Aliev. Soft computing and its applications in business
and economics. In Studies in Fuzziness and Soft Computing, 2004.
[10] Anthony Brabazon and Michael O’Neill, editors. Natural Computing in Computational Finance.
Springer Berlin Heidelberg, 2008.
[11] Ludmila Dymowa. Soft Computing in Economics and Finance. Springer Berlin Heidelberg, 2011.
[12] Shu-Heng Chen, editor. Genetic Algorithms and Genetic Programming in Computational Finance.
Springer US, 2002.
[13] Ma. Guadalupe Castillo Tapia and Carlos A. Coello Coello. Applications of multi-objective evolu-
tionary algorithms in economics and finance: A survey. In 2007 IEEE Congress on Evolutionary
Computation. IEEE, September 2007.
[14] Antonin Ponsich, Antonio Lopez Jaimes, and Carlos A. Coello Coello. A survey on multiobjective
evolutionary algorithms for the solution of the portfolio optimization problem and other finance and
economics applications. IEEE Transactions on Evolutionary Computation, 17(3):321–344, June 2013.
[15] Ruben Aguilar-Rivera, Manuel Valenzuela-Rendon, and J.J. Rodriguez-Ortiz. Genetic algorithms and
darwinian approaches in financial applications: A survey. Expert Systems with Applications, 42(21):
7684–7697, November 2015.
[16] Bo K Wong and Yakup Selvi. Neural network applications in finance: A review and analysis of
literature (1990–1996). Information & Management, 34(3):129–139, October 1998.
[17] Yuhong Li and Weihua Ma. Applications of artificial neural networks in financial economics: A survey.
In 2010 International Symposium on Computational Intelligence and Design. IEEE, October 2010.
[18] B. Elmsili and B. Outtaj. Artificial neural networks applications in economics and management
research: An exploratory literature review. In 2018 4th International Conference on Optimization
and Applications (ICOA), pages 1–6, April 2018.
[19] Blake LeBaron. Chapter 24 agent-based computational finance. In L. Tesfatsion and K.L. Judd,
editors, Handbook of Computational Economics, volume 2 of Handbook of Computational Economics,
pages 1187–1233. Elsevier, 2006.
[20] Stephan K. Chalup and Andreas Mitschele. Kernel methods in finance. In Handbook on Information
Technology in Finance, pages 655–687. Springer Berlin Heidelberg, 2008.
[21] Yann LeCun, Yoshua Bengio, and Geoffrey Hinton. Deep learning. Nature, 521(7553):436–444, 2015.
[22] George Cybenko. Approximation by superpositions of a sigmoidal function. Mathematics of control,
signals and systems, 2(4):303–314, 1989.
43
[23] Barry L Kalman and Stan C Kwasny. Why tanh: choosing a sigmoidal function. In [Proceedings 1992]
IJCNN International Joint Conference on Neural Networks, volume 4, pages 578–581. IEEE, 1992.
[24] Vinod Nair and Geoffrey E Hinton. Rectified linear units improve restricted boltzmann machines.
In Proceedings of the 27th international conference on machine learning (ICML-10), pages 807–814,
2010.
[25] Andrew L Maas, Awni Y Hannun, and Andrew Y Ng. Rectifier nonlinearities improve neural network
acoustic models. In Proc. icml, volume 30, page 3, 2013.
[26] Prajit Ramachandran, Barret Zoph, and Quoc V Le. Searching for activation functions. arXiv preprint
arXiv:1710.05941, 2017.
[27] Ian Goodfellow, Yoshua Bengio, and Aaron Courville. Deep Learning. MIT Press, 2016.
http://www.deeplearningbook.org.
[28] Sepp Hochreiter and Jürgen Schmidhuber. Long short-term memory. Neural computation, 9(8):1735–
1780, 1997.
[29] Xueheng Qiu, Le Zhang, Ye Ren, P. Suganthan, and Gehan Amaratunga. Ensemble deep learning
for regression and time series forecasting. In 2014 IEEE Symposium on Computational Intelligence in
Ensemble Learning (CIEL), pages 1–6, 2014.
[30] Yoshua Bengio. Deep learning of representations for unsupervised and transfer learning. In Proceedings
of ICML workshop on unsupervised and transfer learning, pages 17–36, 2012.
[31] Geoffrey E. Hinton, Simon Osindero, and Yee-Whye Teh. A fast learning algorithm for deep belief
nets. Neural Computation, 18(7):1527–1554, 2006.
[32] Pascal Vincent, Hugo Larochelle, Yoshua Bengio, and Pierre-Antoine Manzagol. Extracting and
composing robust features with denoising autoencoders. In Proceedings of the 25th international
conference on Machine learning, pages 1096–1103. ACM, 2008.
[33] Qinxue Meng, Daniel Catchpoole, David Skillicom, and Paul J Kennedy. Relational autoencoder
for feature extraction. In 2017 International Joint Conference on Neural Networks (IJCNN), pages
364–371. IEEE, 2017.
[34] Yong Hu, Kang Liu, Xiangzhou Zhang, Lijun Su, E.W.T. Ngai, and Mei Liu. Application of evolu-
tionary computation for rule discovery in stock algorithmic trading: A literature review. Applied Soft
Computing, 36:534–551, November 2015.
[35] Sercan Karaoglu and Ugur Arpaci. A deep learning approach for optimization of systematic signal
detection in financial trading systems with big data. International Journal of Intelligent Systems and
Applications in Engineering, SpecialIssue(SpecialIssue):31–36, July 2017.
[36] Wei Bao, Jun Yue, and Yulei Rao. A deep learning framework for financial time series using stacked
autoencoders and long-short term memory. PLOS ONE, 12(7):e0180944, July 2017.
[37] Shuanglong Liu, Chao Zhang, and Jinwen Ma. Cnn-lstm neural network model for quantitative strat-
egy analysis in stock markets. In Neural Information Processing, pages 198–206. Springer International
Publishing, 2017.
[38] Liheng Zhang, Charu Aggarwal, and Guo-Jun Qi. Stock price prediction via discovering multi-
frequency trading patterns. In Proceedings of the 23rd ACM SIGKDD International Conference on
Knowledge Discovery and Data Mining - KDD17. ACM Press, 2017.
[39] Dat Thanh Tran, Martin Magris, Juho Kanniainen, Moncef Gabbouj, and Alexandros Iosifidis. Tensor
representation in high-frequency financial data for price change prediction. In 2017 IEEE Symposium
Series on Computational Intelligence (SSCI). IEEE, November 2017.
[40] Yue Deng, Feng Bao, Youyong Kong, Zhiquan Ren, and Qionghai Dai. Deep direct reinforcement
learning for financial signal representation and trading. IEEE Transactions on Neural Networks and
Learning Systems, 28(3):653–664, March 2017.
[41] Thomas Fischer and Christopher Krauss. Deep learning with long short-term memory networks for
financial market predictions. European Journal of Operational Research, 270(2):654–669, October
2018.
[42] Marios Mourelatos, Christos Alexakos, Thomas Amorgianiotis, and Spiridon Likothanassis. Financial
indices modelling and trading utilizing deep learning techniques: The athens se ftse/ase large cap use
44
case. In 2018 Innovations in Intelligent Systems and Applications (INISTA). IEEE, July 2018.
[43] Weiyu Si, Jinke Li, Peng Ding, and Ruonan Rao. A multi-objective deep reinforcement learning
approach for stock index future’s intraday trading. In 2017 10th International Symposium on Com-
putational Intelligence and Design (ISCID). IEEE, December 2017.
[44] Bang Xiang Yong, Mohd Rozaini Abdul Rahim, and Ahmad Shahidan Abdullah. A stock market
trading system using deep neural network. In Communications in Computer and Information Science,
pages 356–364. Springer Singapore, 2017.
[45] David W. Lu. Agent inspired trading using recurrent reinforcement learning and lstm neural networks,
2017.
[46] Matthew Francis Dixon, Diego Klabjan, and Jin Hoon Bang. Classification-based financial markets
prediction using deep neural networks. SSRN Electronic Journal, 2016.
[47] Jerzy Korczak and Marcin Hernes. Deep learning for financial time series forecasting in a-trader
system. In Proceedings of the 2017 Federated Conference on Computer Science and Information
Systems. IEEE, September 2017.
[48] Bruno Spilak. Deep neural networks for cryptocurrencies price prediction. Master’s thesis, Humboldt-
Universitat zu Berlin, Wirtschaftswissenschaftliche Fakultat, 2018.
[49] Gyeeun Jeong and Ha Young Kim. Improving financial trading decisions using deep q-learning: Pre-
dicting the number of shares, action strategies, and transfer learning. Expert Systems with Applications,
117:125–138, March 2019.
[50] Christopher Krauss, Xuan Anh Do, and Nicolas Huck. Deep neural networks, gradient-boosted trees,
random forests: Statistical arbitrage on the s&p 500. European Journal of Operational Research, 259
(2):689–702, June 2017.
[51] Google. System and method for computer managed funds to outperform benchmarks.
[52] Omer Berat Sezer, Murat Ozbayoglu, and Erdogan Dogdu. A deep neural-network based stock trading
system based on evolutionary optimized technical analysis parameters. Procedia Computer Science,
114:473–480, 2017.
[53] Ariel Navon and Yosi Keller. Financial time series prediction using deep learning, 2017.
[54] Luigi Troiano, Elena Mejuto Villa, and Vincenzo Loia. Replicating a trading strategy by means of
lstm for financial industry applications. IEEE Transactions on Industrial Informatics, 14(7):3226–
3234, July 2018.
[55] Justin Sirignano and Rama Cont. Universal features of price formation in financial markets: Perspec-
tives from deep learning. SSRN Electronic Journal, 2018.
[56] Avraam Tsantekidis, Nikolaos Passalis, Anastasios Tefas, Juho Kanniainen, Moncef Gabbouj, and
Alexandros Iosifidis. Using deep learning to detect price change indications in financial markets. In
2017 25th European Signal Processing Conference (EUSIPCO). IEEE, August 2017.
[57] M. Ugur Gudelek, S. Arda Boluk, and A. Murat Ozbayoglu. A deep learning based stock trading
model with 2-d cnn trend detection. In 2017 IEEE Symposium Series on Computational Intelligence
(SSCI). IEEE, November 2017.
[58] Omer Berat Sezer and Ahmet Murat Ozbayoglu. Algorithmic financial trading with deep convolutional
neural networks: Time series to image conversion approach. Applied Soft Computing, 70:525–538,
September 2018.
[59] Guosheng Hu, Yuxin Hu, Kai Yang, Zehao Yu, Flood Sung, Zhihong Zhang, Fei Xie, Jianguo Liu, Neil
Robertson, Timpathy Hospedales, and Qiangwei Miemie. Deep stock representation learning: From
candlestick charts to investment decisions. In 2018 IEEE International Conference on Acoustics,
Speech and Signal Processing (ICASSP). IEEE, April 2018.
[60] Avraam Tsantekidis, Nikolaos Passalis, Anastasios Tefas, Juho Kanniainen, Moncef Gabbouj, and
Alexandros Iosifidis. Forecasting stock prices from the limit order book using convolutional neural
networks. In 2017 IEEE 19th Conference on Business Informatics (CBI). IEEE, July 2017.
[61] Hakan Gunduz, Yusuf Yaslan, and Zehra Cataltepe. Intraday prediction of borsa istanbul using convo-
lutional neural networks and feature correlations. Knowledge-Based Systems, 137:138–148, December
2017.
45
[62] Omer Berat Sezer and Ahmet Murat Ozbayoglu. Financial trading model with stock bar chart image
time series with deep convolutional neural networks. arXiv preprint arXiv:1903.04610, 2019.
[63] Will Serrano. Fintech model: The random neural network with genetic algorithm. Procedia Computer
Science, 126:537–546, 2018.
[64] E.W. Saad, D.V. Prokhorov, and D.C. Wunsch. Comparative study of stock trend prediction using
time delay, recurrent and probabilistic neural networks. IEEE Transactions on Neural Networks, 9
(6):1456–1470, 1998.
[65] Jonathan Doering, Michael Fairbank, and Sheri Markose. Convolutional neural networks applied
to high-frequency market microstructure forecasting. In 2017 9th Computer Science and Electronic
Engineering (CEEC). IEEE, September 2017.
[66] Zhengyao Jiang, Dixing Xu, and Jinjun Liang. A deep reinforcement learning framework for the
financial portfolio management problem. arXiv preprint arXiv:1706.10059, 2017.
[67] P. Tino, C. Schittenkopf, and G. Dorffner. Financial volatility trading using recurrent neural networks.
IEEE Transactions on Neural Networks, 12(4):865–874, July 2001.
[68] Yu-Ying Chen, Wei-Lun Chen, and Szu-Hao Huang. Developing arbitrage strategy in high-frequency
pairs trading with filterbank cnn algorithm. In 2018 IEEE International Conference on Agents (ICA).
IEEE, July 2018.
[69] Omar A. Bari and Arvin Agah. Ensembles of text and time-series models for automatic generation of
financial trading signals from social media content. Journal of Intelligent Systems, 2018.
[70] Matthew Francis Dixon. Sequence classification of the limit order book using recurrent neural net-
works. SSRN Electronic Journal, 2017.
[71] Chiao-Ting Chen, An-Pin Chen, and Szu-Hao Huang. Cloning strategies from trading records using
agent-based reinforcement learning algorithm. In 2018 IEEE International Conference on Agents
(ICA). IEEE, July 2018.
[72] Yue Wang, Chenwei Zhang, Shen Wang, Philip S. Yu, Lu Bai, and Lixin Cui. Deep co-investment
network learning for financial assets, 2018.
[73] Min-Yuh Day and Chia-Chou Lee. Deep learning for financial sentiment analysis on finance news
providers. In 2016 IEEE/ACM International Conference on Advances in Social Networks Analysis
and Mining (ASONAM). IEEE, August 2016.
[74] Justin Sirignano. Deep learning for limit order books, 2016.
[75] Xiang Gao. Deep reinforcement learning for time series: playing idealized trading games, 2018.
[76] Martin Neil Baily, Robert E. Litan, and Johnson Matthew S. The origins of the financial crisis.
Initiative on Business and Public Policy at Brookings, Fixing Finance Series - Paper 3, 2008.
[77] Cuicui Luo, Desheng Wu, and Dexiang Wu. A deep learning approach for credit scoring using credit
default swaps. Engineering Applications of Artificial Intelligence, 65:465–470, October 2017.
[78] Lean Yu, Rongtian Zhou, Ling Tang, and Rongda Chen. A dbn-based resampling svm ensemble
learning paradigm for credit classification with imbalanced data. Applied Soft Computing, 69:192–202,
August 2018.
[79] Ying Li, Xianghong Lin, Xiangwen Wang, Fanqi Shen, and Zuzheng Gong. Credit risk assessment algo-
rithm using deep neural networks with clustering and merging. In 2017 13th International Conference
on Computational Intelligence and Security (CIS). IEEE, December 2017.
[80] Khiem Tran, Thanh Duong, and Quyen Ho. Credit scoring model: A combination of genetic program-
ming and deep learning. In 2016 Future Technologies Conference (FTC). IEEE, December 2016.
[81] Victor-Emil Neagoe, Adrian-Dumitru Ciotec, and George-Sorin Cucu. Deep convolutional neural
networks versus multilayer perceptron for financial prediction. In 2018 International Conference on
Communications (COMM). IEEE, June 2018.
[82] Bing Zhu, Wenchuan Yang, Huaxuan Wang, and Yuan Yuan. A hybrid deep learning model for
consumer credit scoring. In 2018 International Conference on Artificial Intelligence and Big Data
(ICAIBD). IEEE, May 2018.
[83] Ayahiko Niimi. Deep learning for credit card data analysis. In 2015 World Congress on Internet
Security (WorldCIS). IEEE, October 2015.
46
[84] Efstathios Kirkos and Yannis Manolopoulos. Data mining in finance and accounting: A review of
current research trends. In Proceedings of the 1 st International Conference on Enterprise Systems
and Accounting (ICESAcc, pages 63–78, 2004.
[85] V. Ravi, H. Kurniawan, Peter Nwee Kok Thai, and P. Ravi Kumar. Soft computing system for bank
performance prediction. Applied Soft Computing, 8(1):305–315, January 2008.
[86] Meryem Duygun Fethi and Fotios Pasiouras. Assessing bank efficiency and performance with op-
erational research and artificial intelligence techniques: A survey. European Journal of Operational
Research, 204(2):189–198, July 2010.
[87] Adel Lahsasna, Raja Noor Ainon, and Ying Wah Teh. Credit scoring models using soft computing
methods: A survey. Int. Arab J. Inf. Technol., 7:115–123, 2010.
[88] Ning Chen, Bernardete Ribeiro, and An Chen. Financial credit risk assessment: a recent review.
Artificial Intelligence Review, 45(1):1–23, October 2015.
[89] AI Marques, Vicente García, and José Salvador Sánchez. A literature review on the application of
evolutionary computing to credit scoring. Journal of the Operational Research Society, 64(9):1384–
1399, 2013.
[90] P. Ravi Kumar and V. Ravi. Bankruptcy prediction in banks and firms via statistical and intelligent
techniques – a review. European Journal of Operational Research, 180(1):1–28, July 2007.
[91] Antanas Verikas, Zivile Kalsyte, Marija Bacauskiene, and Adas Gelzinis. Hybrid and ensemble-based
soft computing techniques in bankruptcy prediction: a survey. Soft Computing, 14(9):995–1010,
September 2009.
[92] Jie Sun, Hui Li, Qing-Hua Huang, and Kai-Yu He. Predicting financial distress and corporate failure: A
review from the state-of-the-art definitions, modeling, sampling, and featuring approaches. Knowledge-
Based Systems, 57:41–56, February 2014.
[93] W. Lin, Y. Hu, and C. Tsai. Machine learning in financial crisis prediction: A survey. IEEE Trans-
actions on Systems, Man, and Cybernetics, Part C (Applications and Reviews), 42(4):421–436, July
2012.
[94] Zineb Lanbouri and Said Achchab. A hybrid deep belief network approach for financial distress
prediction. In 2015 10th International Conference on Intelligent Systems: Theories and Applications
(SITA). IEEE, October 2015.
[95] Vipula Rawte, Aparna Gupta, and Mohammed J. Zaki. Analysis of year-over-year changes in risk
factors disclosure in 10-k filings. In Proceedings of the Fourth International Workshop on Data Science
for Macro-Modeling with Financial and Economic Datasets - DSMM18. ACM Press, 2018.
[96] Samuel Ronnqvist and Peter Sarlin. Detect & describe: Deep learning of bank stress in the news. In
2015 IEEE Symposium Series on Computational Intelligence. IEEE, December 2015.
[97] Samuel Ronnqvist and Peter Sarlin. Bank distress in the news: Describing events through deep
learning. Neurocomputing, 264:57–70, November 2017.
[98] Paola Cerchiello, Giancarlo Nicola, Samuel Rönnqvist, and Peter Sarlin. Deep learning bank distress
from news and numerical financial data. CoRR, abs/1706.09627, 2017.
[99] Nikhil Malik, Param Vir Singh, and Urooj Khan. Can banks survive the next financial crisis? an
adversarial deep learning model for bank stress testing. An Adversarial Deep Learning Model for Bank
Stress Testing (June 30, 2018), 2018.
[100] Bernardete Ribeiro and Noel Lopes. Deep belief networks for financial prediction. In Neural Informa-
tion Processing, pages 766–773. Springer Berlin Heidelberg, 2011.
[101] Shu-Hao Yeh, Chuan-Ju Wang, and Ming-Feng Tsai. Deep belief networks for predicting corporate
defaults. In 2015 24th Wireless and Optical Communication Conference (WOCC). IEEE, October
2015.
[102] Tadaaki Hosaka. Bankruptcy prediction using imaged financial ratios and convolutional neural net-
works. Expert Systems with Applications, September 2018.
[103] Justin Sirignano, Apaar Sadhwani, and Kay Giesecke. Deep learning for mortgage risk. SSRN Elec-
tronic Journal, 2018.
[104] Håvard Kvamme, Nikolai Sellereite, Kjersti Aas, and Steffen Sjursen. Predicting mortgage default
47
using convolutional neural networks. Expert Systems with Applications, 102:207–217, July 2018.
[105] Narek Abroyan and. Neural networks for financial market risk classification. Frontiers in Signal
Processing, 1(2), August 2017.
[106] Sotirios P. Chatzis, Vassilis Siakoulis, Anastasios Petropoulos, Evangelos Stavroulakis, and Nikos
Vlachogiannakis. Forecasting stock market crisis events using deep and statistical machine learning
techniques. Expert Systems with Applications, 112:353–371, December 2018.
[107] E Kirkos, C Spathis, and Y Manolopoulos. Data mining techniques for the detection of fraudulent
financial statements. Expert Systems with Applications, 32(4):995–1003, May 2007.
[108] Dianmin Yue, Xiaodan Wu, Yunfeng Wang, Yue Li, and Chao-Hsien Chu. A review of data mining-
based financial fraud detection research. In 2007 International Conference on Wireless Communica-
tions, Networking and Mobile Computing. IEEE, September 2007.
[109] Shiguo Wang. A comprehensive survey of data mining-based accounting-fraud detection research. In
2010 International Conference on Intelligent Computation Technology and Automation. IEEE, May
2010.
[110] Clifton Phua, Vincent C. S. Lee, Kate Smith-Miles, and Ross W. Gayler. A comprehensive survey of
data mining-based fraud detection research. CoRR, abs/1009.6119, 2010.
[111] E.W.T. Ngai, Yong Hu, Y.H. Wong, Yijun Chen, and Xin Sun. The application of data mining tech-
niques in financial fraud detection: A classification framework and an academic review of literature.
Decision Support Systems, 50(3):559–569, February 2011.
[112] Anuj Sharma and Prabin Kumar Panigrahi. A review of financial accounting fraud detection based
on data mining techniques. International Journal of Computer Applications, 39(1):37–47, February
2012.
[113] Jarrod West and Maumita Bhattacharya. Intelligent financial fraud detection: A comprehensive
review. Computers & Security, 57:47–66, March 2016.
[114] Yaya Heryadi and Harco Leslie Hendric Spits Warnars. Learning temporal representation of transaction
amount for fraudulent transaction recognition using cnn, stacked lstm, and cnn-lstm. In 2017 IEEE
International Conference on Cybernetics and Computational Intelligence (CyberneticsCom). IEEE,
November 2017.
[115] Abhimanyu Roy, Jingyi Sun, Robert Mahoney, Loreto Alonzi, Stephen Adams, and Peter Beling. Deep
learning detecting fraud in credit card transactions. In 2018 Systems and Information Engineering
Design Symposium (SIEDS). IEEE, April 2018.
[116] Jon Ander Gómez, Juan Arévalo, Roberto Paredes, and Jordi Nin. End-to-end neural network archi-
tecture for fraud scoring in card payments. Pattern Recognition Letters, 105:175–181, April 2018.
[117] Ishan Sohony, Rameshwar Pratap, and Ullas Nambiar. Ensemble learning for credit card fraud detec-
tion. In Proceedings of the ACM India Joint International Conference on Data Science and Manage-
ment of Data - CoDS-COMAD18. ACM Press, 2018.
[118] Johannes Jurgovsky, Michael Granitzer, Konstantin Ziegler, Sylvie Calabretto, Pierre-Edouard
Portier, Liyun He-Guelton, and Olivier Caelen. Sequence classification for credit-card fraud detection.
Expert Systems with Applications, 100:234–245, June 2018.
[119] Ebberth L. Paula, Marcelo Ladeira, Rommel N. Carvalho, and Thiago Marzagao. Deep learning
anomaly detection as support fraud investigation in brazilian exports and anti-money laundering. In
2016 15th IEEE International Conference on Machine Learning and Applications (ICMLA). IEEE,
December 2016.
[120] Thiago Alencar Gomes, Rommel Novaes Carvalho, and Ricardo Silva Carvalho. Identifying anomalies
in parliamentary expenditures of brazilian chamber of deputies with deep autoencoders. In 2017 16th
IEEE International Conference on Machine Learning and Applications (ICMLA). IEEE, December
2017.
[121] Yibo Wang and Wei Xu. Leveraging deep learning with lda-based text analytics to detect automobile
insurance fraud. Decision Support Systems, 105:87–95, January 2018.
[122] Longfei Li, Jun Zhou, Xiaolong Li, and Tao Chen. Poster: Practical fraud transaction prediction. In
ACM Conference on Computer and Communications Security, 2017.
48
[123] Allan Inocencio de Souza Costa and Luis Silva. Sequence classification of the limit order book using
recurrent neural networks. 2016.
[124] Nikolaos D. Goumagias, Dimitrios Hristu-Varsakelis, and Yannis M. Assael. Using deep q-learning
to understand the tax evasion behavior of risk-averse firms. Expert Systems with Applications, 101:
258–270, July 2018.
[125] Bin Li and Steven C. H. Hoi. Online portfolio selection: A survey. ACM Comput. Surv., 46(3):
35:1–35:36, January 2014.
[126] K. Metaxiotis and K. Liagkouras. Multiobjective evolutionary algorithms for portfolio management:
A comprehensive literature review. Expert Systems with Applications, 39(14):11685–11698, October
2012.
[127] Lawrence Takeuchi. Applying deep learning to enhance momentum trading strategies in stocks. 2013.
[128] Anthony Grace. Can deep learning techniques improve the risk adjusted returns from enhanced
indexing investment strategies. Master’s thesis, 2017.
[129] XingYu Fu, JinHong Du, YiFeng Guo, MingWen Liu, Tao Dong, and XiuWen Duan. A machine
learning framework for stock selection, 2018.
[130] Saurabh Aggarwal and Somya Aggarwal. Deep investment in financial markets using deep learning
models. International Journal of Computer Applications, 162(2):40–43, March 2017.
[131] J.B. Heaton and Nick Polson. Deep learning for finance: Deep portfolios. SSRN Electronic Journal,
2016.
[132] Chi-Ming Lin, Jih-Jeng Huang, Mitsuo Gen, and Gwo-Hshiung Tzeng. Recurrent neural network for
dynamic portfolio selection. Applied Mathematics and Computation, 175(2):1139–1146, April 2006.
[133] Nijolė Maknickienė. Selection of orthogonal investment portfolio using evolino rnn trading model.
Procedia - Social and Behavioral Sciences, 110:1158–1165, January 2014.
[134] Bo Zhou. Deep learning and the cross-section of stock returns: Neural networks combining price and
fundamental information. SSRN Electronic Journal, 2018.
[135] Bilberto Batres-Estrada. Deep learning for multivariate financial time series. Master’s thesis, KTH,
Mathematical Statistics, 2015.
[136] Sang Il Lee and Seong Joon Yoo. Threshold-based portfolio: the role of the threshold and its appli-
cations. The Journal of Supercomputing, September 2018.
[137] Hitoshi Iwasaki and Ying Chen. Topic sentiment asset pricing with dnn supervised learning. SSRN
Electronic Journal, 2018.
[138] Zhipeng Liang, Hao Chen, Junhao Zhu, Kangkang Jiang, and Yanran Li. Adversarial deep reinforce-
ment learning in portfolio management, 2018.
[139] Jiaqi Chen, Wenbo Wu, and Michael Tindall. Hedge fund return prediction and fund selection: A
machine-learning approach. Occasional Papers 16-4, Federal Reserve Bank of Dallas, November 2016.
[140] Zhengyao Jiang and Jinjun Liang. Cryptocurrency portfolio management with deep reinforcement
learning. In 2017 Intelligent Systems Conference (IntelliSys). IEEE, September 2017.
[141] Nijole Maknickiene, Aleksandras Vytautas Rutkauskas, and Algirdas Maknickas. Investigation of
financial market prediction by recurrent neural network. 2014.
[142] Robert Culkin and Sanjiv R. Das. Machine learning in finance: The case of deep learning in option
pricing. 2017.
[143] Pei-Ying Hsu, Chin Chou, Szu-Hao Huang, and An-Pin Chen. A market making quotation strategy
based on dual deep learning agents for option pricing and bid-ask spread estimation. In 2018 IEEE
International Conference on Agents (ICA). IEEE, July 2018.
[144] Guanhao Feng, Nicholas G. Polson, and Jianeng Xu. Deep factor alpha, 2018.
[145] Rui-Yang Chen. A traceability chain algorithm for artificial neural networks using t–s fuzzy cognitive
maps in blockchain. Future Generation Computer Systems, 80:198–210, March 2018.
[146] Lihao Nan and Dacheng Tao. Bitcoin mixing detection using deep autoencoder. In 2018 IEEE Third
International Conference on Data Science in Cyberspace (DSC). IEEE, June 2018.
[147] Gonçalo Duarte Lima Freire Lopes. Deep learning for market forecasts. 2018.
[148] Sean McNally, Jason Roche, and Simon Caton. Predicting the price of bitcoin using machine learn-
49
ing. In 2018 26th Euromicro International Conference on Parallel, Distributed and Network-based
Processing (PDP). IEEE, March 2018.
[149] Colm Kearney and Sha Liu. Textual sentiment in finance: A survey of methods and models. Interna-
tional Review of Financial Analysis, 33:171–185, May 2014.
[150] Qili Wang, Wei Xu, and Han Zheng. Combining the wisdom of crowds and technical analysis for
financial market prediction using deep random subspace ensembles. Neurocomputing, 299:51–61, July
2018.
[151] Lei Shi, Zhiyang Teng, Le Wang, Yue Zhang, and Alexander Binder. Deepclue: Visual interpretation
of text-based deep stock prediction. IEEE Transactions on Knowledge and Data Engineering, pages
1–1, 2018.
[152] Yangtuo Peng and Hui Jiang. Leverage financial news to predict stock price movements using word
embeddings and deep neural networks. In Proceedings of the 2016 Conference of the North American
Chapter of the Association for Computational Linguistics: Human Language Technologies. Association
for Computational Linguistics, 2016.
[153] Qun Zhuge, Lingyu Xu, and Gaowei Zhang. Lstm neural network with emotional analysis for prediction
of stock price. 2017.
[154] Zhongshengz. Measuring financial crisis index for risk warning through analysis of social network.
Master’s thesis, 2018.
[155] Sushree Das, Ranjan Kumar Behera, Mukesh Kumar, and Santanu Kumar Rath. Real-time sentiment
analysis of twitter streaming data for stock prediction. Procedia Computer Science, 132:956–964, 2018.
[156] Jordan Prosky, Xingyou Song, Andrew Tan, and Michael Zhao. Sentiment predictability for stocks.
CoRR, abs/1712.05785, 2017.
[157] Jiahong Li, Hui Bu, and Junjie Wu. Sentiment-aware stock market prediction: A deep learning
method. In 2017 International Conference on Service Systems and Service Management. IEEE, June
2017.
[158] Yifu Huang, Kai Huang, Yang Wang, Hao Zhang, Jihong Guan, and Shuigeng Zhou. Exploiting twitter
moods to boost financial trend prediction based on deep network models. In Intelligent Computing
Methodologies, pages 449–460. Springer International Publishing, 2016.
[159] Leela Mitra and Gautam Mitra. Applications of news analytics in finance: A review. In The Handbook
of News Analytics in Finance, pages 1–39. John Wiley & Sons, Ltd., May 2012.
[160] Feng Li. Textual analysis of corporate disclosures: A survey of the literature. Journal of Accounting
Literature, 29, February 2011.
[161] Tim Loughran and Bill McDonald. Textual analysis in accounting and finance: A survey. Journal of
Accounting Research, 54(4):1187–1230, June 2016.
[162] B. Shravan Kumar and Vadlamani Ravi. A survey of the applications of text mining in financial
domain. Knowledge-Based Systems, 114:128–147, December 2016.
[163] Marc-André Mittermayer and Gerhard F Knolmayer. Text mining systems for market response to
news: A survey. September 2006.
[164] Arman Khadjeh Nassirtoussi, Saeed Aghabozorgi, Teh Ying Wah, and David Chek Ling Ngo. Text
mining for market prediction: A systematic review. Expert Systems with Applications, 41(16):7653–
7670, November 2014.
[165] Huy D. Huynh, L. Minh Dang, and Duc Duong. A new model for stock price movements prediction
using deep neural network. In Proceedings of the Eighth International Symposium on Information and
Communication Technology - SoICT 2017. ACM Press, 2017.
[166] Songqiao Han, Xiaoling Hao, and Hailiang Huang. An event-extraction approach for business analysis
from online chinese news. Electronic Commerce Research and Applications, 28:244–260, March 2018.
[167] Mathias Kraus and Stefan Feuerriegel. Decision support from financial disclosures with deep neural
networks and transfer learning. Decision Support Systems, 104:38–48, December 2017.
[168] L. Minh Dang, Abolghasem Sadeghi-Niaraki, Huy D. Huynh, Kyungbok Min, and Hyeonjoon Moon.
Deep learning approach for short-term stock trends prediction based on two-stream gated recurrent
unit network. IEEE Access, pages 1–1, 2018.
50
[169] Xiao Ding, Yue Zhang, Ting Liu, and Junwen Duan. Deep learning for event-driven stock prediction.
In Proceedings of the 24th International Conference on Artificial Intelligence, IJCAI’15, pages 2327–
2333. AAAI Press, 2015.
[170] Manuel R. Vargas, Beatriz S. L. P. de Lima, and Alexandre G. Evsukoff. Deep learning for stock market
prediction from financial news articles. In 2017 IEEE International Conference on Computational
Intelligence and Virtual Environments for Measurement Systems and Applications (CIVEMSA). IEEE,
June 2017.
[171] Ryo Akita, Akira Yoshihara, Takashi Matsubara, and Kuniaki Uehara. Deep learning for stock pre-
diction using numerical and textual information. In 2016 IEEE/ACIS 15th International Conference
on Computer and Information Science (ICIS). IEEE, June 2016.
[172] Ishan Verma, Lipika Dey, and Hardik Meisheri. Detecting, quantifying and accessing impact of news
events on indian stock indices. In Proceedings of the International Conference on Web Intelligence -
WI17. ACM Press, 2017.
[173] Xi Zhang, Yunjia Zhang, Senzhang Wang, Yuntao Yao, Binxing Fang, and Philip S. Yu. Improving
stock market prediction via heterogeneous information fusion. Knowledge-Based Systems, 143:236–247,
March 2018.
[174] Weiling Chen, Chai Kiat Yeo, Chiew Tong Lau, and Bu Sung Lee. Leveraging social media news to
predict stock index movement using rnn-boost. Data & Knowledge Engineering, August 2018.
[175] Ziniu Hu, Weiqing Liu, Jiang Bian, Xuanzhe Liu, and Tie-Yan Liu. Listening to chaotic whispers:
A deep learning framework for news-oriented stock trend prediction. In Proceedings of the Eleventh
ACM International Conference on Web Search and Data Mining, WSDM ’18, pages 261–269, New
York, NY, USA, 2018. ACM.
[176] Xiaodong Li, Jingjing Cao, and Zhaoqing Pan. Market impact analysis via deep learned architectures.
Neural Computing and Applications, March 2018.
[177] Che-Yu Lee and Von-Wun Soo. Predict stock price with financial news based on recurrent convolu-
tional neural networks. In 2017 Conference on Technologies and Applications of Artificial Intelligence
(TAAI). IEEE, December 2017.
[178] Shotaro Minami. Predicting equity price with corporate action events using lstm-rnn. Journal of
Mathematical Finance, 08(01):58–63, 2018.
[179] Akira Yoshihara, Kazuki Fujikawa, Kazuhiro Seki, and Kuniaki Uehara. Predicting stock market
trends by recurrent deep neural networks. In Lecture Notes in Computer Science, pages 759–769.
Springer International Publishing, 2014.
[180] Przemyslaw Buczkowski. Predicting stock trends based on expert recommendations using gru/lstm
neural networks. In Lecture Notes in Computer Science, pages 708–717. Springer International Pub-
lishing, 2017.
[181] Leonardo dos Santos Pinheiro and Mark Dras. Stock market prediction with deep learning: A
character-based neural language model for event-based trading. In Proceedings of the Australasian
Language Technology Association Workshop 2017, pages 6–15, 2017.
[182] Yang Liu, Qingguo Zeng, Huanrui Yang, and Adrian Carrio. Stock price movement prediction from
financial news with deep learning and knowledge graph embedding. In Knowledge Management and
Acquisition for Intelligent Systems, pages 102–113. Springer International Publishing, 2018.
[183] Takashi Matsubara, Ryo Akita, and Kuniaki Uehara. Stock price prediction by deep neural generative
model of news articles. IEICE Transactions on Information and Systems, E101.D(4):901–908, 2018.
[184] Janderson B. Nascimento and Marco Cristo. The impact of structured event embeddings on scalable
stock forecasting models. In Proceedings of the 21st Brazilian Symposium on Multimedia and the Web
- WebMedia15. ACM Press, 2015.
[185] Md Shad Akhtar, Abhishek Kumar, Deepanway Ghosal, Asif Ekbal, and Pushpak Bhattacharyya.
A multilayer perceptron based ensemble technique for fine-grained financial sentiment analysis. In
Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, pages
540–546. Association for Computational Linguistics, 2017.
[186] Ching-Yun Chang, Yue Zhang, Zhiyang Teng, Zahn Bozanic, and Bin Ke. Measuring the information
51
content of financial news. In COLING, 2016.
[187] Hitkul Jangid, Shivangi Singhal, Rajiv Ratn Shah, and Roger Zimmermann. Aspect-based financial
sentiment analysis using deep learning. In Companion of the The Web Conference 2018 on The Web
Conference 2018 - WWW18. ACM Press, 2018.
[188] Shijia E., Li Yang, Mohan Zhang, and Yang Xiang. Aspect-based financial sentiment analysis with
deep neural networks. In Companion of the The Web Conference 2018 on The Web Conference 2018
- WWW18. ACM Press, 2018.
[189] Sahar Sohangir, Dingding Wang, Anna Pomeranets, and Taghi M. Khoshgoftaar. Big data: Deep
learning for financial sentiment analysis. Journal of Big Data, 5(1), January 2018.
[190] Nader Mahmoudi, Paul Docherty, and Pablo Moscato. Deep neural networks understand investors
better. Decision Support Systems, 112:23–34, August 2018.
[191] Shiori Kitamori, Hiroyuki Sakai, and Hiroki Sakaji. Extraction of sentences concerning business
performance forecast and economic forecast from summaries of financial statements by deep learning.
In 2017 IEEE Symposium Series on Computational Intelligence (SSCI). IEEE, November 2017.
[192] Guangyuan Piao and John G. Breslin. Financial aspect and sentiment predictions with deep neural
networks. In Companion of the The Web Conference 2018 on The Web Conference 2018 - WWW18.
ACM Press, 2018.
[193] Weifeng Li and Hsinchun Chen. Identifying top sellers in underground economy using deep learning-
based sentiment analysis. In 2014 IEEE Joint Intelligence and Security Informatics Conference. IEEE,
September 2014.
[194] Andrew Moore and Paul Rayson. Lancaster a at semeval-2017 task 5: Evaluation metrics matter: pre-
dicting sentiment from financial news headlines. In Proceedings of the 11th International Workshop on
Semantic Evaluation (SemEval-2017), pages 581–585, Vancouver, Canada, August 2017. Association
for Computational Linguistics.
[195] Josh Jia-Ching Ying, Po-Yu Huang, Chih-Kai Chang, and Don-Lin Yang. A preliminary study on deep
learning for predicting social insurance payment behavior. In 2017 IEEE International Conference on
Big Data (Big Data). IEEE, December 2017.
[196] Sahar Sohangir and Dingding Wang. Finding expert authors in financial forum using deep learning
methods. In 2018 Second IEEE International Conference on Robotic Computing (IRC). IEEE, January
2018.
[197] Vadim Sokolov. Discussion of ’deep learning for finance: deep portfolios’. Applied Stochastic Models
in Business and Industry, 33(1):16–18, 2017.
[198] Abdelali El Bouchti, Ahmed Chakroun, Hassan Abbar, and Chafik Okar. Fraud detection in banking
using deep reinforcement learning. In 2017 Seventh International Conference on Innovative Computing
Technology (INTECH). IEEE, August 2017.
[199] Matthew Dixon, Diego Klabjan, and Jin Hoon Bang. Implementing deep neural networks for financial
market prediction on the intel xeon phi. In Proceedings of the 8th Workshop on High Performance
Computational Finance - WHPCF15. ACM Press, 2015.
[200] John Alberg and Zachary Chase Lipton. Improving factor-based quantitative investing by forecasting
company fundamentals. CoRR, abs/1711.04837, 2017.
[201] Kee-Hoon Kim, Chang-Seok Lee, Sang-Muk Jo, and Sung-Bae Cho. Predicting the success of bank
telemarketing using deep convolutional neural network. In 2015 7th International Conference of Soft
Computing and Pattern Recognition (SoCPaR). IEEE, November 2015.
[202] Joonhyuck Lee, Dong Sik Jang, and Sangsung Park. Deep learning-based corporate performance
prediction model considering technical capability. Sustainability, 9(6), May 2017.
52