TY - GEN
T1 - Algorithmic trading behavior identification using reward learning method
AU - Yang, Steve Y.
AU - Qiao, Qifeng
AU - Beling, Peter A.
AU - Scherer, William T.
N1 - Publisher Copyright:
© 2014 IEEE.
PY - 2014/9/3
Y1 - 2014/9/3
N2 - Identifying and understanding the impact of algorithmic trading on financial markets has become a critical issue for market operators and regulators. Advanced data feed and audit trail information from market operators now make the full observation of market participants' actions possible. A key question is the extent to which it is possible to understand and characterize the behavior of individual participants from observations of trading actions. In this paper, we consider the basic problems of categorizing and recognizing traders (or, equivalently, trading algorithms) on the basis observed limit orders. Our approach, which is based on inverse reinforcement learning (IRL), is to model trading decisions as a Markov decision process and then use observations of an optimal decision policy to find the reward function. The approach strikes a balance between two desirable features in that it captures key empirical properties of order book dynamics and yet remains computationally tractable. Making use of a real-world data set from the E-Mini futures contract, we compare two principal IRL variants, linear IRL and Gaussian process IRL. Results suggest that IRL-based feature spaces support accurate classification and meaningful clustering.
AB - Identifying and understanding the impact of algorithmic trading on financial markets has become a critical issue for market operators and regulators. Advanced data feed and audit trail information from market operators now make the full observation of market participants' actions possible. A key question is the extent to which it is possible to understand and characterize the behavior of individual participants from observations of trading actions. In this paper, we consider the basic problems of categorizing and recognizing traders (or, equivalently, trading algorithms) on the basis observed limit orders. Our approach, which is based on inverse reinforcement learning (IRL), is to model trading decisions as a Markov decision process and then use observations of an optimal decision policy to find the reward function. The approach strikes a balance between two desirable features in that it captures key empirical properties of order book dynamics and yet remains computationally tractable. Making use of a real-world data set from the E-Mini futures contract, we compare two principal IRL variants, linear IRL and Gaussian process IRL. Results suggest that IRL-based feature spaces support accurate classification and meaningful clustering.
KW - Algorithmic Trading
KW - Behavioral Finance
KW - Gaussian Process
KW - High Frequency Trading
KW - Inverse Reinforcement Learning
KW - Markov Decision Process
KW - Support Vector Machine
UR - http://www.scopus.com/inward/record.url?scp=84908472703&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=84908472703&partnerID=8YFLogxK
U2 - 10.1109/IJCNN.2014.6889878
DO - 10.1109/IJCNN.2014.6889878
M3 - Conference contribution
AN - SCOPUS:84908472703
T3 - Proceedings of the International Joint Conference on Neural Networks
SP - 3807
EP - 3814
BT - Proceedings of the International Joint Conference on Neural Networks
T2 - 2014 International Joint Conference on Neural Networks, IJCNN 2014
Y2 - 6 July 2014 through 11 July 2014
ER -