File Download

There are no files associated with this item.

  • Find it @ UNIST can give you direct access to the published full text of this article. (UNISTARs only)
Related Researcher

이용재

Lee, Yongjae
Financial Engineering Lab.
Read More

Views & Downloads

Detailed Information

Cited time in webofscience Cited time in scopus
Metadata Downloads

Full metadata record

DC Field Value Language
dc.citation.conferencePlace US -
dc.citation.conferencePlace NYC; Online -
dc.citation.endPage 231 -
dc.citation.startPage 223 -
dc.citation.title ACM International Conference on AI in Finance -
dc.contributor.author Chung, Guhyuk -
dc.contributor.author Chung, Munki -
dc.contributor.author Lee, Yongjae -
dc.contributor.author Kim, Woo Chang -
dc.date.accessioned 2023-12-19T11:27:01Z -
dc.date.available 2023-12-19T11:27:01Z -
dc.date.created 2022-11-15 -
dc.date.issued 2022-11-02 -
dc.description.abstract Market making strategy is one of the most popular high frequency trading strategies, where a market maker continuously quotes on both bid and ask side of the limit order book to profit from capturing bid-ask spread and to provide liquidity to the market. A market maker should consider three types of risk: 1) inventory risk, 2) adverse selection risk, and 3) non-execution risk. While there have been a lot of studies on market making via deep reinforcement learning, most of them focus on the first risk. However, in highly competitive markets, the latter two risks are very important to make stable profit from market making. For better control of the latter two risks, it is important to reserve good queue position of their resting limit orders. For this purpose, practitioners frequently adopt order stacking framework where their limit orders are quoted at multiple price levels beyond the best limit price. To the best of our knowledge, there have been no studies that adopt order stacking framework for market making. In this regard, we develop a deep reinforcement learning model for market making under order stacking framework. We use a modified state representation to efficiently encode the queue positions of the resting limit orders. We conduct comprehensive ablation study to show that by utilizing deep reinforcement learning, a market making agent under order stacking framework successfully learns to improve the PL while reducing various risks. For the training and testing of our model, we use complete limit order book data of KOSPI200 Index Futures from November 1, 2019 to January 31, 2020 which is comprised of 61 trading days. -
dc.identifier.bibliographicCitation ACM International Conference on AI in Finance, pp.223 - 231 -
dc.identifier.doi 10.1145/3533271.3561789 -
dc.identifier.uri https://scholarworks.unist.ac.kr/handle/201301/60028 -
dc.publisher ACM -
dc.title Market Making under Order Stacking Framework: A Deep Reinforcement Learning Approach -
dc.type Conference Paper -
dc.date.conferenceDate 2022-11-02 -

qrcode

Items in Repository are protected by copyright, with all rights reserved, unless otherwise indicated.