Chuqiao Zong, Chaojie Wang, Molei Qin, Lei Feng, Xinrun Wang, Bo An
{"title":"MacroHFT:高频交易中的记忆增强型情境感知强化学习","authors":"Chuqiao Zong, Chaojie Wang, Molei Qin, Lei Feng, Xinrun Wang, Bo An","doi":"arxiv-2406.14537","DOIUrl":null,"url":null,"abstract":"High-frequency trading (HFT) that executes algorithmic trading in short time\nscales, has recently occupied the majority of cryptocurrency market. Besides\ntraditional quantitative trading methods, reinforcement learning (RL) has\nbecome another appealing approach for HFT due to its terrific ability of\nhandling high-dimensional financial data and solving sophisticated sequential\ndecision-making problems, \\emph{e.g.,} hierarchical reinforcement learning\n(HRL) has shown its promising performance on second-level HFT by training a\nrouter to select only one sub-agent from the agent pool to execute the current\ntransaction. However, existing RL methods for HFT still have some defects: 1)\nstandard RL-based trading agents suffer from the overfitting issue, preventing\nthem from making effective policy adjustments based on financial context; 2)\ndue to the rapid changes in market conditions, investment decisions made by an\nindividual agent are usually one-sided and highly biased, which might lead to\nsignificant loss in extreme markets. To tackle these problems, we propose a\nnovel Memory Augmented Context-aware Reinforcement learning method On HFT,\n\\emph{a.k.a.} MacroHFT, which consists of two training phases: 1) we first\ntrain multiple types of sub-agents with the market data decomposed according to\nvarious financial indicators, specifically market trend and volatility, where\neach agent owns a conditional adapter to adjust its trading policy according to\nmarket conditions; 2) then we train a hyper-agent to mix the decisions from\nthese sub-agents and output a consistently profitable meta-policy to handle\nrapid market fluctuations, equipped with a memory mechanism to enhance the\ncapability of decision-making. Extensive experiments on various cryptocurrency\nmarkets demonstrate that MacroHFT can achieve state-of-the-art performance on\nminute-level trading tasks.","PeriodicalId":501478,"journal":{"name":"arXiv - QuantFin - Trading and Market Microstructure","volume":"97 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"MacroHFT: Memory Augmented Context-aware Reinforcement Learning On High Frequency Trading\",\"authors\":\"Chuqiao Zong, Chaojie Wang, Molei Qin, Lei Feng, Xinrun Wang, Bo An\",\"doi\":\"arxiv-2406.14537\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"High-frequency trading (HFT) that executes algorithmic trading in short time\\nscales, has recently occupied the majority of cryptocurrency market. Besides\\ntraditional quantitative trading methods, reinforcement learning (RL) has\\nbecome another appealing approach for HFT due to its terrific ability of\\nhandling high-dimensional financial data and solving sophisticated sequential\\ndecision-making problems, \\\\emph{e.g.,} hierarchical reinforcement learning\\n(HRL) has shown its promising performance on second-level HFT by training a\\nrouter to select only one sub-agent from the agent pool to execute the current\\ntransaction. However, existing RL methods for HFT still have some defects: 1)\\nstandard RL-based trading agents suffer from the overfitting issue, preventing\\nthem from making effective policy adjustments based on financial context; 2)\\ndue to the rapid changes in market conditions, investment decisions made by an\\nindividual agent are usually one-sided and highly biased, which might lead to\\nsignificant loss in extreme markets. To tackle these problems, we propose a\\nnovel Memory Augmented Context-aware Reinforcement learning method On HFT,\\n\\\\emph{a.k.a.} MacroHFT, which consists of two training phases: 1) we first\\ntrain multiple types of sub-agents with the market data decomposed according to\\nvarious financial indicators, specifically market trend and volatility, where\\neach agent owns a conditional adapter to adjust its trading policy according to\\nmarket conditions; 2) then we train a hyper-agent to mix the decisions from\\nthese sub-agents and output a consistently profitable meta-policy to handle\\nrapid market fluctuations, equipped with a memory mechanism to enhance the\\ncapability of decision-making. Extensive experiments on various cryptocurrency\\nmarkets demonstrate that MacroHFT can achieve state-of-the-art performance on\\nminute-level trading tasks.\",\"PeriodicalId\":501478,\"journal\":{\"name\":\"arXiv - QuantFin - Trading and Market Microstructure\",\"volume\":\"97 1\",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-06-20\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"arXiv - QuantFin - Trading and Market Microstructure\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/arxiv-2406.14537\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - QuantFin - Trading and Market Microstructure","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2406.14537","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
MacroHFT: Memory Augmented Context-aware Reinforcement Learning On High Frequency Trading
High-frequency trading (HFT) that executes algorithmic trading in short time
scales, has recently occupied the majority of cryptocurrency market. Besides
traditional quantitative trading methods, reinforcement learning (RL) has
become another appealing approach for HFT due to its terrific ability of
handling high-dimensional financial data and solving sophisticated sequential
decision-making problems, \emph{e.g.,} hierarchical reinforcement learning
(HRL) has shown its promising performance on second-level HFT by training a
router to select only one sub-agent from the agent pool to execute the current
transaction. However, existing RL methods for HFT still have some defects: 1)
standard RL-based trading agents suffer from the overfitting issue, preventing
them from making effective policy adjustments based on financial context; 2)
due to the rapid changes in market conditions, investment decisions made by an
individual agent are usually one-sided and highly biased, which might lead to
significant loss in extreme markets. To tackle these problems, we propose a
novel Memory Augmented Context-aware Reinforcement learning method On HFT,
\emph{a.k.a.} MacroHFT, which consists of two training phases: 1) we first
train multiple types of sub-agents with the market data decomposed according to
various financial indicators, specifically market trend and volatility, where
each agent owns a conditional adapter to adjust its trading policy according to
market conditions; 2) then we train a hyper-agent to mix the decisions from
these sub-agents and output a consistently profitable meta-policy to handle
rapid market fluctuations, equipped with a memory mechanism to enhance the
capability of decision-making. Extensive experiments on various cryptocurrency
markets demonstrate that MacroHFT can achieve state-of-the-art performance on
minute-level trading tasks.