{"title":"金融工程的 BERT 与 GPT","authors":"Edward Sharkey, Philip Treleaven","doi":"arxiv-2405.12990","DOIUrl":null,"url":null,"abstract":"The paper benchmarks several Transformer models [4], to show how these models\ncan judge sentiment from a news event. This signal can then be used for\ndownstream modelling and signal identification for commodity trading. We find\nthat fine-tuned BERT models outperform fine-tuned or vanilla GPT models on this\ntask. Transformer models have revolutionized the field of natural language\nprocessing (NLP) in recent years, achieving state-of-the-art results on various\ntasks such as machine translation, text summarization, question answering, and\nnatural language generation. Among the most prominent transformer models are\nBidirectional Encoder Representations from Transformers (BERT) and Generative\nPre-trained Transformer (GPT), which differ in their architectures and\nobjectives. A CopBERT model training data and process overview is provided. The CopBERT\nmodel outperforms similar domain specific BERT trained models such as FinBERT.\nThe below confusion matrices show the performance on CopBERT & CopGPT\nrespectively. We see a ~10 percent increase in f1_score when compare CopBERT vs\nGPT4 and 16 percent increase vs CopGPT. Whilst GPT4 is dominant It highlights\nthe importance of considering alternatives to GPT models for financial\nengineering tasks, given risks of hallucinations, and challenges with\ninterpretability. We unsurprisingly see the larger LLMs outperform the BERT\nmodels, with predictive power. In summary BERT is partially the new XGboost,\nwhat it lacks in predictive power it provides with higher levels of\ninterpretability. Concluding that BERT models might not be the next XGboost\n[2], but represent an interesting alternative for financial engineering tasks,\nthat require a blend of interpretability and accuracy.","PeriodicalId":501139,"journal":{"name":"arXiv - QuantFin - Statistical Finance","volume":"19 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-04-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"BERT vs GPT for financial engineering\",\"authors\":\"Edward Sharkey, Philip Treleaven\",\"doi\":\"arxiv-2405.12990\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"The paper benchmarks several Transformer models [4], to show how these models\\ncan judge sentiment from a news event. This signal can then be used for\\ndownstream modelling and signal identification for commodity trading. We find\\nthat fine-tuned BERT models outperform fine-tuned or vanilla GPT models on this\\ntask. Transformer models have revolutionized the field of natural language\\nprocessing (NLP) in recent years, achieving state-of-the-art results on various\\ntasks such as machine translation, text summarization, question answering, and\\nnatural language generation. Among the most prominent transformer models are\\nBidirectional Encoder Representations from Transformers (BERT) and Generative\\nPre-trained Transformer (GPT), which differ in their architectures and\\nobjectives. A CopBERT model training data and process overview is provided. The CopBERT\\nmodel outperforms similar domain specific BERT trained models such as FinBERT.\\nThe below confusion matrices show the performance on CopBERT & CopGPT\\nrespectively. We see a ~10 percent increase in f1_score when compare CopBERT vs\\nGPT4 and 16 percent increase vs CopGPT. Whilst GPT4 is dominant It highlights\\nthe importance of considering alternatives to GPT models for financial\\nengineering tasks, given risks of hallucinations, and challenges with\\ninterpretability. We unsurprisingly see the larger LLMs outperform the BERT\\nmodels, with predictive power. In summary BERT is partially the new XGboost,\\nwhat it lacks in predictive power it provides with higher levels of\\ninterpretability. Concluding that BERT models might not be the next XGboost\\n[2], but represent an interesting alternative for financial engineering tasks,\\nthat require a blend of interpretability and accuracy.\",\"PeriodicalId\":501139,\"journal\":{\"name\":\"arXiv - QuantFin - Statistical Finance\",\"volume\":\"19 1\",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-04-24\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"arXiv - QuantFin - Statistical Finance\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/arxiv-2405.12990\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - QuantFin - Statistical Finance","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2405.12990","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
The paper benchmarks several Transformer models [4], to show how these models
can judge sentiment from a news event. This signal can then be used for
downstream modelling and signal identification for commodity trading. We find
that fine-tuned BERT models outperform fine-tuned or vanilla GPT models on this
task. Transformer models have revolutionized the field of natural language
processing (NLP) in recent years, achieving state-of-the-art results on various
tasks such as machine translation, text summarization, question answering, and
natural language generation. Among the most prominent transformer models are
Bidirectional Encoder Representations from Transformers (BERT) and Generative
Pre-trained Transformer (GPT), which differ in their architectures and
objectives. A CopBERT model training data and process overview is provided. The CopBERT
model outperforms similar domain specific BERT trained models such as FinBERT.
The below confusion matrices show the performance on CopBERT & CopGPT
respectively. We see a ~10 percent increase in f1_score when compare CopBERT vs
GPT4 and 16 percent increase vs CopGPT. Whilst GPT4 is dominant It highlights
the importance of considering alternatives to GPT models for financial
engineering tasks, given risks of hallucinations, and challenges with
interpretability. We unsurprisingly see the larger LLMs outperform the BERT
models, with predictive power. In summary BERT is partially the new XGboost,
what it lacks in predictive power it provides with higher levels of
interpretability. Concluding that BERT models might not be the next XGboost
[2], but represent an interesting alternative for financial engineering tasks,
that require a blend of interpretability and accuracy.