{"title":"金融股票研究报告的结构 -- 利用 Llama 3 和 GPT-4 识别金融分析师报告中最常见的问题,实现股票研究自动化","authors":"Adria Pop, Jan Spörer, Siegfried Handschuh","doi":"arxiv-2407.18327","DOIUrl":null,"url":null,"abstract":"This research dissects financial equity research reports (ERRs) by mapping\ntheir content into categories. There is insufficient empirical analysis of the questions answered in ERRs.\nIn particular, it is not understood how frequently certain information appears,\nwhat information is considered essential, and what information requires human\njudgment to distill into an ERR. The study analyzes 72 ERRs sentence-by-sentence, classifying their 4940\nsentences into 169 unique question archetypes. We did not predefine the\nquestions but derived them solely from the statements in the ERRs. This\napproach provides an unbiased view of the content of the observed ERRs.\nSubsequently, we used public corporate reports to classify the questions'\npotential for automation. Answers were labeled \"text-extractable\" if the\nanswers to the question were accessible in corporate reports. 78.7% of the questions in ERRs can be automated. Those automatable question\nconsist of 48.2% text-extractable (suited to processing by large language\nmodels, LLMs) and 30.5% database-extractable questions. Only 21.3% of questions\nrequire human judgment to answer. We empirically validate using Llama-3-70B and GPT-4-turbo-2024-04-09 that\nrecent advances in language generation and information extraction enable the\nautomation of approximately 80% of the statements in ERRs. Surprisingly, the\nmodels complement each other's strengths and weaknesses well. The research confirms that the current writing process of ERRs can likely\nbenefit from additional automation, improving quality and efficiency. The\nresearch thus allows us to quantify the potential impacts of introducing large\nlanguage models in the ERR writing process. The full question list, including the archetypes and their frequency, will be\nmade available online after peer review.","PeriodicalId":501294,"journal":{"name":"arXiv - QuantFin - Computational Finance","volume":"50 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-07-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"The Structure of Financial Equity Research Reports -- Identification of the Most Frequently Asked Questions in Financial Analyst Reports to Automate Equity Research Using Llama 3 and GPT-4\",\"authors\":\"Adria Pop, Jan Spörer, Siegfried Handschuh\",\"doi\":\"arxiv-2407.18327\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"This research dissects financial equity research reports (ERRs) by mapping\\ntheir content into categories. There is insufficient empirical analysis of the questions answered in ERRs.\\nIn particular, it is not understood how frequently certain information appears,\\nwhat information is considered essential, and what information requires human\\njudgment to distill into an ERR. The study analyzes 72 ERRs sentence-by-sentence, classifying their 4940\\nsentences into 169 unique question archetypes. We did not predefine the\\nquestions but derived them solely from the statements in the ERRs. This\\napproach provides an unbiased view of the content of the observed ERRs.\\nSubsequently, we used public corporate reports to classify the questions'\\npotential for automation. Answers were labeled \\\"text-extractable\\\" if the\\nanswers to the question were accessible in corporate reports. 78.7% of the questions in ERRs can be automated. Those automatable question\\nconsist of 48.2% text-extractable (suited to processing by large language\\nmodels, LLMs) and 30.5% database-extractable questions. Only 21.3% of questions\\nrequire human judgment to answer. We empirically validate using Llama-3-70B and GPT-4-turbo-2024-04-09 that\\nrecent advances in language generation and information extraction enable the\\nautomation of approximately 80% of the statements in ERRs. Surprisingly, the\\nmodels complement each other's strengths and weaknesses well. The research confirms that the current writing process of ERRs can likely\\nbenefit from additional automation, improving quality and efficiency. The\\nresearch thus allows us to quantify the potential impacts of introducing large\\nlanguage models in the ERR writing process. The full question list, including the archetypes and their frequency, will be\\nmade available online after peer review.\",\"PeriodicalId\":501294,\"journal\":{\"name\":\"arXiv - QuantFin - Computational Finance\",\"volume\":\"50 1\",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-07-04\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"arXiv - QuantFin - Computational Finance\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/arxiv-2407.18327\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - QuantFin - Computational Finance","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2407.18327","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
The Structure of Financial Equity Research Reports -- Identification of the Most Frequently Asked Questions in Financial Analyst Reports to Automate Equity Research Using Llama 3 and GPT-4
This research dissects financial equity research reports (ERRs) by mapping
their content into categories. There is insufficient empirical analysis of the questions answered in ERRs.
In particular, it is not understood how frequently certain information appears,
what information is considered essential, and what information requires human
judgment to distill into an ERR. The study analyzes 72 ERRs sentence-by-sentence, classifying their 4940
sentences into 169 unique question archetypes. We did not predefine the
questions but derived them solely from the statements in the ERRs. This
approach provides an unbiased view of the content of the observed ERRs.
Subsequently, we used public corporate reports to classify the questions'
potential for automation. Answers were labeled "text-extractable" if the
answers to the question were accessible in corporate reports. 78.7% of the questions in ERRs can be automated. Those automatable question
consist of 48.2% text-extractable (suited to processing by large language
models, LLMs) and 30.5% database-extractable questions. Only 21.3% of questions
require human judgment to answer. We empirically validate using Llama-3-70B and GPT-4-turbo-2024-04-09 that
recent advances in language generation and information extraction enable the
automation of approximately 80% of the statements in ERRs. Surprisingly, the
models complement each other's strengths and weaknesses well. The research confirms that the current writing process of ERRs can likely
benefit from additional automation, improving quality and efficiency. The
research thus allows us to quantify the potential impacts of introducing large
language models in the ERR writing process. The full question list, including the archetypes and their frequency, will be
made available online after peer review.