Hanyu Zhao, Li Du, Yiming Ju, Chengwei Wu, Tengfei Pan
{"title":"Beyond IID: Optimizing Instruction Learning from the Perspective of Instruction Interaction and Dependency","authors":"Hanyu Zhao, Li Du, Yiming Ju, Chengwei Wu, Tengfei Pan","doi":"arxiv-2409.07045","DOIUrl":null,"url":null,"abstract":"With the availability of various instruction datasets, a pivotal challenge is\nhow to effectively select and integrate these instructions to fine-tune large\nlanguage models (LLMs). Previous research mainly focuses on selecting\nindividual high-quality instructions. However, these works overlooked the joint\ninteractions and dependencies between different categories of instructions,\nleading to suboptimal selection strategies. Moreover, the nature of these\ninteraction patterns remains largely unexplored, let alone optimize the\ninstruction set with regard to them. To fill these gaps, in this paper, we: (1)\nsystemically investigate interaction and dependency patterns between different\ncategories of instructions, (2) manage to optimize the instruction set\nconcerning the interaction patterns using a linear programming-based method,\nand optimize the learning schema of SFT using an instruction dependency\ntaxonomy guided curriculum learning. Experimental results across different LLMs\ndemonstrate improved performance over strong baselines on widely adopted\nbenchmarks.","PeriodicalId":501030,"journal":{"name":"arXiv - CS - Computation and Language","volume":"106 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-09-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Computation and Language","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.07045","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
With the availability of various instruction datasets, a pivotal challenge is
how to effectively select and integrate these instructions to fine-tune large
language models (LLMs). Previous research mainly focuses on selecting
individual high-quality instructions. However, these works overlooked the joint
interactions and dependencies between different categories of instructions,
leading to suboptimal selection strategies. Moreover, the nature of these
interaction patterns remains largely unexplored, let alone optimize the
instruction set with regard to them. To fill these gaps, in this paper, we: (1)
systemically investigate interaction and dependency patterns between different
categories of instructions, (2) manage to optimize the instruction set
concerning the interaction patterns using a linear programming-based method,
and optimize the learning schema of SFT using an instruction dependency
taxonomy guided curriculum learning. Experimental results across different LLMs
demonstrate improved performance over strong baselines on widely adopted
benchmarks.