Yifan Wang, David Stevens, Pranay Shah, Wenwen Jiang, Miao Liu, Xu Chen, Robert Kuo, Na Li, Boying Gong, Daniel Lee, Jiabo Hu, Ning Zhang, Bob Kamma
{"title":"Model-in-the-Loop (MILO): Accelerating Multimodal AI Data Annotation with LLMs","authors":"Yifan Wang, David Stevens, Pranay Shah, Wenwen Jiang, Miao Liu, Xu Chen, Robert Kuo, Na Li, Boying Gong, Daniel Lee, Jiabo Hu, Ning Zhang, Bob Kamma","doi":"arxiv-2409.10702","DOIUrl":null,"url":null,"abstract":"The growing demand for AI training data has transformed data annotation into\na global industry, but traditional approaches relying on human annotators are\noften time-consuming, labor-intensive, and prone to inconsistent quality. We\npropose the Model-in-the-Loop (MILO) framework, which integrates AI/ML models\ninto the annotation process. Our research introduces a collaborative paradigm\nthat leverages the strengths of both professional human annotators and large\nlanguage models (LLMs). By employing LLMs as pre-annotation and real-time\nassistants, and judges on annotator responses, MILO enables effective\ninteraction patterns between human annotators and LLMs. Three empirical studies\non multimodal data annotation demonstrate MILO's efficacy in reducing handling\ntime, improving data quality, and enhancing annotator experiences. We also\nintroduce quality rubrics for flexible evaluation and fine-grained feedback on\nopen-ended annotations. The MILO framework has implications for accelerating\nAI/ML development, reducing reliance on human annotation alone, and promoting\nbetter alignment between human and machine values.","PeriodicalId":501541,"journal":{"name":"arXiv - CS - Human-Computer Interaction","volume":"54 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Human-Computer Interaction","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.10702","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
The growing demand for AI training data has transformed data annotation into
a global industry, but traditional approaches relying on human annotators are
often time-consuming, labor-intensive, and prone to inconsistent quality. We
propose the Model-in-the-Loop (MILO) framework, which integrates AI/ML models
into the annotation process. Our research introduces a collaborative paradigm
that leverages the strengths of both professional human annotators and large
language models (LLMs). By employing LLMs as pre-annotation and real-time
assistants, and judges on annotator responses, MILO enables effective
interaction patterns between human annotators and LLMs. Three empirical studies
on multimodal data annotation demonstrate MILO's efficacy in reducing handling
time, improving data quality, and enhancing annotator experiences. We also
introduce quality rubrics for flexible evaluation and fine-grained feedback on
open-ended annotations. The MILO framework has implications for accelerating
AI/ML development, reducing reliance on human annotation alone, and promoting
better alignment between human and machine values.