{"title":"TexShape:语言模型的信息论句子嵌入","authors":"H. Kaan Kale, Homa Esfahanizadeh, Noel Elias, Oguzhan Baser, Muriel Medard, Sriram Vishwanath","doi":"arxiv-2402.05132","DOIUrl":null,"url":null,"abstract":"With the exponential growth in data volume and the emergence of\ndata-intensive applications, particularly in the field of machine learning,\nconcerns related to resource utilization, privacy, and fairness have become\nparamount. This paper focuses on the textual domain of data and addresses\nchallenges regarding encoding sentences to their optimized representations\nthrough the lens of information-theory. In particular, we use empirical\nestimates of mutual information, using the Donsker-Varadhan definition of\nKullback-Leibler divergence. Our approach leverages this estimation to train an\ninformation-theoretic sentence embedding, called TexShape, for (task-based)\ndata compression or for filtering out sensitive information, enhancing privacy\nand fairness. In this study, we employ a benchmark language model for initial\ntext representation, complemented by neural networks for information-theoretic\ncompression and mutual information estimations. Our experiments demonstrate\nsignificant advancements in preserving maximal targeted information and minimal\nsensitive information over adverse compression ratios, in terms of predictive\naccuracy of downstream models that are trained using the compressed data.","PeriodicalId":501433,"journal":{"name":"arXiv - CS - Information Theory","volume":"19 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-02-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"TexShape: Information Theoretic Sentence Embedding for Language Models\",\"authors\":\"H. Kaan Kale, Homa Esfahanizadeh, Noel Elias, Oguzhan Baser, Muriel Medard, Sriram Vishwanath\",\"doi\":\"arxiv-2402.05132\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"With the exponential growth in data volume and the emergence of\\ndata-intensive applications, particularly in the field of machine learning,\\nconcerns related to resource utilization, privacy, and fairness have become\\nparamount. This paper focuses on the textual domain of data and addresses\\nchallenges regarding encoding sentences to their optimized representations\\nthrough the lens of information-theory. In particular, we use empirical\\nestimates of mutual information, using the Donsker-Varadhan definition of\\nKullback-Leibler divergence. Our approach leverages this estimation to train an\\ninformation-theoretic sentence embedding, called TexShape, for (task-based)\\ndata compression or for filtering out sensitive information, enhancing privacy\\nand fairness. In this study, we employ a benchmark language model for initial\\ntext representation, complemented by neural networks for information-theoretic\\ncompression and mutual information estimations. Our experiments demonstrate\\nsignificant advancements in preserving maximal targeted information and minimal\\nsensitive information over adverse compression ratios, in terms of predictive\\naccuracy of downstream models that are trained using the compressed data.\",\"PeriodicalId\":501433,\"journal\":{\"name\":\"arXiv - CS - Information Theory\",\"volume\":\"19 1\",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-02-05\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"arXiv - CS - Information Theory\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/arxiv-2402.05132\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Information Theory","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2402.05132","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
TexShape: Information Theoretic Sentence Embedding for Language Models
With the exponential growth in data volume and the emergence of
data-intensive applications, particularly in the field of machine learning,
concerns related to resource utilization, privacy, and fairness have become
paramount. This paper focuses on the textual domain of data and addresses
challenges regarding encoding sentences to their optimized representations
through the lens of information-theory. In particular, we use empirical
estimates of mutual information, using the Donsker-Varadhan definition of
Kullback-Leibler divergence. Our approach leverages this estimation to train an
information-theoretic sentence embedding, called TexShape, for (task-based)
data compression or for filtering out sensitive information, enhancing privacy
and fairness. In this study, we employ a benchmark language model for initial
text representation, complemented by neural networks for information-theoretic
compression and mutual information estimations. Our experiments demonstrate
significant advancements in preserving maximal targeted information and minimal
sensitive information over adverse compression ratios, in terms of predictive
accuracy of downstream models that are trained using the compressed data.