首页 > 最新文献

AAAI Workshop on Artificial Intelligence with Biased or Scarce Data (AIBSD)最新文献

英文 中文
Statement of Peer Review 同行评审声明
Pub Date : 2022-05-31 DOI: 10.3390/cmsf2022003012
Kuan-Chuan Peng, Ziyan Wu
{"title":"Statement of Peer Review","authors":"Kuan-Chuan Peng, Ziyan Wu","doi":"10.3390/cmsf2022003012","DOIUrl":"https://doi.org/10.3390/cmsf2022003012","url":null,"abstract":"","PeriodicalId":127261,"journal":{"name":"AAAI Workshop on Artificial Intelligence with Biased or Scarce Data (AIBSD)","volume":"41 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-05-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116282857","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Age Should Not Matter: Towards More Accurate Pedestrian Detection via Self-Training 年龄无关紧要:通过自我训练实现更准确的行人检测
Pub Date : 2022-05-24 DOI: 10.3390/cmsf2022003011
Shunsuke Kogure, Kai Watabe, Ryosuke Yamada, Y. Aoki, Akio Nakamura, Hirokatsu Kataoka
Why is there a the disparity in the miss rates of pedestrian detection between different age attributes? In this study, we propose to (i) improve the accuracy of pedestrian detection using our pre-trained model and (ii) explore the causes of this disparity. In order to improve detection accuracy, we extend a pedestrian detection pre-training dataset, the Weakly Supervised Pedestrian Dataset (WSPD), by means of self-training, to construct our Self-Trained Person Dataset (STPD). More-over, we hypothesise the cause of the miss rate as being due to three biases: 1) the apparent bias towards “adults” versus “children,” 2) the quantity of training data bias against “chil- dren,” and 3) the scale bias of the bounding box. In addition, we constructed an evaluation dataset by manually annotat- ing “adult” and “child” bounding boxes to the INRIA Person Dataset. As a result, we confirm that the miss rate was re- duced by up to 0.4% for adults and up to 3.9% for children. In addition, we discuss the impact of the size and appearance of the bounding boxes on the disparity in miss rates and pro-vide an outlook for future research.
为什么不同年龄属性的行人检测失检率存在差异?在本研究中,我们建议(i)使用我们的预训练模型提高行人检测的准确性,(ii)探索这种差异的原因。为了提高检测精度,我们扩展了行人检测预训练数据集弱监督行人数据集(WSPD),通过自我训练来构建我们的自我训练人员数据集(STPD)。此外,我们假设缺失率的原因是由于三种偏差:1)对“成人”与“儿童”的明显偏差,2)对“儿童”的训练数据量偏差,以及3)边界框的尺度偏差。此外,我们通过在INRIA Person dataset上手动标注“成人”和“儿童”边界框,构建了一个评估数据集。结果,我们确认,失误率降低了——成人减少了0.4%,儿童减少了3.9%。此外,我们还讨论了边界框的大小和外观对脱靶率差异的影响,并对未来的研究进行了展望。
{"title":"Age Should Not Matter: Towards More Accurate Pedestrian Detection via Self-Training","authors":"Shunsuke Kogure, Kai Watabe, Ryosuke Yamada, Y. Aoki, Akio Nakamura, Hirokatsu Kataoka","doi":"10.3390/cmsf2022003011","DOIUrl":"https://doi.org/10.3390/cmsf2022003011","url":null,"abstract":"Why is there a the disparity in the miss rates of pedestrian detection between different age attributes? In this study, we propose to (i) improve the accuracy of pedestrian detection using our pre-trained model and (ii) explore the causes of this disparity. In order to improve detection accuracy, we extend a pedestrian detection pre-training dataset, the Weakly Supervised Pedestrian Dataset (WSPD), by means of self-training, to construct our Self-Trained Person Dataset (STPD). More-over, we hypothesise the cause of the miss rate as being due to three biases: 1) the apparent bias towards “adults” versus “children,” 2) the quantity of training data bias against “chil- dren,” and 3) the scale bias of the bounding box. In addition, we constructed an evaluation dataset by manually annotat- ing “adult” and “child” bounding boxes to the INRIA Person Dataset. As a result, we confirm that the miss rate was re- duced by up to 0.4% for adults and up to 3.9% for children. In addition, we discuss the impact of the size and appearance of the bounding boxes on the disparity in miss rates and pro-vide an outlook for future research.","PeriodicalId":127261,"journal":{"name":"AAAI Workshop on Artificial Intelligence with Biased or Scarce Data (AIBSD)","volume":"69 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-05-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117146113","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Dual Complementary Prototype Learning for Few-Shot Segmentation 基于双互补原型学习的少镜头分割
Pub Date : 2022-04-29 DOI: 10.3390/cmsf2022003008
Q. Ren, Jie Chen
: Few-shot semantic segmentation aims to transfer knowledge from base classes with sufficient data to represent novel classes with limited few-shot samples. Recent methods follow a metric learning framework with prototypes for foreground representation. However, they still face the challenge of segmentation of novel classes due to inadequate representation of foreground and lack of discriminability between foreground and background. To address this problem, we propose the Dual Complementary prototype Network (DCNet). Firstly, we design a training-free Complementary Prototype Generation (CPG) module to extract comprehensive information from the mask region in the support image. Secondly, we design a Background Guided Learning (BGL) as a complementary branch of the foreground segmentation branch, which enlarges difference between the foreground and its corresponding background so that the representation of novel class in the foreground could be more discriminative. Extensive experiments on PASCAL-5 i and COCO-20 i demonstrate that our DCNet achieves state-of-the-art
: Few-shot语义分割旨在从具有足够数据的基类中转移知识,以有限的Few-shot样本表示新类。最近的方法遵循带有前景表示原型的度量学习框架。然而,由于前景的代表性不足,前景和背景之间缺乏可辨别性,它们仍然面临着新类别分割的挑战。为了解决这个问题,我们提出了双互补原型网络(DCNet)。首先,我们设计了一个无需训练的互补原型生成(CPG)模块,从支持图像的掩模区域中提取综合信息。其次,我们设计了一个背景引导学习(BGL)作为前景分割分支的补充分支,扩大前景与其对应背景之间的差异,使前景中新类的表示更具判别性。在PASCAL-5 i和COCO-20 i上进行的大量实验表明,我们的DCNet达到了最先进的水平
{"title":"Dual Complementary Prototype Learning for Few-Shot Segmentation","authors":"Q. Ren, Jie Chen","doi":"10.3390/cmsf2022003008","DOIUrl":"https://doi.org/10.3390/cmsf2022003008","url":null,"abstract":": Few-shot semantic segmentation aims to transfer knowledge from base classes with sufficient data to represent novel classes with limited few-shot samples. Recent methods follow a metric learning framework with prototypes for foreground representation. However, they still face the challenge of segmentation of novel classes due to inadequate representation of foreground and lack of discriminability between foreground and background. To address this problem, we propose the Dual Complementary prototype Network (DCNet). Firstly, we design a training-free Complementary Prototype Generation (CPG) module to extract comprehensive information from the mask region in the support image. Secondly, we design a Background Guided Learning (BGL) as a complementary branch of the foreground segmentation branch, which enlarges difference between the foreground and its corresponding background so that the representation of novel class in the foreground could be more discriminative. Extensive experiments on PASCAL-5 i and COCO-20 i demonstrate that our DCNet achieves state-of-the-art","PeriodicalId":127261,"journal":{"name":"AAAI Workshop on Artificial Intelligence with Biased or Scarce Data (AIBSD)","volume":"30 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-04-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128747245","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Extracting Salient Facts from Company Reviews with Scarce Labels 从缺少标签的公司评论中提取重要事实
Pub Date : 2022-04-29 DOI: 10.3390/cmsf2022003009
Jinfeng Li, Nikita Bhutani, Alexander Whedon, Chieh-Yang Huang, Estevam Hruschka, Yoshihiko Suhara
In this paper, we propose the task of extracting salient facts from online company reviews. Salient facts present unique and distinctive information about a company, which helps the user in deciding whether to apply to the company. We formulate the salient fact extraction task as a text classification problem, and leverage pretrained language models to tackle the problem. However, the scarcity of salient facts in company reviews causes a serious label imbalance issue, which hinders taking full advantage of pretrained language models. To address the issue, we developed two data enrichment methods: first, representation enrichment, which highlights uncommon tokens by appending special tokens, and second, label propagation, which interactively creates pseudopositive examples from unlabeled data. Experimental results on an online company review corpus show that our approach improves the performance of pretrained language models by up to an F1 score of 0.24. We also confirm that our approach competitively performs well against the state-of-the-art data augmentation method on the SemEval 2019 benchmark even when trained with only 20% of
在本文中,我们提出了从在线公司评论中提取重要事实的任务。突出事实表示关于公司的独特和独特的信息,这有助于用户决定是否申请该公司。我们将显著事实提取任务表述为文本分类问题,并利用预训练的语言模型来解决该问题。然而,由于公司审查中缺乏突出事实,导致了严重的标签不平衡问题,这阻碍了预训练语言模型的充分利用。为了解决这个问题,我们开发了两种数据充实方法:第一种是表示充实,通过附加特殊令牌来突出显示不常见的令牌;第二种是标签传播,它从未标记的数据中交互式地创建伪正示例。在一个在线公司评论语料库上的实验结果表明,我们的方法将预训练语言模型的性能提高了0.24的F1分数。我们还证实,即使只使用20%的数据增强方法进行训练,我们的方法在SemEval 2019基准上也能与最先进的数据增强方法相比表现良好
{"title":"Extracting Salient Facts from Company Reviews with Scarce Labels","authors":"Jinfeng Li, Nikita Bhutani, Alexander Whedon, Chieh-Yang Huang, Estevam Hruschka, Yoshihiko Suhara","doi":"10.3390/cmsf2022003009","DOIUrl":"https://doi.org/10.3390/cmsf2022003009","url":null,"abstract":"In this paper, we propose the task of extracting salient facts from online company reviews. Salient facts present unique and distinctive information about a company, which helps the user in deciding whether to apply to the company. We formulate the salient fact extraction task as a text classification problem, and leverage pretrained language models to tackle the problem. However, the scarcity of salient facts in company reviews causes a serious label imbalance issue, which hinders taking full advantage of pretrained language models. To address the issue, we developed two data enrichment methods: first, representation enrichment, which highlights uncommon tokens by appending special tokens, and second, label propagation, which interactively creates pseudopositive examples from unlabeled data. Experimental results on an online company review corpus show that our approach improves the performance of pretrained language models by up to an F1 score of 0.24. We also confirm that our approach competitively performs well against the state-of-the-art data augmentation method on the SemEval 2019 benchmark even when trained with only 20% of","PeriodicalId":127261,"journal":{"name":"AAAI Workshop on Artificial Intelligence with Biased or Scarce Data (AIBSD)","volume":"26 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-04-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125432127","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Super-Resolution for Brain MR Images from a Significantly Small Amount of Training Data 基于少量训练数据的脑MR图像的超分辨率
Pub Date : 2022-04-27 DOI: 10.3390/cmsf2022003007
Kumpei Ikuta, H. Iyatomi, K. Oishi, on behalf of the Alzheimer’s Disease Neuroimaging Initiative
article were obtained from the Alzheimer’s Disease Neuroimaging Initiative (ADNI) As such, the investigators within the ADNI contributed to the design and implementation of ADNI and/or provided data but did not participate in analysis or writing of this report. A complete listing of ADNI investigators can be found at Abstract: We propose two essential techniques to effectively train generative adversarial network-based super-resolution networks for brain magnetic resonance images, even when only a small number of training samples are available. First, stochastic patch sampling is proposed, which in-creases training samples by sampling many small patches from the input image. However, sampling patches and combining them causes unpleasant artifacts around patch boundaries. The second proposed method, an artifact-suppressing discriminator, suppresses the artifacts by taking two-channel input containing an original high-resolution image and a generated image. With the introduction of the proposed techniques, the network achieved generation of natural-looking MR images from only ~40 training images, and improved the area-under-curve score on Alzheimer’s disease from 76.17% to 81.57%.
因此,ADNI的研究人员参与了ADNI的设计和实施和/或提供了数据,但没有参与本报告的分析或撰写。摘要:我们提出了两种基本技术,即使只有少量的训练样本,也可以有效地训练基于生成对抗网络的脑磁共振图像超分辨率网络。首先,提出了随机斑块采样,通过从输入图像中采样许多小块来增加训练样本。然而,对斑块进行采样和组合会在斑块边界周围产生令人不快的伪影。第二种方法是伪影抑制鉴别器,它通过采用包含原始高分辨率图像和生成图像的双通道输入来抑制伪影。引入上述技术后,该网络仅用~40张训练图像就能生成外观自然的MR图像,并将阿尔茨海默病的曲线下面积评分从76.17%提高到81.57%。
{"title":"Super-Resolution for Brain MR Images from a Significantly Small Amount of Training Data","authors":"Kumpei Ikuta, H. Iyatomi, K. Oishi, on behalf of the Alzheimer’s Disease Neuroimaging Initiative","doi":"10.3390/cmsf2022003007","DOIUrl":"https://doi.org/10.3390/cmsf2022003007","url":null,"abstract":"article were obtained from the Alzheimer’s Disease Neuroimaging Initiative (ADNI) As such, the investigators within the ADNI contributed to the design and implementation of ADNI and/or provided data but did not participate in analysis or writing of this report. A complete listing of ADNI investigators can be found at Abstract: We propose two essential techniques to effectively train generative adversarial network-based super-resolution networks for brain magnetic resonance images, even when only a small number of training samples are available. First, stochastic patch sampling is proposed, which in-creases training samples by sampling many small patches from the input image. However, sampling patches and combining them causes unpleasant artifacts around patch boundaries. The second proposed method, an artifact-suppressing discriminator, suppresses the artifacts by taking two-channel input containing an original high-resolution image and a generated image. With the introduction of the proposed techniques, the network achieved generation of natural-looking MR images from only ~40 training images, and improved the area-under-curve score on Alzheimer’s disease from 76.17% to 81.57%.","PeriodicalId":127261,"journal":{"name":"AAAI Workshop on Artificial Intelligence with Biased or Scarce Data (AIBSD)","volume":"29 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-04-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126994606","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Quantifying Bias in a Face Verification System 人脸验证系统中的量化偏差
Pub Date : 2022-04-20 DOI: 10.3390/cmsf2022003006
Megan Frisella, Pooya Khorrami, J. Matterer, K. Kratkiewicz, P. Torres-Carrasquillo
: Machine learning models perform face verification (FV) for a variety of highly consequential applications, such as biometric authentication, face identification, and surveillance. Many state-of-the-art FV systems suffer from unequal performance across demographic groups, which is commonly overlooked by evaluation measures that do not assess population-specific performance. Deployed systems with bias may result in serious harm against individuals or groups who experience underperformance. We explore several fairness definitions and metrics, attempting to quantify bias in Google’s FaceNet model. In addition to statistical fairness metrics, we analyze clustered face embeddings produced by the FV model. We link well-clustered embeddings (well-defined, dense clusters) for a demographic group to biased model performance against that group. We present the intuition that FV systems underperform on protected demographic groups because they are less sensitive to differences between features within those groups, as evidenced by clustered embeddings. We show how this performance discrepancy results from a combination of representation and aggregation bias. death times for White face embeddings to later than other race groups ( p < 0.05 for W × A , W × I , and W × B t -tests), indicating that White embeddings are more in the embedding space. The other race groups have peak death times that are taller and earlier than the White race group. The shorter and wider peak for the White subgroup means that there is more variety (higher variance) in H 0 death times, rather than the consistent peak around 0.8 with less variance for other race groups. This shows that there is more variance for White face distribution in the embedding space compared to other race groups, a trend that was not present in the centroid distance distribution for race groups, which showed four bell-shaped density plots. Thus, our analysis of the ( H 0 ) death times supports previous findings that the White race group is clustered differently to other race groups. We note that there is less inequality in H 0 death times for female vs. male faces, despite our p -value indicating that this discrepancy may be significant ( p < 0.05).
机器学习模型执行人脸验证(FV)的各种高度重要的应用,如生物识别认证,人脸识别和监控。许多最先进的FV系统在不同人口群体中表现不平等,这一点通常被不评估特定人群表现的评估措施所忽视。部署有偏见的系统可能会对表现不佳的个人或群体造成严重伤害。我们探讨了几个公平的定义和指标,试图量化谷歌的FaceNet模型中的偏见。除了统计公平性指标外,我们还分析了由FV模型产生的聚类人脸嵌入。我们将人口统计群体的良好聚类嵌入(定义良好的密集聚类)与针对该群体的有偏差模型性能联系起来。我们提出的直觉是,FV系统在受保护的人口群体上表现不佳,因为它们对这些群体内部特征之间的差异不太敏感,聚类嵌入证明了这一点。我们展示了这种性能差异是如何由表示和聚集偏差共同造成的。白色面孔嵌入的死亡时间比其他种族组晚(W × A、W × I和W × B -t检验p < 0.05),说明白色面孔嵌入在嵌入空间中更多。其他种族群体的死亡高峰时间比白种人群体更高更早。白种人亚组的峰值更短更宽,这意味着h0死亡时间的变化更大(方差更大),而不是其他种族群体在0.8左右的一致峰值,方差更小。这表明,与其他种族相比,白人面孔分布在嵌入空间中的方差更大,这一趋势在种族群体的质心距离分布中没有出现,呈现出四个钟形密度图。因此,我们对(H 0)死亡时间的分析支持了之前的发现,即白种人群体与其他种族群体的聚集方式不同。我们注意到,尽管我们的p值表明这种差异可能是显著的(p < 0.05),但女性与男性面孔在H 0死亡时间上的不平等程度较小。
{"title":"Quantifying Bias in a Face Verification System","authors":"Megan Frisella, Pooya Khorrami, J. Matterer, K. Kratkiewicz, P. Torres-Carrasquillo","doi":"10.3390/cmsf2022003006","DOIUrl":"https://doi.org/10.3390/cmsf2022003006","url":null,"abstract":": Machine learning models perform face verification (FV) for a variety of highly consequential applications, such as biometric authentication, face identification, and surveillance. Many state-of-the-art FV systems suffer from unequal performance across demographic groups, which is commonly overlooked by evaluation measures that do not assess population-specific performance. Deployed systems with bias may result in serious harm against individuals or groups who experience underperformance. We explore several fairness definitions and metrics, attempting to quantify bias in Google’s FaceNet model. In addition to statistical fairness metrics, we analyze clustered face embeddings produced by the FV model. We link well-clustered embeddings (well-defined, dense clusters) for a demographic group to biased model performance against that group. We present the intuition that FV systems underperform on protected demographic groups because they are less sensitive to differences between features within those groups, as evidenced by clustered embeddings. We show how this performance discrepancy results from a combination of representation and aggregation bias. death times for White face embeddings to later than other race groups ( p < 0.05 for W × A , W × I , and W × B t -tests), indicating that White embeddings are more in the embedding space. The other race groups have peak death times that are taller and earlier than the White race group. The shorter and wider peak for the White subgroup means that there is more variety (higher variance) in H 0 death times, rather than the consistent peak around 0.8 with less variance for other race groups. This shows that there is more variance for White face distribution in the embedding space compared to other race groups, a trend that was not present in the centroid distance distribution for race groups, which showed four bell-shaped density plots. Thus, our analysis of the ( H 0 ) death times supports previous findings that the White race group is clustered differently to other race groups. We note that there is less inequality in H 0 death times for female vs. male faces, despite our p -value indicating that this discrepancy may be significant ( p < 0.05).","PeriodicalId":127261,"journal":{"name":"AAAI Workshop on Artificial Intelligence with Biased or Scarce Data (AIBSD)","volume":"9 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-04-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132036575","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
DAP-SDD: Distribution-Aware Pseudo Labeling for Small Defect Detection 小缺陷检测的分布感知伪标记
Pub Date : 2022-04-20 DOI: 10.3390/cmsf2022003005
Xiaoyan Zhuo, Wolfgang Rahfeldt, Xiaoqian Zhang, Ted Doros, S. Son
: Detecting defects, especially when they are small in the early manufacturing stages, is critical to achieving a high yield in industrial applications. While numerous modern deep learning models can improve detection performance, they become less effective in detecting small defects in practical applications due to the scarcity of labeled data and significant class imbalance in multiple dimensions. In this work, we propose a distribution-aware pseudo labeling method (DAP-SDD) to detect small defects accurately while using limited labeled data effectively. Specifically, we apply bootstrapping on limited labeled data and then utilize the approximated label distribution to guide pseudo label propagation. Moreover, we propose to use the t-distribution confidence interval for threshold setting to generate more pseudo labels with high confidence. DAP-SDD also incorporates data augmentation to enhance the model’s performance and robustness. We conduct extensive experiments on various datasets to validate the proposed method. Our evaluation results show that, overall, our proposed method requires less than 10% of labeled data to achieve comparable results of using a fully-labeled (100%) dataset and outperforms the state-of-the-art methods. For a dataset of wafer images, our proposed model can achieve above 0.93 of AP (average precision) with only four labeled images (i.e., 2% of labeled data).
在工业应用中,检测缺陷,特别是在早期制造阶段的小缺陷,是实现高产量的关键。虽然许多现代深度学习模型可以提高检测性能,但由于标记数据的稀缺性和多维度的显著类不平衡,它们在实际应用中检测小缺陷的效率较低。在这项工作中,我们提出了一种分布感知伪标记方法(DAP-SDD),在有效使用有限标记数据的情况下准确检测小缺陷。具体来说,我们在有限的标记数据上应用自举,然后利用近似的标签分布来指导伪标签传播。此外,我们建议使用t分布置信区间进行阈值设置,以生成更多具有高置信度的伪标签。DAP-SDD还结合了数据增强,以提高模型的性能和鲁棒性。我们在不同的数据集上进行了大量的实验来验证所提出的方法。我们的评估结果表明,总体而言,我们提出的方法只需要不到10%的标记数据就可以达到使用完全标记(100%)数据集的可比结果,并且优于最先进的方法。对于晶圆图像数据集,我们提出的模型仅使用四张标记图像(即标记数据的2%)就可以达到0.93 AP(平均精度)以上。
{"title":"DAP-SDD: Distribution-Aware Pseudo Labeling for Small Defect Detection","authors":"Xiaoyan Zhuo, Wolfgang Rahfeldt, Xiaoqian Zhang, Ted Doros, S. Son","doi":"10.3390/cmsf2022003005","DOIUrl":"https://doi.org/10.3390/cmsf2022003005","url":null,"abstract":": Detecting defects, especially when they are small in the early manufacturing stages, is critical to achieving a high yield in industrial applications. While numerous modern deep learning models can improve detection performance, they become less effective in detecting small defects in practical applications due to the scarcity of labeled data and significant class imbalance in multiple dimensions. In this work, we propose a distribution-aware pseudo labeling method (DAP-SDD) to detect small defects accurately while using limited labeled data effectively. Specifically, we apply bootstrapping on limited labeled data and then utilize the approximated label distribution to guide pseudo label propagation. Moreover, we propose to use the t-distribution confidence interval for threshold setting to generate more pseudo labels with high confidence. DAP-SDD also incorporates data augmentation to enhance the model’s performance and robustness. We conduct extensive experiments on various datasets to validate the proposed method. Our evaluation results show that, overall, our proposed method requires less than 10% of labeled data to achieve comparable results of using a fully-labeled (100%) dataset and outperforms the state-of-the-art methods. For a dataset of wafer images, our proposed model can achieve above 0.93 of AP (average precision) with only four labeled images (i.e., 2% of labeled data).","PeriodicalId":127261,"journal":{"name":"AAAI Workshop on Artificial Intelligence with Biased or Scarce Data (AIBSD)","volume":"28 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-04-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125236256","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
The Details Matter: Preventing Class Collapse in Supervised Contrastive Learning 细节至关重要:在监督对比学习中防止班级崩溃
Pub Date : 2022-04-15 DOI: 10.3390/cmsf2022003004
Daniel Y. Fu, Mayee F. Chen, Michael Zhang, K. Fatahalian, C. Ré
: Supervised contrastive learning optimizes a loss that pushes together embeddings of points from the same class while pulling apart embeddings of points from different classes. Class collapse—when every point from the same class has the same embedding—minimizes this loss but loses critical information that is not encoded in the class labels. For instance, the “cat” label does not capture unlabeled categories such as breeds, poses, or backgrounds (which we call “strata”). As a result, class collapse produces embeddings that are less useful for downstream applications such as transfer learning and achieves suboptimal generalization error when there are strata. We explore a simple modification to supervised contrastive loss that aims to prevent class collapse by uniformly pulling apart individual points from the same class. We seek to understand the effects of this loss by examining how it embeds strata of different sizes, finding that it clusters larger strata more tightly than smaller strata. As a result, our loss function produces embeddings that better distinguish strata in embedding space, which produces lift on three downstream applications: 4.4 points on coarse-to-fine transfer learning, 2.5 points on worst-group robustness, and 1.0 points on minimal coreset construction. Our loss also produces more accurate models, with up to 4.0 points of lift across 9 tasks.
:有监督的对比学习优化了一种损失,它将来自同一类的点的嵌入推到一起,同时将来自不同类的点的嵌入拉开。类崩溃(当来自同一类的每个点具有相同的嵌入时)将这种损失最小化,但会丢失未在类标签中编码的关键信息。例如,“猫”标签不能捕获未标记的类别,如品种、姿势或背景(我们称之为“分层”)。因此,类崩溃产生的嵌入对下游应用(如迁移学习)不太有用,并且在存在分层时产生次优泛化误差。我们探索了对监督对比损失的一种简单修改,旨在通过均匀地从同一类中分离单个点来防止类崩溃。我们试图通过研究它如何嵌入不同大小的地层来了解这种损失的影响,发现它比较小的地层更紧密地聚集较大的地层。因此,我们的损失函数产生的嵌入可以更好地区分嵌入空间中的地层,这在三个下游应用中产生提升:粗到细迁移学习4.4分,最差组鲁棒性2.5分,最小核心集构建1.0分。我们的损失也产生了更准确的模型,在9个任务中高达4.0点的升力。
{"title":"The Details Matter: Preventing Class Collapse in Supervised Contrastive Learning","authors":"Daniel Y. Fu, Mayee F. Chen, Michael Zhang, K. Fatahalian, C. Ré","doi":"10.3390/cmsf2022003004","DOIUrl":"https://doi.org/10.3390/cmsf2022003004","url":null,"abstract":": Supervised contrastive learning optimizes a loss that pushes together embeddings of points from the same class while pulling apart embeddings of points from different classes. Class collapse—when every point from the same class has the same embedding—minimizes this loss but loses critical information that is not encoded in the class labels. For instance, the “cat” label does not capture unlabeled categories such as breeds, poses, or backgrounds (which we call “strata”). As a result, class collapse produces embeddings that are less useful for downstream applications such as transfer learning and achieves suboptimal generalization error when there are strata. We explore a simple modification to supervised contrastive loss that aims to prevent class collapse by uniformly pulling apart individual points from the same class. We seek to understand the effects of this loss by examining how it embeds strata of different sizes, finding that it clusters larger strata more tightly than smaller strata. As a result, our loss function produces embeddings that better distinguish strata in embedding space, which produces lift on three downstream applications: 4.4 points on coarse-to-fine transfer learning, 2.5 points on worst-group robustness, and 1.0 points on minimal coreset construction. Our loss also produces more accurate models, with up to 4.0 points of lift across 9 tasks.","PeriodicalId":127261,"journal":{"name":"AAAI Workshop on Artificial Intelligence with Biased or Scarce Data (AIBSD)","volume":"252 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-04-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123750541","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Measuring Embedded Human-Like Biases in Face Recognition Models 测量人脸识别模型中嵌入的类人偏见
Pub Date : 2022-04-11 DOI: 10.3390/cmsf2022003002
Sangeun Lee, Soyoung Oh, Minji Kim, Eunil Park
: Recent works in machine learning have focused on understanding and mitigating bias in data and algorithms. Because the pre-trained models are trained on large real-world data, they are known to learn implicit biases in a way that humans unconsciously constructed for a long time. However, there has been little discussion about social biases with pre-trained face recognition models. Thus, this study investigates the robustness of the models against racial, gender, age, and an intersectional bias. We also present the racial bias with a different ethnicity other than white and black: Asian. In detail, we introduce the Face Embedding Association Test (FEAT) to measure the social biases in image vectors of faces with different race, gender, and age. It measures social bias in the face recognition models under the hypothesis that a specific group is more likely to be associated with a particular attribute in a biased manner. The presence of these biases within DeepFace, DeepID, VGGFace, FaceNet, OpenFace, and ArcFace critically mitigate the fairness in our society.
最近在机器学习方面的工作主要集中在理解和减轻数据和算法中的偏见。由于预先训练的模型是在大量现实世界数据上训练的,因此已知它们以一种人类长期无意识地构建的方式学习隐性偏见。然而,关于预先训练的人脸识别模型的社会偏见的讨论很少。因此,本研究考察了模型对种族、性别、年龄和交叉偏差的稳健性。我们还呈现了一个不同于白人和黑人的种族偏见:亚洲人。详细地,我们引入了人脸嵌入关联测试(FEAT)来测量不同种族、性别和年龄的人脸图像向量中的社会偏见。它测量了人脸识别模型中的社会偏见,假设一个特定的群体更有可能以一种偏见的方式与特定的属性联系在一起。这些偏见在DeepFace、DeepID、VGGFace、FaceNet、OpenFace和ArcFace中的存在严重削弱了我们社会的公平性。
{"title":"Measuring Embedded Human-Like Biases in Face Recognition Models","authors":"Sangeun Lee, Soyoung Oh, Minji Kim, Eunil Park","doi":"10.3390/cmsf2022003002","DOIUrl":"https://doi.org/10.3390/cmsf2022003002","url":null,"abstract":": Recent works in machine learning have focused on understanding and mitigating bias in data and algorithms. Because the pre-trained models are trained on large real-world data, they are known to learn implicit biases in a way that humans unconsciously constructed for a long time. However, there has been little discussion about social biases with pre-trained face recognition models. Thus, this study investigates the robustness of the models against racial, gender, age, and an intersectional bias. We also present the racial bias with a different ethnicity other than white and black: Asian. In detail, we introduce the Face Embedding Association Test (FEAT) to measure the social biases in image vectors of faces with different race, gender, and age. It measures social bias in the face recognition models under the hypothesis that a specific group is more likely to be associated with a particular attribute in a biased manner. The presence of these biases within DeepFace, DeepID, VGGFace, FaceNet, OpenFace, and ArcFace critically mitigate the fairness in our society.","PeriodicalId":127261,"journal":{"name":"AAAI Workshop on Artificial Intelligence with Biased or Scarce Data (AIBSD)","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-04-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121005011","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Measuring Gender Bias in Contextualized Embeddings 情境化嵌入中的性别偏见测量
Pub Date : 2022-04-11 DOI: 10.3390/cmsf2022003003
Styliani Katsarou, Borja Rodríguez-Gálvez, Jesse Shanahan
: Transformer models are now increasingly being used in real-world applications. Indiscrim-inately using these models as automated tools may propagate biases in ways we do not realize. To responsibly direct actions that will combat this problem, it is of crucial importance that we detect and quantify these biases. Robust methods have been developed to measure bias in non-contextualized embeddings. Nevertheless, these methods fail to apply to contextualized embeddings due to their mutable nature. Our study focuses on the detection and measurement of stereotypical biases associated with gender in the embeddings of T5 and mT5. We quantify bias by measuring the gender polarity of T5’s word embeddings for various professions. To measure gender polarity, we use a stable gender direction that we detect in the model’s embedding space. We also measure gender bias with respect to a specific downstream task and compare Swedish with English, as well as various sizes of the T5 model and its multilingual variant. The insights from our exploration indicate that the use of a stable gender direction, even in a Transformer’s mutable embedding space, can be a robust method to measure bias. We show that higher status professions are associated more with the male gender than the female gender. In addition, our method suggests that the Swedish language carries less bias associated with gender than English, and the higher manifestation of gender bias is associated with the use of larger language models.
:变压器模型现在越来越多地用于实际应用程序。不加选择地使用这些模型作为自动化工具可能会以我们没有意识到的方式传播偏见。为了负责任地指导应对这一问题的行动,我们发现并量化这些偏见至关重要。鲁棒的方法已经开发,以衡量偏差在非情境化嵌入。然而,由于其可变性,这些方法不能应用于上下文化嵌入。我们的研究重点是T5和mT5嵌入中与性别相关的刻板印象偏见的检测和测量。我们通过测量T5对不同职业的词嵌入的性别极性来量化偏见。为了测量性别极性,我们使用在模型嵌入空间中检测到的稳定性别方向。我们还测量了关于特定下游任务的性别偏见,并比较了瑞典语和英语,以及T5模型及其多语言变体的不同大小。从我们的探索中得出的见解表明,即使在Transformer的可变嵌入空间中,使用稳定的性别方向也可以是测量偏差的稳健方法。我们发现,地位较高的职业与男性的关系比与女性的关系更大。此外,我们的方法表明,瑞典语比英语带有更少的与性别相关的偏见,并且性别偏见的更高表现与使用更大的语言模型有关。
{"title":"Measuring Gender Bias in Contextualized Embeddings","authors":"Styliani Katsarou, Borja Rodríguez-Gálvez, Jesse Shanahan","doi":"10.3390/cmsf2022003003","DOIUrl":"https://doi.org/10.3390/cmsf2022003003","url":null,"abstract":": Transformer models are now increasingly being used in real-world applications. Indiscrim-inately using these models as automated tools may propagate biases in ways we do not realize. To responsibly direct actions that will combat this problem, it is of crucial importance that we detect and quantify these biases. Robust methods have been developed to measure bias in non-contextualized embeddings. Nevertheless, these methods fail to apply to contextualized embeddings due to their mutable nature. Our study focuses on the detection and measurement of stereotypical biases associated with gender in the embeddings of T5 and mT5. We quantify bias by measuring the gender polarity of T5’s word embeddings for various professions. To measure gender polarity, we use a stable gender direction that we detect in the model’s embedding space. We also measure gender bias with respect to a specific downstream task and compare Swedish with English, as well as various sizes of the T5 model and its multilingual variant. The insights from our exploration indicate that the use of a stable gender direction, even in a Transformer’s mutable embedding space, can be a robust method to measure bias. We show that higher status professions are associated more with the male gender than the female gender. In addition, our method suggests that the Swedish language carries less bias associated with gender than English, and the higher manifestation of gender bias is associated with the use of larger language models.","PeriodicalId":127261,"journal":{"name":"AAAI Workshop on Artificial Intelligence with Biased or Scarce Data (AIBSD)","volume":"1 4","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-04-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131539652","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
期刊
AAAI Workshop on Artificial Intelligence with Biased or Scarce Data (AIBSD)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1