A Comprehensive Review of Feature Selection and Feature Selection Stability in Machine Learning

IF 1 Q3 MULTIDISCIPLINARY SCIENCES gazi university journal of science Pub Date : 2022-09-04 DOI:10.35378/gujs.993763
Mustafa Büyükkeçeci̇, M. C. Okur
{"title":"A Comprehensive Review of Feature Selection and Feature Selection Stability in Machine Learning","authors":"Mustafa Büyükkeçeci̇, M. C. Okur","doi":"10.35378/gujs.993763","DOIUrl":null,"url":null,"abstract":"Feature selection is a data preprocessing method used to reduce the number of features in the datasets. Feature selection techniques search the entire feature space to find an optimal feature set that is free of redundant and irrelevant features. Reducing the dimensionality of a dataset by removing redundant and irrelevant features has a pivotal role in improving the performance, i.e., accuracy, of inductive learners and building simple models. Thus, feature selection is an imperative task of machine learning. The apparent need for feature selection raised considerable interest and became an important research topic in a wide range of fields, including bioinformatics, text classification, image recognition, and computer vision. As a result, a large pool of feature selection methods has been proposed, and a considerable amount of literature has been published on feature selection. The quality of feature selection algorithms is measured not only by the performance of the features they prefer but also by their stability. Therefore, this study focuses on two topics: feature selection and feature selection stability. In the pages that follow, general concepts and methods of feature selection are discussed and then an overview of feature selection stability and stability measures are given.","PeriodicalId":12615,"journal":{"name":"gazi university journal of science","volume":" ","pages":""},"PeriodicalIF":1.0000,"publicationDate":"2022-09-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"gazi university journal of science","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.35378/gujs.993763","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"MULTIDISCIPLINARY SCIENCES","Score":null,"Total":0}
引用次数: 0

Abstract

Feature selection is a data preprocessing method used to reduce the number of features in the datasets. Feature selection techniques search the entire feature space to find an optimal feature set that is free of redundant and irrelevant features. Reducing the dimensionality of a dataset by removing redundant and irrelevant features has a pivotal role in improving the performance, i.e., accuracy, of inductive learners and building simple models. Thus, feature selection is an imperative task of machine learning. The apparent need for feature selection raised considerable interest and became an important research topic in a wide range of fields, including bioinformatics, text classification, image recognition, and computer vision. As a result, a large pool of feature selection methods has been proposed, and a considerable amount of literature has been published on feature selection. The quality of feature selection algorithms is measured not only by the performance of the features they prefer but also by their stability. Therefore, this study focuses on two topics: feature selection and feature selection stability. In the pages that follow, general concepts and methods of feature selection are discussed and then an overview of feature selection stability and stability measures are given.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
机器学习中的特征选择和特征选择稳定性综述
特征选择是一种用于减少数据集中特征数量的数据预处理方法。特征选择技术通过搜索整个特征空间来找到一个没有冗余和不相关特征的最优特征集。通过去除冗余和不相关的特征来降低数据集的维数,对于提高归纳学习器的性能(即准确性)和构建简单模型具有关键作用。因此,特征选择是机器学习的一项重要任务。对特征选择的明显需求引起了相当大的兴趣,并成为包括生物信息学、文本分类、图像识别和计算机视觉在内的广泛领域的重要研究课题。因此,人们提出了大量的特征选择方法,并发表了大量关于特征选择的文献。特征选择算法的质量不仅取决于它们所选择的特征的性能,还取决于它们的稳定性。因此,本研究主要关注两个主题:特征选择和特征选择稳定性。在接下来的章节中,讨论了特征选择的一般概念和方法,然后概述了特征选择稳定性和稳定性措施。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
gazi university journal of science
gazi university journal of science MULTIDISCIPLINARY SCIENCES-
CiteScore
1.60
自引率
11.10%
发文量
87
期刊介绍: The scope of the “Gazi University Journal of Science” comprises such as original research on all aspects of basic science, engineering and technology. Original research results, scientific reviews and short communication notes in various fields of science and technology are considered for publication. The publication language of the journal is English. Manuscripts previously published in another journal are not accepted. Manuscripts with a suitable balance of practice and theory are preferred. A review article is expected to give in-depth information and satisfying evaluation of a specific scientific or technologic subject, supported with an extensive list of sources. Short communication notes prepared by researchers who would like to share the first outcomes of their on-going, original research work are welcome.
期刊最新文献
Experimental Investigation of a Small-Scale Parabolic Trough Concentrated Solar Power ‎Systems EVALUATION OF BUILDING ENERGY PERFORMANCE BASED ON SETTLEMENT SCALE DESIGN PARAMETERS USING BIM TOOLS Invariants of Immersions on n-Dimensional Affine Manifold MIL, SIL and PIL Implementation for Closed Loop Control of Flyback Converter Lean Six Sigma Studies in Türkiye: A Literature Review
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1