High dimensional microarray datasets are difficult to classify since they have many features with small number ofinstances and imbalanced distribution of classes. This paper proposes a filter-based feature selection method to improvethe classification performance of microarray datasets by selecting the significant features. Combining the concepts ofrough sets, weighted rough set, fuzzy rough set and hesitant fuzzy sets for developing an effective algorithm is the maincontribution of this paper. The mentioned method has two steps, in the first step, four discretization approaches areapplied to discretize continuous datasets and selects a primary subset of features by combining of weighted rough setdependency degree and information gain via hesitant fuzzy aggregation approach. In the second step, a significancemeasure of features (defined by fuzzy rough concepts) is employed to remove redundant features from primary set.The Wilcoxon Signed Ranked tes (A Non-parametric statistical test) is conducted for comparing the presented methodwith ten feature selection methods across seven datasets. The results of experiments show that the proposed methodis able to select a significant subset of features and it is an effective method in the literature in terms of classificationperformance and simplicity.
由于高维微阵列数据集特征多、样本少、类别分布不平衡等特点,给分类带来困难。本文提出了一种基于滤波器的特征选择方法,通过选择显著特征来提高微阵列数据集的分类性能。结合粗糙集、加权粗糙集、模糊粗糙集和犹豫模糊集的概念,提出一种有效的算法是本文的主要贡献。该方法分为两步,第一步采用四种离散化方法对连续数据集进行离散化,并结合加权粗糙集依赖度和犹豫模糊聚集法的信息增益选择特征的主要子集;在第二步中,使用特征的显著性度量(由模糊粗糙概念定义)从原始集中去除冗余特征。进行了Wilcoxon Signed rank tes(一种非参数统计检验),将所提出的方法与七个数据集上的十种特征选择方法进行了比较。实验结果表明,所提出的方法能够选择出大量的特征子集,在分类性能和简单性方面是文献中有效的方法。
{"title":"A Hybrid Filter-Based Feature Selection Method via Hesitant Fuzzy and Rough Sets Concepts","authors":"Mohammad Mohtashami, M. Eftekhari","doi":"10.22111/ijfs.2018.4140","DOIUrl":"https://doi.org/10.22111/ijfs.2018.4140","url":null,"abstract":"High dimensional microarray datasets are difficult to classify since they have many features with small number ofinstances and imbalanced distribution of classes. This paper proposes a filter-based feature selection method to improvethe classification performance of microarray datasets by selecting the significant features. Combining the concepts ofrough sets, weighted rough set, fuzzy rough set and hesitant fuzzy sets for developing an effective algorithm is the maincontribution of this paper. The mentioned method has two steps, in the first step, four discretization approaches areapplied to discretize continuous datasets and selects a primary subset of features by combining of weighted rough setdependency degree and information gain via hesitant fuzzy aggregation approach. In the second step, a significancemeasure of features (defined by fuzzy rough concepts) is employed to remove redundant features from primary set.The Wilcoxon Signed Ranked tes (A Non-parametric statistical test) is conducted for comparing the presented methodwith ten feature selection methods across seven datasets. The results of experiments show that the proposed methodis able to select a significant subset of features and it is an effective method in the literature in terms of classificationperformance and simplicity.","PeriodicalId":212493,"journal":{"name":"How Fuzzy Concepts Contribute to Machine Learning","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-10-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131854976","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Fuzzy Decision Tree (FDT) classifiers combine decision trees with approximate reasoning offered by fuzzy representation to deal with language and measurement uncertainties. When a FDT induction algorithm utilizes stopping criteria for early stopping of the tree's growth, threshold values of stopping criteria will control the number of nodes. Finding a proper threshold value for a stopping criterion is one of the greatest challenges to be faced in FDT induction. In this paper, we propose a new method named Iterative Deepening Fuzzy ID3 (IDFID3) for FDT induction that has the ability of controlling the tree’s growth via dynamically setting the threshold value of stopping criterion in an iterative procedure. The final FDT induced by IDFID3 and the one obtained by common FID3 are the same when the numbers of nodes of induced FDTs are equal, but our main intention for introducing IDFID3 is the comparison of different stopping criteria through this algorithm. Therefore, a new stopping criterion named Normalized Maximum fuzzy information Gain multiplied by Number of Instances (NMGNI) is proposed and IDFID3 is used for comparing it against the other stopping criteria. Generally speaking, this paper presents a method to compare different stopping criteria independent of their threshold values utilizing IDFID3. The comparison results show that FDTs induced by the proposed stopping criterion in most situations are superior to the others and number of instances stopping criterion performs better than fuzzy information gain stopping criterion in terms of complexity (i.e. number of nodes) and classification accuracy. Also, both tree depth and fuzzy information gain stopping criteria, outperform fuzzy entropy, accuracy and number of instances in terms of mean depth of generated FDTs.
{"title":"Comparing Different Stopping Criteria for Fuzzy Decision Tree Induction Through IDFID3","authors":"M. Zeinalkhani, M. Eftekhari","doi":"10.22111/IJFS.2014.1394","DOIUrl":"https://doi.org/10.22111/IJFS.2014.1394","url":null,"abstract":"Fuzzy Decision Tree (FDT) classifiers combine decision trees with approximate reasoning offered by fuzzy representation to deal with language and measurement uncertainties. When a FDT induction algorithm utilizes stopping criteria for early stopping of the tree's growth, threshold values of stopping criteria will control the number of nodes. Finding a proper threshold value for a stopping criterion is one of the greatest challenges to be faced in FDT induction. In this paper, we propose a new method named Iterative Deepening Fuzzy ID3 (IDFID3) for FDT induction that has the ability of controlling the tree’s growth via dynamically setting the threshold value of stopping criterion in an iterative procedure. The final FDT induced by IDFID3 and the one obtained by common FID3 are the same when the numbers of nodes of induced FDTs are equal, but our main intention for introducing IDFID3 is the comparison of different stopping criteria through this algorithm. Therefore, a new stopping criterion named Normalized Maximum fuzzy information Gain multiplied by Number of Instances (NMGNI) is proposed and IDFID3 is used for comparing it against the other stopping criteria. Generally speaking, this paper presents a method to compare different stopping criteria independent of their threshold values utilizing IDFID3. The comparison results show that FDTs induced by the proposed stopping criterion in most situations are superior to the others and number of instances stopping criterion performs better than fuzzy information gain stopping criterion in terms of complexity (i.e. number of nodes) and classification accuracy. Also, both tree depth and fuzzy information gain stopping criteria, outperform fuzzy entropy, accuracy and number of instances in terms of mean depth of generated FDTs.","PeriodicalId":212493,"journal":{"name":"How Fuzzy Concepts Contribute to Machine Learning","volume":"31 ","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-02-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134128153","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 1900-01-01DOI: 10.1007/978-3-030-94066-9_3
M. Eftekhari, Adel Mehrpooya, Farid Saberi-Movahed, V. Torra
{"title":"Unsupervised Feature Selection Method Based on Sensitivity and Correlation Concepts for Multiclass Problems","authors":"M. Eftekhari, Adel Mehrpooya, Farid Saberi-Movahed, V. Torra","doi":"10.1007/978-3-030-94066-9_3","DOIUrl":"https://doi.org/10.1007/978-3-030-94066-9_3","url":null,"abstract":"","PeriodicalId":212493,"journal":{"name":"How Fuzzy Concepts Contribute to Machine Learning","volume":"36 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116009883","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 1900-01-01DOI: 10.1007/978-3-030-94066-9_8
M. Eftekhari, Adel Mehrpooya, Farid Saberi-Movahed, V. Torra
{"title":"Ensemble of Feature Selection Methods: A Hesitant Fuzzy Set Based Approach","authors":"M. Eftekhari, Adel Mehrpooya, Farid Saberi-Movahed, V. Torra","doi":"10.1007/978-3-030-94066-9_8","DOIUrl":"https://doi.org/10.1007/978-3-030-94066-9_8","url":null,"abstract":"","PeriodicalId":212493,"journal":{"name":"How Fuzzy Concepts Contribute to Machine Learning","volume":"40 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114899816","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}