Unseen family member classification using mixture of experts

M. Ghahramani, H. L. Wang, W. Yau, E. Teoh
{"title":"Unseen family member classification using mixture of experts","authors":"M. Ghahramani, H. L. Wang, W. Yau, E. Teoh","doi":"10.1109/ICIEA.2010.5516872","DOIUrl":null,"url":null,"abstract":"All family members resemble each other in different ways which is recognizable by our brain. In this paper, we have developed family classification using AdaBoost, Support Vector Machines and K-Nearest Neighbor classifiers with different patches of training data. In some cases family classification involve unseen data classification in which the classifiers' performance drop significantly. Therefore Mixture of Experts is conducted to improve their performance. To have a fair comparison of mentioned approaches 3 different families from 3 different ethnic groups are used. Experimental results show that we can achieve an average accuracy rate of 76 percent and up to 27 percent accuracy improvement by using majority voting of mixture of experts depending on the family data.","PeriodicalId":234296,"journal":{"name":"2010 5th IEEE Conference on Industrial Electronics and Applications","volume":"20 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2010-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"5","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2010 5th IEEE Conference on Industrial Electronics and Applications","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICIEA.2010.5516872","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 5

Abstract

All family members resemble each other in different ways which is recognizable by our brain. In this paper, we have developed family classification using AdaBoost, Support Vector Machines and K-Nearest Neighbor classifiers with different patches of training data. In some cases family classification involve unseen data classification in which the classifiers' performance drop significantly. Therefore Mixture of Experts is conducted to improve their performance. To have a fair comparison of mentioned approaches 3 different families from 3 different ethnic groups are used. Experimental results show that we can achieve an average accuracy rate of 76 percent and up to 27 percent accuracy improvement by using majority voting of mixture of experts depending on the family data.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
看不见的家庭成员分类使用混合专家
所有的家庭成员都以不同的方式彼此相似,这是我们的大脑可以识别的。在本文中,我们使用AdaBoost,支持向量机和k -最近邻分类器对不同patch的训练数据进行了家庭分类。在某些情况下,家族分类涉及不可见数据分类,分类器的性能明显下降。因此,对专家进行混合,以提高他们的表现。为了对上述方法进行公平比较,使用了来自3个不同民族的3个不同家庭。实验结果表明,采用基于家庭数据的专家混合多数投票方法,平均准确率可达76%,准确率可提高27%。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Forecasting next-day electricity prices with Hidden Markov Models Design of HTS Linear Induction Motor using GA and the Finite Element Method Hybrid recurrent fuzzy neural network control for permanent magnet synchronous motor applied in electric scooter Integrating human factors into nanotech sustainability assessment and communication An ID-based content extraction signatures without trusted party
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1