{"title":"高维域的增量贝叶斯网络结构学习","authors":"Amanullah Yasin, Philippe Leray","doi":"10.1109/ICMSAO.2013.6552635","DOIUrl":null,"url":null,"abstract":"The recent advances in hardware and software has led to development of applications generating a large amount of data in real-time. To keep abreast with latest trends, learning algorithms need to incorporate novel data continuously. One of the efficient ways is revising the existing knowledge so as to save time and memory. In this paper, we proposed an incremental algorithm for Bayesian network structure learning. It could deal with high dimensional domains, where whole dataset is not completely available, but grows continuously. Our algorithm learns local models by limiting search space and performs a constrained greedy hill-climbing search to obtain a global model. We evaluated our method on different datasets having several hundreds of variables, in terms of performance and accuracy. The empirical evaluation shows that our method is significantly better than existing state of the art methods and justifies its effectiveness for incremental use.","PeriodicalId":339666,"journal":{"name":"2013 5th International Conference on Modeling, Simulation and Applied Optimization (ICMSAO)","volume":"11 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2013-04-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"4","resultStr":"{\"title\":\"Incremental Bayesian network structure learning in high dimensional domains\",\"authors\":\"Amanullah Yasin, Philippe Leray\",\"doi\":\"10.1109/ICMSAO.2013.6552635\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"The recent advances in hardware and software has led to development of applications generating a large amount of data in real-time. To keep abreast with latest trends, learning algorithms need to incorporate novel data continuously. One of the efficient ways is revising the existing knowledge so as to save time and memory. In this paper, we proposed an incremental algorithm for Bayesian network structure learning. It could deal with high dimensional domains, where whole dataset is not completely available, but grows continuously. Our algorithm learns local models by limiting search space and performs a constrained greedy hill-climbing search to obtain a global model. We evaluated our method on different datasets having several hundreds of variables, in terms of performance and accuracy. The empirical evaluation shows that our method is significantly better than existing state of the art methods and justifies its effectiveness for incremental use.\",\"PeriodicalId\":339666,\"journal\":{\"name\":\"2013 5th International Conference on Modeling, Simulation and Applied Optimization (ICMSAO)\",\"volume\":\"11 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2013-04-28\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"4\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2013 5th International Conference on Modeling, Simulation and Applied Optimization (ICMSAO)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICMSAO.2013.6552635\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2013 5th International Conference on Modeling, Simulation and Applied Optimization (ICMSAO)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICMSAO.2013.6552635","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Incremental Bayesian network structure learning in high dimensional domains
The recent advances in hardware and software has led to development of applications generating a large amount of data in real-time. To keep abreast with latest trends, learning algorithms need to incorporate novel data continuously. One of the efficient ways is revising the existing knowledge so as to save time and memory. In this paper, we proposed an incremental algorithm for Bayesian network structure learning. It could deal with high dimensional domains, where whole dataset is not completely available, but grows continuously. Our algorithm learns local models by limiting search space and performs a constrained greedy hill-climbing search to obtain a global model. We evaluated our method on different datasets having several hundreds of variables, in terms of performance and accuracy. The empirical evaluation shows that our method is significantly better than existing state of the art methods and justifies its effectiveness for incremental use.