首页 > 最新文献

2006 IEEE Conference on Cybernetics and Intelligent Systems最新文献

英文 中文
An Efficient MPC Algorithm based on a Priori Zone Control 基于先验区域控制的高效MPC算法
Pub Date : 2006-06-07 DOI: 10.1109/ICCIS.2006.252347
P. Park, S. Kim, J. Moon, M. Shin
This paper presents an efficient MPC algorithm for uncertain time-varying systems with input constraints. The proposed algorithm adopts the method of increasing free control horizon in the dual mode (i.e., a free control mode in the first finite horizon and a state-feedback mode in the following infinite horizon) paradigm so as to enlarge the set of stabilizable initial states. In the method, however, since the number of LMIs growing exponentially with the free control horizon makes the corresponding optimization problems intractable even for small horizon, it is impracticable to blindly increase the free control horizon. The objective of this paper is to relax the restriction on increase of the free control horizon, incurred on computational burdens in the method. By choosing a combination of hyper-boxes including a possible region of the initial states and then by designing a priori zone controller for each hyper-box so as to send any initial states in the hyper-box into the invariant ellipsoidal target set, the algorithm can dramatically reduce the on-line computational burden for enlarging the set of stabilizable initial states
针对具有输入约束的不确定时变系统,提出了一种高效的MPC算法。该算法采用双模式(即第一个有限视界的自由控制模式和下一个无限视界的状态反馈模式)范式中增加自由控制视界的方法,以扩大可稳定初始状态集。然而,由于lmi的数量随自由控制水平呈指数增长,使得即使在很小的水平下,相应的优化问题也难以解决,因此盲目地增加自由控制水平是不可行的。本文的目的是放宽该方法对自由控制水平的增加所带来的计算负担的限制。该算法通过选择包含可能初始状态区域的超盒组合,然后为每个超盒设计一个先验区域控制器,将超盒中的任意初始状态发送到不变椭球目标集中,从而大大减少了扩大可稳定初始状态集的在线计算量
{"title":"An Efficient MPC Algorithm based on a Priori Zone Control","authors":"P. Park, S. Kim, J. Moon, M. Shin","doi":"10.1109/ICCIS.2006.252347","DOIUrl":"https://doi.org/10.1109/ICCIS.2006.252347","url":null,"abstract":"This paper presents an efficient MPC algorithm for uncertain time-varying systems with input constraints. The proposed algorithm adopts the method of increasing free control horizon in the dual mode (i.e., a free control mode in the first finite horizon and a state-feedback mode in the following infinite horizon) paradigm so as to enlarge the set of stabilizable initial states. In the method, however, since the number of LMIs growing exponentially with the free control horizon makes the corresponding optimization problems intractable even for small horizon, it is impracticable to blindly increase the free control horizon. The objective of this paper is to relax the restriction on increase of the free control horizon, incurred on computational burdens in the method. By choosing a combination of hyper-boxes including a possible region of the initial states and then by designing a priori zone controller for each hyper-box so as to send any initial states in the hyper-box into the invariant ellipsoidal target set, the algorithm can dramatically reduce the on-line computational burden for enlarging the set of stabilizable initial states","PeriodicalId":296028,"journal":{"name":"2006 IEEE Conference on Cybernetics and Intelligent Systems","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133740552","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Mean Values of Fuzzy Numbers and the Measurement of Fuzziness by Evaluation Measures 模糊数的均值与评价指标对模糊性的度量
Pub Date : 2006-06-07 DOI: 10.1109/ICCIS.2006.252245
Y. Yoshida
In this paper, we discuss an evaluation method of fuzzy numbers as mean values and measurement of fuzziness defined by fuzzy measures, and the presented method is applicable to fuzzy numbers and fuzzy stochastic process defined by fuzzy numbers/fuzzy random variables in decision making. We compare the measurement of fuzziness and the variance as a factor to measure uncertainty. Formulae are also given to apply the results to triangle-type fuzzy numbers and trapezoidal-type fuzzy numbers
本文讨论了模糊数作为均值的评价方法和模糊测度定义的模糊性度量方法,该方法适用于决策中的模糊数和模糊数/模糊随机变量定义的模糊随机过程。我们比较了模糊度量和方差作为度量不确定性的一个因素。并给出了将结果应用于三角形模糊数和梯形模糊数的公式
{"title":"Mean Values of Fuzzy Numbers and the Measurement of Fuzziness by Evaluation Measures","authors":"Y. Yoshida","doi":"10.1109/ICCIS.2006.252245","DOIUrl":"https://doi.org/10.1109/ICCIS.2006.252245","url":null,"abstract":"In this paper, we discuss an evaluation method of fuzzy numbers as mean values and measurement of fuzziness defined by fuzzy measures, and the presented method is applicable to fuzzy numbers and fuzzy stochastic process defined by fuzzy numbers/fuzzy random variables in decision making. We compare the measurement of fuzziness and the variance as a factor to measure uncertainty. Formulae are also given to apply the results to triangle-type fuzzy numbers and trapezoidal-type fuzzy numbers","PeriodicalId":296028,"journal":{"name":"2006 IEEE Conference on Cybernetics and Intelligent Systems","volume":"60 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131395221","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
OWL/XDD: A Formal Language for Application Profiles OWL/XDD:一种用于应用程序概要的形式化语言
Pub Date : 2006-06-07 DOI: 10.1109/ICCIS.2006.252313
Photchanan Ratanajaipan, E. Nantajeewarawat, V. Wuwongse
An application profile specifies a set of terms, drawn from one or more standard namespaces, for annotation of data, and constrains their usage and interpretations in a particular local application. An approach to defining an application profile using the OWL and OWL/XDD languages is proposed - the former is a standard Web ontology language and the latter is a definite-clause-style knowledge representation language that uses XML expressions as their underlying data structure. Constraints are defined in terms of rules, which are represented as XDD clauses. As an illustration, application of the approach to defining Dublin core metadata initiative's library application profile (DC-Lib), along with the possibility of extending it by describing finer-grained semantic constraints, is demonstrated. A prototype catalog validation system has been implemented, and some experimental results are shown
应用程序概要文件指定一组术语,从一个或多个标准名称空间中提取,用于数据注释,并限制它们在特定本地应用程序中的使用和解释。提出了一种使用OWL和OWL/XDD语言定义应用程序概要文件的方法——前者是标准的Web本体语言,后者是使用XML表达式作为其底层数据结构的定义子句样式的知识表示语言。约束是根据规则定义的,这些规则表示为XDD子句。作为示例,演示了定义Dublin核心元数据计划的图书馆应用程序概要文件(DC-Lib)的方法的应用,以及通过描述细粒度语义约束对其进行扩展的可能性。实现了一个原型目录验证系统,并给出了一些实验结果
{"title":"OWL/XDD: A Formal Language for Application Profiles","authors":"Photchanan Ratanajaipan, E. Nantajeewarawat, V. Wuwongse","doi":"10.1109/ICCIS.2006.252313","DOIUrl":"https://doi.org/10.1109/ICCIS.2006.252313","url":null,"abstract":"An application profile specifies a set of terms, drawn from one or more standard namespaces, for annotation of data, and constrains their usage and interpretations in a particular local application. An approach to defining an application profile using the OWL and OWL/XDD languages is proposed - the former is a standard Web ontology language and the latter is a definite-clause-style knowledge representation language that uses XML expressions as their underlying data structure. Constraints are defined in terms of rules, which are represented as XDD clauses. As an illustration, application of the approach to defining Dublin core metadata initiative's library application profile (DC-Lib), along with the possibility of extending it by describing finer-grained semantic constraints, is demonstrated. A prototype catalog validation system has been implemented, and some experimental results are shown","PeriodicalId":296028,"journal":{"name":"2006 IEEE Conference on Cybernetics and Intelligent Systems","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129824083","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Adaptive Language Independent Spell Checking using Intelligent Traverse on a Tree 在树上使用智能遍历自适应语言独立拼写检查
Pub Date : 2006-06-07 DOI: 10.1109/ICCIS.2006.252325
Behrang Q. Zadeh, A. Ilkhani, A. Ganjeii
This paper introduces an adaptive, language independent, and 'built-in error pattern free' spell checker. Proposed system suggests proper form of misspelled words using non deterministic traverse of 'ternary search tree' data structure. In other words the problem of spell checking is addressed by traverse a tree with variable weighted edges. The proposed system uses interaction with user to learn error pattern of media. In this way, system improves its suggestions as time goes by
本文介绍了一种自适应的、独立于语言的、“内置无错误模式”的拼写检查器。提出的系统建议使用“三元搜索树”数据结构的非确定性遍历正确形式的拼写错误的单词。换句话说,拼写检查问题是通过遍历具有可变加权边的树来解决的。该系统通过与用户的交互来学习媒体的错误模式。通过这种方式,系统会随着时间的推移而改进其建议
{"title":"Adaptive Language Independent Spell Checking using Intelligent Traverse on a Tree","authors":"Behrang Q. Zadeh, A. Ilkhani, A. Ganjeii","doi":"10.1109/ICCIS.2006.252325","DOIUrl":"https://doi.org/10.1109/ICCIS.2006.252325","url":null,"abstract":"This paper introduces an adaptive, language independent, and 'built-in error pattern free' spell checker. Proposed system suggests proper form of misspelled words using non deterministic traverse of 'ternary search tree' data structure. In other words the problem of spell checking is addressed by traverse a tree with variable weighted edges. The proposed system uses interaction with user to learn error pattern of media. In this way, system improves its suggestions as time goes by","PeriodicalId":296028,"journal":{"name":"2006 IEEE Conference on Cybernetics and Intelligent Systems","volume":"25 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134634729","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 11
A Forecasting Model of Dynamic Grey Rough Set and its Application on Stock Selection 动态灰色粗糙集预测模型及其在选股中的应用
Pub Date : 2006-06-07 DOI: 10.1109/ICCIS.2006.252320
Ting-Cheng Chang, Chuen-Jiuan Jane, Yuan-Paio Lee
The main purpose of paper is to establish a system, which combines rough set and grey theory. This model is used to let the time-serial, season-serial or regular data have the dynamic trend concepts by grey prediction, then, select the data sets with trend value through rough set screening system. It mainly is applied for a portfolio prediction in the stock market. Our study first predicts each listed company's attributes of condition and decision-making by grey prediction, secondly groups their attributes by K-means grouping tools, then filters and categorizes the groups with the classified capacity of rough set for uncertain and non-sufficient information and selects the stock portfolio. And then we evaluate the company shares from the portfolio according to their past EPS and ROE and elect the better ones again. Finally, the selected companies are arranged in order with grey relation and determine the weight of each share in the portfolio according to it. The experimental result in Taiwan: during five years (2000-2004), the average annual rate of return was 38.1%. The portfolio determined by the model overran the market dramatically
本文的主要目的是建立一个将粗糙集理论与灰色理论相结合的系统。该模型通过灰色预测让时间序列、季节序列或规律数据具有动态趋势概念,然后通过粗集筛选系统选择具有趋势值的数据集。它主要应用于股票市场的投资组合预测。本文首先利用灰色预测方法对各上市公司的条件属性和决策属性进行预测,然后利用K-means分组工具对各上市公司的属性进行分组,然后利用粗糙集对不确定信息和不充分信息的分类能力对分组进行过滤和分类,选择股票组合。然后我们根据公司过去的每股收益和净资产收益率对投资组合中的公司股票进行评估,并再次选出较好的股票。最后,将选取的公司按灰色关联排序,并据此确定投资组合中各股份的权重。台湾地区的实验结果:在2000-2004年的五年间,平均年收益率为38.1%。这个模型决定的投资组合大大超过了市场
{"title":"A Forecasting Model of Dynamic Grey Rough Set and its Application on Stock Selection","authors":"Ting-Cheng Chang, Chuen-Jiuan Jane, Yuan-Paio Lee","doi":"10.1109/ICCIS.2006.252320","DOIUrl":"https://doi.org/10.1109/ICCIS.2006.252320","url":null,"abstract":"The main purpose of paper is to establish a system, which combines rough set and grey theory. This model is used to let the time-serial, season-serial or regular data have the dynamic trend concepts by grey prediction, then, select the data sets with trend value through rough set screening system. It mainly is applied for a portfolio prediction in the stock market. Our study first predicts each listed company's attributes of condition and decision-making by grey prediction, secondly groups their attributes by K-means grouping tools, then filters and categorizes the groups with the classified capacity of rough set for uncertain and non-sufficient information and selects the stock portfolio. And then we evaluate the company shares from the portfolio according to their past EPS and ROE and elect the better ones again. Finally, the selected companies are arranged in order with grey relation and determine the weight of each share in the portfolio according to it. The experimental result in Taiwan: during five years (2000-2004), the average annual rate of return was 38.1%. The portfolio determined by the model overran the market dramatically","PeriodicalId":296028,"journal":{"name":"2006 IEEE Conference on Cybernetics and Intelligent Systems","volume":"48 11","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114060099","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Linguistic Knowledge Extraction from Neural Networks Using Maximum Weight and Frequency Data Representation 基于最大权值和频率数据表示的神经网络语言知识提取
Pub Date : 2006-06-07 DOI: 10.1109/ICCIS.2006.252314
W. Wettayaprasit, U. Sangket
This paper presents a method of linguistic rule extraction from neural networks nodes pruning using frequency interval data representation. The method composes of two steps which are 1) neural networks nodes pruning by analysis on the maximum weight and 2) linguistic rule extraction using frequency interval data representation. The study has tested with the benchmark data sets such as heart disease, Wisconsin breast cancer, Pima Indians diabetes, and electrocardiography data set of heart disease patients from hospitals in Thailand. The study found that the linguistic rules received had high accuracy and easy to understand. The number of rules and the number of conjunction of conditions were small and the training time was also decreased
提出了一种基于频率区间数据表示的神经网络节点剪枝语言规则提取方法。该方法由两个步骤组成:1)通过最大权值分析对神经网络节点进行剪枝;2)使用频率区间数据表示进行语言规则提取。本研究以心脏病、威斯康星乳腺癌、皮马印第安人糖尿病等基准数据集和泰国医院心脏病患者的心电图数据集进行了测试。研究发现,接收到的语言规则准确率高,易于理解。规则数和条件连接数少,训练时间也大大缩短
{"title":"Linguistic Knowledge Extraction from Neural Networks Using Maximum Weight and Frequency Data Representation","authors":"W. Wettayaprasit, U. Sangket","doi":"10.1109/ICCIS.2006.252314","DOIUrl":"https://doi.org/10.1109/ICCIS.2006.252314","url":null,"abstract":"This paper presents a method of linguistic rule extraction from neural networks nodes pruning using frequency interval data representation. The method composes of two steps which are 1) neural networks nodes pruning by analysis on the maximum weight and 2) linguistic rule extraction using frequency interval data representation. The study has tested with the benchmark data sets such as heart disease, Wisconsin breast cancer, Pima Indians diabetes, and electrocardiography data set of heart disease patients from hospitals in Thailand. The study found that the linguistic rules received had high accuracy and easy to understand. The number of rules and the number of conjunction of conditions were small and the training time was also decreased","PeriodicalId":296028,"journal":{"name":"2006 IEEE Conference on Cybernetics and Intelligent Systems","volume":"65 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124850371","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Particle Swarm Assisted Incremental Evolution Strategy for Function Optimization 粒子群辅助的函数优化增量进化策略
Pub Date : 2006-06-07 DOI: 10.1109/ICCIS.2006.252276
W. Mo, S. Guan
This paper presents a new evolutionary approach for function optimization problems particle swarm assisted incremental evolution strategy (PIES). Two strategies are proposed. One is incremental optimization that the whole evolution consists of several phases and one more variable is focused in each phase. The number of phases is equal to the number of variables in maximum. Each phase is composed of two stages: in the single-variable evolution (SVE) stage, a population is evolved with respect to one independent variable in a series of cutting planes; in the multi-variable evolving (MVE) stage, the initial population is formed by integrating the population obtained by the SVE in current phase and by the MVE in the last phase. And then the MVE is taken on the incremented variable set. The second strategy is a hybrid of particle swarm optimization (PSO) and the evolution strategy (ES). PSO is applied to adjust the cutting planes (in SVEs) or hyper-planes (in MVEs) while ES is applied to searching optima in the cutting planes/hyper-planes. The results of experiments show that PIES generally outperforms three other evolutionary algorithms, improved normal GA, PSO and SADEXERAF, in the sense that PIES finds solutions with more optimal objective values and closer to the true optima
提出了一种新的求解函数优化问题的进化方法——粒子群辅助增量进化策略。提出了两种策略。一种是增量优化,即整个进化过程分为几个阶段,每个阶段多关注一个变量。相位的数量等于最大值中变量的数量。每个阶段由两个阶段组成:在单变量进化(SVE)阶段,种群相对于一系列切割平面中的一个自变量进化;在多变量进化(MVE)阶段,初始种群是由当前阶段的SVE和上一阶段的MVE得到的种群积分形成的。然后对增量变量集取最小方差。第二种策略是混合粒子群优化(PSO)和进化策略(ES)。粒子群算法用于调整切割平面(sve)或超平面(mve),粒子群算法用于搜索切割平面/超平面中的最优点。实验结果表明,PIES总体上优于其他三种进化算法,即改进的普通遗传算法、PSO算法和SADEXERAF算法,因为PIES找到的解具有更优的目标值,并且更接近真实最优值
{"title":"Particle Swarm Assisted Incremental Evolution Strategy for Function Optimization","authors":"W. Mo, S. Guan","doi":"10.1109/ICCIS.2006.252276","DOIUrl":"https://doi.org/10.1109/ICCIS.2006.252276","url":null,"abstract":"This paper presents a new evolutionary approach for function optimization problems particle swarm assisted incremental evolution strategy (PIES). Two strategies are proposed. One is incremental optimization that the whole evolution consists of several phases and one more variable is focused in each phase. The number of phases is equal to the number of variables in maximum. Each phase is composed of two stages: in the single-variable evolution (SVE) stage, a population is evolved with respect to one independent variable in a series of cutting planes; in the multi-variable evolving (MVE) stage, the initial population is formed by integrating the population obtained by the SVE in current phase and by the MVE in the last phase. And then the MVE is taken on the incremented variable set. The second strategy is a hybrid of particle swarm optimization (PSO) and the evolution strategy (ES). PSO is applied to adjust the cutting planes (in SVEs) or hyper-planes (in MVEs) while ES is applied to searching optima in the cutting planes/hyper-planes. The results of experiments show that PIES generally outperforms three other evolutionary algorithms, improved normal GA, PSO and SADEXERAF, in the sense that PIES finds solutions with more optimal objective values and closer to the true optima","PeriodicalId":296028,"journal":{"name":"2006 IEEE Conference on Cybernetics and Intelligent Systems","volume":"34 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125145324","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
A Robust Algorithm for Classification Using Decision Trees 基于决策树的鲁棒分类算法
Pub Date : 2006-06-07 DOI: 10.1109/ICCIS.2006.252336
B. Chandra, P. Paul V
Decision trees algorithms have been suggested in the past for classification of numeric as well as categorical attributes. SLIQ algorithm was proposed (Mehta et al., 1996) as an improvement over ID3 and C4.5 algorithms (Quinlan, 1993). Elegant Decision Tree Algorithm was proposed (Chandra et al. 2002) to improve the performance of SLIQ. In this paper a novel approach has been presented for the choice of split value of attributes. The issue of reducing the number of split points has been addressed. It has been shown on various datasets taken from UCI machine learning data repository that this approach gives better classification accuracy as compared to C4.5, SLIQ and Elegant Decision Tree Algorithm (EDTA) and at the same time the number of split points to be evaluated is much less compared to that of SLIQ and EDTA
决策树算法在过去被建议用于数字和分类属性的分类。作为对ID3和C4.5算法的改进(Quinlan, 1993),提出了SLIQ算法(Mehta et al., 1996)。为了提高SLIQ的性能,提出了优雅决策树算法(Chandra et al. 2002)。本文提出了一种新的属性分割值选择方法。减少分界点数目的问题已得到解决。从UCI机器学习数据库中获取的各种数据集表明,与C4.5、SLIQ和优雅决策树算法(EDTA)相比,这种方法具有更好的分类精度,同时与SLIQ和EDTA相比,需要评估的分裂点数量要少得多
{"title":"A Robust Algorithm for Classification Using Decision Trees","authors":"B. Chandra, P. Paul V","doi":"10.1109/ICCIS.2006.252336","DOIUrl":"https://doi.org/10.1109/ICCIS.2006.252336","url":null,"abstract":"Decision trees algorithms have been suggested in the past for classification of numeric as well as categorical attributes. SLIQ algorithm was proposed (Mehta et al., 1996) as an improvement over ID3 and C4.5 algorithms (Quinlan, 1993). Elegant Decision Tree Algorithm was proposed (Chandra et al. 2002) to improve the performance of SLIQ. In this paper a novel approach has been presented for the choice of split value of attributes. The issue of reducing the number of split points has been addressed. It has been shown on various datasets taken from UCI machine learning data repository that this approach gives better classification accuracy as compared to C4.5, SLIQ and Elegant Decision Tree Algorithm (EDTA) and at the same time the number of split points to be evaluated is much less compared to that of SLIQ and EDTA","PeriodicalId":296028,"journal":{"name":"2006 IEEE Conference on Cybernetics and Intelligent Systems","volume":"12 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125810324","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
Multi-criteria Intelligent Dispatching Control of Automated Guided Vehicles in FMS FMS中自动引导车辆的多准则智能调度控制
Pub Date : 2006-06-07 DOI: 10.1109/ICCIS.2006.252292
N. Umashankar, V. Karthik
In flexible manufacturing systems (FMS), automated guided vehicles (AGVs) are used for transportation of the processed materials between various pickup and delivery points. The assignment of an AGV to a workcentre from a set of workcentres simultaneously requesting the service for transport of a part is often solved in real-time with simple dispatching rules. This paper proposes an intelligent dispatching approach for the AGVs based on multi-criteria fuzzy logic controller, which simultaneously takes into account multiple aspects in every dispatching decision. The controller operates in two stages in which the second stage is constructed as a conflict resolving tool between two equally ranked AGVs for a particular workcentre. The control system is being implemented using MATLAB and its fuzzy inference engine. Sample runs have been provided to illustrate the controller implementation
在柔性制造系统(FMS)中,自动导引车(agv)用于在各个取货点和交货点之间运输加工过的材料。将AGV从同时请求零件运输服务的一组工作中心分配到一个工作中心,通常使用简单的调度规则实时解决。提出了一种基于多准则模糊控制器的agv智能调度方法,该方法在每次调度决策中同时考虑多个方面。控制器分两个阶段运行,其中第二阶段被构建为特定工作中心的两个同等排名的agv之间的冲突解决工具。利用MATLAB及其模糊推理引擎实现了该控制系统。提供了示例运行来说明控制器的实现
{"title":"Multi-criteria Intelligent Dispatching Control of Automated Guided Vehicles in FMS","authors":"N. Umashankar, V. Karthik","doi":"10.1109/ICCIS.2006.252292","DOIUrl":"https://doi.org/10.1109/ICCIS.2006.252292","url":null,"abstract":"In flexible manufacturing systems (FMS), automated guided vehicles (AGVs) are used for transportation of the processed materials between various pickup and delivery points. The assignment of an AGV to a workcentre from a set of workcentres simultaneously requesting the service for transport of a part is often solved in real-time with simple dispatching rules. This paper proposes an intelligent dispatching approach for the AGVs based on multi-criteria fuzzy logic controller, which simultaneously takes into account multiple aspects in every dispatching decision. The controller operates in two stages in which the second stage is constructed as a conflict resolving tool between two equally ranked AGVs for a particular workcentre. The control system is being implemented using MATLAB and its fuzzy inference engine. Sample runs have been provided to illustrate the controller implementation","PeriodicalId":296028,"journal":{"name":"2006 IEEE Conference on Cybernetics and Intelligent Systems","volume":"363 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126808497","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Pre-Eliminating Features for Fast Training in Real Time Object Detection in Images with a Novel Variant of AdaBoost AdaBoost的一种新变体,用于图像中实时目标检测的快速训练的预消除功能
Pub Date : 2006-06-07 DOI: 10.1109/ICCIS.2006.252285
M. Stojmenovic
Our primary interest is to build fast and reliable object recognizers in images based on small training sets. This is important in cases where the training set needs to be built mostly manually, as in the case that we studied, the recognition of the Honda Accord 2004 from rear views. We described a novel variant of the AdaBoost based learning algorithm, which builds a strong classifier by incremental addition of weak classifiers (WCs) that minimize the combined error of the already selected WCs. Each WC is trained only once, and examples do not change their weights. We proposed to pre-eliminate features whose cumulative error of corresponding best WCs exceeds a predetermined threshold value. We tested two straightforward definitions of cumulative error. In both cases, we showed that, when over 97% of the initial features are eliminated at the very beginning from further training, training time is drastically reduced while having little impact on the quality of the pool of available WCs. This is a novel method that has reduced the training set WC quantity to less than 3% of its original number, greatly speeding up training time, and showing no negative impact on the quality of the final classifier. Our experiments indicated that the set of features used by Viola and Jones and others for face recognition was inefficient for our problem; therefore, each object requires its own custom-made set of features for real time and accurate recognition. Our training method, combined with the selection of appropriate features, has resulted in finding a very accurate classifier containing merely 30 weak classifiers. Compared to existing literature, we have overall achieved the design of a real time object detection machine with the least number of examples, the least number of weak classifiers, the fastest training time, and with competitive detection and false positive rates
我们的主要兴趣是在基于小训练集的图像中构建快速可靠的目标识别器。这在训练集主要需要手工构建的情况下很重要,就像我们研究的例子一样,从后视图识别本田雅阁2004。我们描述了一种基于AdaBoost的学习算法的新变体,该算法通过增量添加弱分类器(WCs)来构建强分类器,从而最小化已经选择的WCs的组合误差。每个WC只训练一次,并且示例不改变其权重。我们提出对相应最佳wc的累积误差超过预定阈值的特征进行预剔除。我们测试了累积误差的两个简单定义。在这两种情况下,我们都表明,当超过97%的初始特征在一开始就从进一步的训练中消除时,训练时间大大减少,而对可用wc池的质量几乎没有影响。这是一种新颖的方法,将训练集WC数量减少到原来的3%以下,大大加快了训练时间,并且对最终分类器的质量没有负面影响。我们的实验表明,Viola和Jones等人用于人脸识别的特征集对于我们的问题是低效的;因此,每个对象都需要定制自己的一套特征,以便实时准确地识别。我们的训练方法,结合适当特征的选择,已经找到了一个非常准确的分类器,只包含30个弱分类器。与现有文献相比,我们总体上实现了用最少的样例、最少的弱分类器、最快的训练时间、具有竞争性检测率和假阳性率的实时目标检测机的设计
{"title":"Pre-Eliminating Features for Fast Training in Real Time Object Detection in Images with a Novel Variant of AdaBoost","authors":"M. Stojmenovic","doi":"10.1109/ICCIS.2006.252285","DOIUrl":"https://doi.org/10.1109/ICCIS.2006.252285","url":null,"abstract":"Our primary interest is to build fast and reliable object recognizers in images based on small training sets. This is important in cases where the training set needs to be built mostly manually, as in the case that we studied, the recognition of the Honda Accord 2004 from rear views. We described a novel variant of the AdaBoost based learning algorithm, which builds a strong classifier by incremental addition of weak classifiers (WCs) that minimize the combined error of the already selected WCs. Each WC is trained only once, and examples do not change their weights. We proposed to pre-eliminate features whose cumulative error of corresponding best WCs exceeds a predetermined threshold value. We tested two straightforward definitions of cumulative error. In both cases, we showed that, when over 97% of the initial features are eliminated at the very beginning from further training, training time is drastically reduced while having little impact on the quality of the pool of available WCs. This is a novel method that has reduced the training set WC quantity to less than 3% of its original number, greatly speeding up training time, and showing no negative impact on the quality of the final classifier. Our experiments indicated that the set of features used by Viola and Jones and others for face recognition was inefficient for our problem; therefore, each object requires its own custom-made set of features for real time and accurate recognition. Our training method, combined with the selection of appropriate features, has resulted in finding a very accurate classifier containing merely 30 weak classifiers. Compared to existing literature, we have overall achieved the design of a real time object detection machine with the least number of examples, the least number of weak classifiers, the fastest training time, and with competitive detection and false positive rates","PeriodicalId":296028,"journal":{"name":"2006 IEEE Conference on Cybernetics and Intelligent Systems","volume":"28 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116314138","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
期刊
2006 IEEE Conference on Cybernetics and Intelligent Systems
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1