Pub Date : 2010-12-31DOI: 10.3745/KIPSTD.2010.17D.6.463
Hyun-Kyung Shin
We presented a very fast and robust method of text line segmentation based on the DCT blocks of color image without decompression and binary transformation processes. Using DC and another three primary AC coefficients from block DCT we created a gray-scale image having reduced size by 8x8. In order to detect and locate white strips between text lines we analyzed horizontal and vertical projection profiles of the image and we applied a direct markov model to recover the missing white strips by estimating hidden periodicity. We presented performance results. The results showed that our method was 40 - 100 times faster than traditional method.
{"title":"Fast Text Line Segmentation Model Based on DCT for Color Image","authors":"Hyun-Kyung Shin","doi":"10.3745/KIPSTD.2010.17D.6.463","DOIUrl":"https://doi.org/10.3745/KIPSTD.2010.17D.6.463","url":null,"abstract":"We presented a very fast and robust method of text line segmentation based on the DCT blocks of color image without decompression and binary transformation processes. Using DC and another three primary AC coefficients from block DCT we created a gray-scale image having reduced size by 8x8. In order to detect and locate white strips between text lines we analyzed horizontal and vertical projection profiles of the image and we applied a direct markov model to recover the missing white strips by estimating hidden periodicity. We presented performance results. The results showed that our method was 40 - 100 times faster than traditional method.","PeriodicalId":348746,"journal":{"name":"The Kips Transactions:partd","volume":"21 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128461463","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2010-10-31DOI: 10.3745/KIPSTD.2010.17D.5.353
Hye-Kyeong Ko
Unlike the traditional databases, queries on XML streams are restricted to a real time processing and memory usage. In this paper, a robust labeling scheme is proposed, which quickly identifies structural relationship between XML fragments. The proposed labeling scheme provides an effective query processing by removing many redundant operations and minimizing the number of fragments being processed. In experimental results, the proposed labeling scheme efficiently processes query processing and optimizes memory usage.
{"title":"Efficient Labeling Scheme for Query Processing over XML Fragment Stream in Wireless Computing","authors":"Hye-Kyeong Ko","doi":"10.3745/KIPSTD.2010.17D.5.353","DOIUrl":"https://doi.org/10.3745/KIPSTD.2010.17D.5.353","url":null,"abstract":"Unlike the traditional databases, queries on XML streams are restricted to a real time processing and memory usage. In this paper, a robust labeling scheme is proposed, which quickly identifies structural relationship between XML fragments. The proposed labeling scheme provides an effective query processing by removing many redundant operations and minimizing the number of fragments being processed. In experimental results, the proposed labeling scheme efficiently processes query processing and optimizes memory usage.","PeriodicalId":348746,"journal":{"name":"The Kips Transactions:partd","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-10-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129800639","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2010-10-31DOI: 10.3745/KIPSTC.2010.17D.5.371
JongKeun Kim, Heeae Ko, Kunjung Sim, Zhao Meihua, Young-Jae Lim
One of the advantages of the mobile internet is that it is accessible everywhere. However, producing the mobile internet contents is hard because the standards of the contents are different depending on the telecommunications companies and the mobile phone terminals. Because of this, mobile contents production has not been activated and profit models are rarely found. In order to vitalize the mobile contents, we developed the Mobile UCC (User Created Contents) Phone Page Service with which general users can produce and provide the mobile phone pages easily. In addition, we began to provide advertisement on the UCC phone pages as a profit model. However, as mentioned already, the contents that meet each of the standards should be prepared in advance to be compatible with the different standards. For this reason, only one advertisement could be provided on one UCC phone page. Previous studies were focused on the standardization of different standards by integrating them. Though the standardization may be applicable to the mobile phones that will be newly released, it is hardly applied to the conventional mobile phones. To solve the abovementioned problems and improve the drawbacks of the conventional studies, we proposed in this article the technology to synthesize the phone page contents and independent advertisement at the moment when the UCC phone page prepared by the users is provided in real time. This technology allows to provide various advertisement on one UCC phone page and administrate the start, end and frequency of specific advertisements. This method was applied in this study to vitalize the mobile contents and improve the profit model.
{"title":"Composition Technique for Independent Real-time Advertisement on Mobile UCC Phonepage","authors":"JongKeun Kim, Heeae Ko, Kunjung Sim, Zhao Meihua, Young-Jae Lim","doi":"10.3745/KIPSTC.2010.17D.5.371","DOIUrl":"https://doi.org/10.3745/KIPSTC.2010.17D.5.371","url":null,"abstract":"One of the advantages of the mobile internet is that it is accessible everywhere. However, producing the mobile internet contents is hard because the standards of the contents are different depending on the telecommunications companies and the mobile phone terminals. Because of this, mobile contents production has not been activated and profit models are rarely found. In order to vitalize the mobile contents, we developed the Mobile UCC (User Created Contents) Phone Page Service with which general users can produce and provide the mobile phone pages easily. In addition, we began to provide advertisement on the UCC phone pages as a profit model. However, as mentioned already, the contents that meet each of the standards should be prepared in advance to be compatible with the different standards. For this reason, only one advertisement could be provided on one UCC phone page. Previous studies were focused on the standardization of different standards by integrating them. Though the standardization may be applicable to the mobile phones that will be newly released, it is hardly applied to the conventional mobile phones. To solve the abovementioned problems and improve the drawbacks of the conventional studies, we proposed in this article the technology to synthesize the phone page contents and independent advertisement at the moment when the UCC phone page prepared by the users is provided in real time. This technology allows to provide various advertisement on one UCC phone page and administrate the start, end and frequency of specific advertisements. This method was applied in this study to vitalize the mobile contents and improve the profit model.","PeriodicalId":348746,"journal":{"name":"The Kips Transactions:partd","volume":"56 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-10-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130315026","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2010-10-31DOI: 10.3745/KIPSTD.2010.17D.5.347
Chan-ki Jung, Sangkyu Hwang, Y. Byun
Automatic service identification and quality evaluation is one of key characteristics for a Service-Oriented Computing, being receiving a lot of attention from researchers in recent years. However, most researchers focus on identifying and evaluating application services and do not present methods for automatically identifying and evaluating business services from business processes. In general, the manual business service identification process by a human expert is a highly expensive and ambiguous task and may result in the service design with bad quality because of errors and misconception. We propose an automatic business service identification and quality evaluation method using Enterprise Architecture as a machine understandable knowledge-base. We verify the effectiveness of the proposed method through a case study on Department of Defense Enterprise Architecture.
{"title":"A Business Service Identification and Quality Evaluation Using Enterprise Architecture","authors":"Chan-ki Jung, Sangkyu Hwang, Y. Byun","doi":"10.3745/KIPSTD.2010.17D.5.347","DOIUrl":"https://doi.org/10.3745/KIPSTD.2010.17D.5.347","url":null,"abstract":"Automatic service identification and quality evaluation is one of key characteristics for a Service-Oriented Computing, being receiving a lot of attention from researchers in recent years. However, most researchers focus on identifying and evaluating application services and do not present methods for automatically identifying and evaluating business services from business processes. In general, the manual business service identification process by a human expert is a highly expensive and ambiguous task and may result in the service design with bad quality because of errors and misconception. We propose an automatic business service identification and quality evaluation method using Enterprise Architecture as a machine understandable knowledge-base. We verify the effectiveness of the proposed method through a case study on Department of Defense Enterprise Architecture.","PeriodicalId":348746,"journal":{"name":"The Kips Transactions:partd","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-10-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130414854","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2010-10-31DOI: 10.3745/KIPSTD.2010.17D.5.327
Sang-Un Lee
This paper proposed software equation that is relation with effort and duration based on function point (FP) software size. Existent software equation based on lines of code (LOC). LOC sees big difference according to development language and there are a lot of difficulties in software size estimation. First, considered method that change LOC to FP. But, this method is not decided definitely conversion ratio between LOC and FP by development language. Also, failed though the conversion ratio motives software formula because was not presented about specification development language. Therefore, we derived software formula directly to large project data that was developed by FP. Firstly, datas that reasonable development period is set among development projects. Secondly, FP through regression analysis about this data and effort, motived relation with FP and duration. Finally, software equation was derived from these relation. Proposed model solves application problems that LOC-based model has and has advantage that application is possible easily in business.
{"title":"Software Equation Based on Function Points","authors":"Sang-Un Lee","doi":"10.3745/KIPSTD.2010.17D.5.327","DOIUrl":"https://doi.org/10.3745/KIPSTD.2010.17D.5.327","url":null,"abstract":"This paper proposed software equation that is relation with effort and duration based on function point (FP) software size. Existent software equation based on lines of code (LOC). LOC sees big difference according to development language and there are a lot of difficulties in software size estimation. First, considered method that change LOC to FP. But, this method is not decided definitely conversion ratio between LOC and FP by development language. Also, failed though the conversion ratio motives software formula because was not presented about specification development language. Therefore, we derived software formula directly to large project data that was developed by FP. Firstly, datas that reasonable development period is set among development projects. Secondly, FP through regression analysis about this data and effort, motived relation with FP and duration. Finally, software equation was derived from these relation. Proposed model solves application problems that LOC-based model has and has advantage that application is possible easily in business.","PeriodicalId":348746,"journal":{"name":"The Kips Transactions:partd","volume":"35 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-10-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128325506","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2010-10-31DOI: 10.3745/KIPSTD.2010.17D.5.337
Hongsuk Lee, K. Chung, Kyunghee Choi
Timing diagram is popularly utilized for the reason of its advantages; it is convenient for timing diagram to describe behavior of system and it is simple for described behaviors to recognize it. Various techniques are needed to test systems described in timing diagram. One of them is a technique to derive the system into a certain condition under which a test case is effective. This paper proposes a technique to automatically generate the test input sequence to reach the condition for systems described in timing diagram. It requires a proper input set which satisfy transition condition restricted by input waveform and timing constraints to generate a test input sequence automatically. To solve the problem, this paper chooses an approach utilizing the linear programming, and solving procedure is as follows: 1) Get a Timing diagram model as an input, and transforms the timing diagram model into a linear programming problem. 2) Solve the linear programming problem using a linear programming tool. 3) Generate test input sequences of a timing diagram model from the solution of linear programming problem. This paper addresses the formal method to drive the linear programming model from a given timing diagram, shows the feasibility of our approach by prove it, and demonstrates the usability of our paper by showing that our implemented tool solves an example of a timing diagram model.
{"title":"Test Input Sequence Generation Strategy for Timing Diagram using Linear Programming","authors":"Hongsuk Lee, K. Chung, Kyunghee Choi","doi":"10.3745/KIPSTD.2010.17D.5.337","DOIUrl":"https://doi.org/10.3745/KIPSTD.2010.17D.5.337","url":null,"abstract":"Timing diagram is popularly utilized for the reason of its advantages; it is convenient for timing diagram to describe behavior of system and it is simple for described behaviors to recognize it. Various techniques are needed to test systems described in timing diagram. One of them is a technique to derive the system into a certain condition under which a test case is effective. This paper proposes a technique to automatically generate the test input sequence to reach the condition for systems described in timing diagram. It requires a proper input set which satisfy transition condition restricted by input waveform and timing constraints to generate a test input sequence automatically. To solve the problem, this paper chooses an approach utilizing the linear programming, and solving procedure is as follows: 1) Get a Timing diagram model as an input, and transforms the timing diagram model into a linear programming problem. 2) Solve the linear programming problem using a linear programming tool. 3) Generate test input sequences of a timing diagram model from the solution of linear programming problem. This paper addresses the formal method to drive the linear programming model from a given timing diagram, shows the feasibility of our approach by prove it, and demonstrates the usability of our paper by showing that our implemented tool solves an example of a timing diagram model.","PeriodicalId":348746,"journal":{"name":"The Kips Transactions:partd","volume":"157 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-10-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128173485","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2010-08-31DOI: 10.3745/KIPSTD.2010.17D.4.283
Hongsuk Lee, K. Chung, Kyunghee Choi
Timing diagram is a useful tool for describing the specification of system, but there is no study for test case strategy of a timing diagram. To solve this problem, we followed the steps to generate test cases from timing diagram in this paper. 1) We defined a timing diagram formally. 2) We describe the method of transforming from a timing diagram model into a Stateflow model which has an equivalent relationship between a timing diagram model and a transformed Stateflow model. 3) We generated test cases from a transformed Stateflow model using SDV which is plugged in Simulink. To show that our approach is useful, we made an experiment with a surveillance model and arbitrary timing diagram models. In the experiment we transformed timing diagram models into Stateflow models, generated test cases from transformed Stateflow models using SDV, and analyzed the generation results. The conclusion that can be obtained from this study is that timing diagram is not only a specification tool but also a useful tool when users are trying to generate test cases based on model.
{"title":"Test Case Generation Strategy for Timing Diagram","authors":"Hongsuk Lee, K. Chung, Kyunghee Choi","doi":"10.3745/KIPSTD.2010.17D.4.283","DOIUrl":"https://doi.org/10.3745/KIPSTD.2010.17D.4.283","url":null,"abstract":"Timing diagram is a useful tool for describing the specification of system, but there is no study for test case strategy of a timing diagram. To solve this problem, we followed the steps to generate test cases from timing diagram in this paper. 1) We defined a timing diagram formally. 2) We describe the method of transforming from a timing diagram model into a Stateflow model which has an equivalent relationship between a timing diagram model and a transformed Stateflow model. 3) We generated test cases from a transformed Stateflow model using SDV which is plugged in Simulink. To show that our approach is useful, we made an experiment with a surveillance model and arbitrary timing diagram models. In the experiment we transformed timing diagram models into Stateflow models, generated test cases from transformed Stateflow models using SDV, and analyzed the generation results. The conclusion that can be obtained from this study is that timing diagram is not only a specification tool but also a useful tool when users are trying to generate test cases based on model.","PeriodicalId":348746,"journal":{"name":"The Kips Transactions:partd","volume":"46 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-08-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128187877","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2010-08-31DOI: 10.3745/KIPSTD.2010.17D.4.271
Ju-Young Kim, S. Rhew, ManSoo Hwang
Insufficient requirement management accounts for 54% ofunsuccessful software development projects and 22% of insufficient requirement management comes from requirement change management. Hence, requirement management activities are important to reduce failure rates and a tracing method is suggested as the major factor in requirements change management. A traceability table is easy to use because of its legibility accurate tracing. However, traceability tables of existing studies have failed to concretely suggest method of change management and effect of traceability. Also, studies of methods to estimate change impact is complex. Hence, this study suggests how to use a traceability table to manage changes in requirements. Together, in comparison to existing studies, this study suggests easier methods to estimate change rate and change impact. Also Fifteen projects were sampled to test the hypothesis that traceability table influences the success of projects and that it decreases the failure rate that comes from the insufficient requirements management.
{"title":"A Study of Requirement Change Management and Traceability Effect Using Traceability Table","authors":"Ju-Young Kim, S. Rhew, ManSoo Hwang","doi":"10.3745/KIPSTD.2010.17D.4.271","DOIUrl":"https://doi.org/10.3745/KIPSTD.2010.17D.4.271","url":null,"abstract":"Insufficient requirement management accounts for 54% ofunsuccessful software development projects and 22% of insufficient requirement management comes from requirement change management. Hence, requirement management activities are important to reduce failure rates and a tracing method is suggested as the major factor in requirements change management. A traceability table is easy to use because of its legibility accurate tracing. However, traceability tables of existing studies have failed to concretely suggest method of change management and effect of traceability. Also, studies of methods to estimate change impact is complex. Hence, this study suggests how to use a traceability table to manage changes in requirements. Together, in comparison to existing studies, this study suggests easier methods to estimate change rate and change impact. Also Fifteen projects were sampled to test the hypothesis that traceability table influences the success of projects and that it decreases the failure rate that comes from the insufficient requirements management.","PeriodicalId":348746,"journal":{"name":"The Kips Transactions:partd","volume":"11 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-08-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123671776","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2010-08-31DOI: 10.3745/KIPSTD.2010.17D.4.259
Jae-won Park, Jaehyun Choi, P. Cho, Nam-Yong Lee
Information systems nowadays are heterogeneous and distributed which integrate the enterprise information by processes. They are also very complex, because they are linked together by processes. It aims to integrate the systems so that these systems work as one system. A process is a framework which contains all of the business activities in an enterprise, and has a lot of information which is needed for measuring performance. A process consists of activities, and an activity contains events which can be considered information sources. In most cases, it is very valuable to determine if a process is meaningful, but it is difficult because of the complexity in measuring performance, and also because finding relationships between business factors and events is not a simple problem. So it would reduce operation cost and allow efficient process execution if I could evaluate the process before it ends. In this paper we propose an event based process measurement model. First, we propose the concept of process performance measurement, and a model for selecting process and activity indexes from the events which are collected from information systems. Second, we propose at methodologies and data schema that can store and manage the selected process indexes, the mapping methods between indexes and events. Finally, we propose a process Performance measurement model using the collected events which gives users a valuable managerial information.
{"title":"Process Performance Measurement Model Based on Event for an efficient Decision-Making","authors":"Jae-won Park, Jaehyun Choi, P. Cho, Nam-Yong Lee","doi":"10.3745/KIPSTD.2010.17D.4.259","DOIUrl":"https://doi.org/10.3745/KIPSTD.2010.17D.4.259","url":null,"abstract":"Information systems nowadays are heterogeneous and distributed which integrate the enterprise information by processes. They are also very complex, because they are linked together by processes. It aims to integrate the systems so that these systems work as one system. A process is a framework which contains all of the business activities in an enterprise, and has a lot of information which is needed for measuring performance. A process consists of activities, and an activity contains events which can be considered information sources. In most cases, it is very valuable to determine if a process is meaningful, but it is difficult because of the complexity in measuring performance, and also because finding relationships between business factors and events is not a simple problem. So it would reduce operation cost and allow efficient process execution if I could evaluate the process before it ends. In this paper we propose an event based process measurement model. First, we propose the concept of process performance measurement, and a model for selecting process and activity indexes from the events which are collected from information systems. Second, we propose at methodologies and data schema that can store and manage the selected process indexes, the mapping methods between indexes and events. Finally, we propose a process Performance measurement model using the collected events which gives users a valuable managerial information.","PeriodicalId":348746,"journal":{"name":"The Kips Transactions:partd","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-08-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125456841","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2010-08-31DOI: 10.3745/KIPSTD.2010.17D.4.253
Byeong-Soo Jeong, Ahmed S. Farhan
Traditional frequent pattern mining considers equal profit/weight value of every item. Weighted Frequent Pattern (WFP) mining becomes an important research issue in data mining and knowledge discovery by considering different weights for different items. Existing algorithms in this area are based on fixed weight. But in our real world scenarios the price/weight/importance of a pattern may vary frequently due to some unavoidable situations. Tracking these dynamic changes is very necessary in different application area such as retail market basket data analysis and web click stream management. In this paper, we propose a novel concept of dynamic weight and an algorithm DWFPM (dynamic weighted frequent pattern mining). Our algorithm can handle the situation where price/weight of a pattern may vary dynamically. It scans the database exactly once and also eligible for real time data processing. To our knowledge, this is the first research work to mine weighted frequent patterns using dynamic weights. Extensive performance analyses show that our algorithm is very efficient and scalable for WFP mining using dynamic weights.
传统的频繁模式挖掘考虑每个项目的利润/权重值相等。加权频繁模式挖掘(Weighted frequency Pattern, WFP)是数据挖掘和知识发现领域的一个重要研究课题。该领域现有的算法都是基于固定权值的。但在现实世界中,由于某些不可避免的情况,模式的价格/权重/重要性可能会经常变化。在零售市场购物篮数据分析和网络点击流管理等不同的应用领域,跟踪这些动态变化是非常必要的。本文提出了一种新的动态权值概念和动态加权频繁模式挖掘算法(DWFPM)。我们的算法可以处理价格/权重可能动态变化的情况。它只扫描数据库一次,也可以进行实时数据处理。据我们所知,这是第一个使用动态权重挖掘加权频繁模式的研究工作。广泛的性能分析表明,我们的算法对于使用动态权重的WFP采矿非常有效和可扩展。
{"title":"Efficient Dynamic Weighted Frequent Pattern Mining by using a Prefix-Tree","authors":"Byeong-Soo Jeong, Ahmed S. Farhan","doi":"10.3745/KIPSTD.2010.17D.4.253","DOIUrl":"https://doi.org/10.3745/KIPSTD.2010.17D.4.253","url":null,"abstract":"Traditional frequent pattern mining considers equal profit/weight value of every item. Weighted Frequent Pattern (WFP) mining becomes an important research issue in data mining and knowledge discovery by considering different weights for different items. Existing algorithms in this area are based on fixed weight. But in our real world scenarios the price/weight/importance of a pattern may vary frequently due to some unavoidable situations. Tracking these dynamic changes is very necessary in different application area such as retail market basket data analysis and web click stream management. In this paper, we propose a novel concept of dynamic weight and an algorithm DWFPM (dynamic weighted frequent pattern mining). Our algorithm can handle the situation where price/weight of a pattern may vary dynamically. It scans the database exactly once and also eligible for real time data processing. To our knowledge, this is the first research work to mine weighted frequent patterns using dynamic weights. Extensive performance analyses show that our algorithm is very efficient and scalable for WFP mining using dynamic weights.","PeriodicalId":348746,"journal":{"name":"The Kips Transactions:partd","volume":"123 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-08-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117131099","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}