Pub Date : 2012-05-16DOI: 10.1109/RCIS.2012.6240447
Elena Kornyshova, R. Deneckère
Information Systems (IS) engineering (ISE) processes contain steps where decisions must be made. Moreover, the growing role of IS involves requirements for their engineering, such as quality, cost, time, and so on. On the one hand, considering these aspects implies that the number of researches dealing with decision-making (DM) in ISE increasingly grows. On the other hand, many DM methods exist and are applied in several fields of ISE. The main characteristic of these applications is that they resolve one DM problem at a time. We have developed a generic DM approach MADISE, which aims at guiding IS engineers through DM activities. The goal of this paper is to check the completeness and flexibility of MADISE by comparing it with the five well-known requirements prioritization approaches.
{"title":"Decision-making method family MADISE: Validation within the requirements engineering domain","authors":"Elena Kornyshova, R. Deneckère","doi":"10.1109/RCIS.2012.6240447","DOIUrl":"https://doi.org/10.1109/RCIS.2012.6240447","url":null,"abstract":"Information Systems (IS) engineering (ISE) processes contain steps where decisions must be made. Moreover, the growing role of IS involves requirements for their engineering, such as quality, cost, time, and so on. On the one hand, considering these aspects implies that the number of researches dealing with decision-making (DM) in ISE increasingly grows. On the other hand, many DM methods exist and are applied in several fields of ISE. The main characteristic of these applications is that they resolve one DM problem at a time. We have developed a generic DM approach MADISE, which aims at guiding IS engineers through DM activities. The goal of this paper is to check the completeness and flexibility of MADISE by comparing it with the five well-known requirements prioritization approaches.","PeriodicalId":130476,"journal":{"name":"2012 Sixth International Conference on Research Challenges in Information Science (RCIS)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125806042","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2012-05-16DOI: 10.1109/RCIS.2012.6240458
Sarah Ayad
The scientific problem addressed in this paper consists in modeling and improving Business Process (BP) models quality. This problem is of growing interest as companies are realizing the undeniable impact of a better understanding of business processes (BP) on the effectiveness, consistency and transparency of their business operations. The research aims at proposing methods and tools for BP model quality measurement and improvement. We propose a semantic approach of quality exploiting domain knowledge.
{"title":"A quality based approach for the analysis and design of Business Process models","authors":"Sarah Ayad","doi":"10.1109/RCIS.2012.6240458","DOIUrl":"https://doi.org/10.1109/RCIS.2012.6240458","url":null,"abstract":"The scientific problem addressed in this paper consists in modeling and improving Business Process (BP) models quality. This problem is of growing interest as companies are realizing the undeniable impact of a better understanding of business processes (BP) on the effectiveness, consistency and transparency of their business operations. The research aims at proposing methods and tools for BP model quality measurement and improvement. We propose a semantic approach of quality exploiting domain knowledge.","PeriodicalId":130476,"journal":{"name":"2012 Sixth International Conference on Research Challenges in Information Science (RCIS)","volume":"55 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130680839","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2012-05-16DOI: 10.1109/RCIS.2012.6240427
Watsawee Sansrimahachai, M. Weal, L. Moreau
Applications that require continuous processing of high-volume data streams have grown in prevalence and importance. These systems process streaming data in real-time and provide instantaneous response to support precise and ontime decisions. In such systems, it is difficult to know exactly how a particular result is generated or more particularly how to precisely trace stream events that caused a particular result. However, such information is extremely important for validating stream processing results. Therefore, it is crucial that stream processing systems have a mechanism for capturing and querying provenance information - the information pertaining to the process that produced result data - at the level of individual stream events, which we refer to as fine-grained provenance. In this paper, we propose a novel fine-grained provenance solution called Stream Ancestor Function - a reverse mapping function used to express precise dependencies between input and output stream elements. We demonstrate how to utilize stream ancestor functions by means of a stream provenance query and replay execution algorithm. Finally, we evaluate the stream ancestor function in terms of storage consumption for provenance collection and system throughput, demonstrating significant reductions in storage size and reasonable processing overheads.
{"title":"Stream ancestor function: A mechanism for fine-grained provenance in stream processing systems","authors":"Watsawee Sansrimahachai, M. Weal, L. Moreau","doi":"10.1109/RCIS.2012.6240427","DOIUrl":"https://doi.org/10.1109/RCIS.2012.6240427","url":null,"abstract":"Applications that require continuous processing of high-volume data streams have grown in prevalence and importance. These systems process streaming data in real-time and provide instantaneous response to support precise and ontime decisions. In such systems, it is difficult to know exactly how a particular result is generated or more particularly how to precisely trace stream events that caused a particular result. However, such information is extremely important for validating stream processing results. Therefore, it is crucial that stream processing systems have a mechanism for capturing and querying provenance information - the information pertaining to the process that produced result data - at the level of individual stream events, which we refer to as fine-grained provenance. In this paper, we propose a novel fine-grained provenance solution called Stream Ancestor Function - a reverse mapping function used to express precise dependencies between input and output stream elements. We demonstrate how to utilize stream ancestor functions by means of a stream provenance query and replay execution algorithm. Finally, we evaluate the stream ancestor function in terms of storage consumption for provenance collection and system throughput, demonstrating significant reductions in storage size and reasonable processing overheads.","PeriodicalId":130476,"journal":{"name":"2012 Sixth International Conference on Research Challenges in Information Science (RCIS)","volume":"4299 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133309086","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2012-05-16DOI: 10.1109/RCIS.2012.6240448
A. Maalel, L. Mejri, Habib Hadj-Mabrouk, H. Ghézala
The work developed in the context of this article stems from the thesis work in progress (done in RIADI labs. at the National School of Computer Sciences, Tunisia in collaboration with the Research Unit of Evaluation of Automated Transport Systems and Safety IFSTTAR, French). The works treat the problem of knowledge management of a critical area, that of security and in particular the railroad accidents. The goal is to provide methodological support and tools to support the capitalization and exploitation of produced knowledge and/or used by domain experts. We have proposed an approach based on domain ontology and Case-based Reasoning CBR. We will present as part of this article, the first realized works in our approach.
{"title":"Toward a knowledge management approach based on an ontology and Case-based Reasoning (CBR): Application to railroad accidents","authors":"A. Maalel, L. Mejri, Habib Hadj-Mabrouk, H. Ghézala","doi":"10.1109/RCIS.2012.6240448","DOIUrl":"https://doi.org/10.1109/RCIS.2012.6240448","url":null,"abstract":"The work developed in the context of this article stems from the thesis work in progress (done in RIADI labs. at the National School of Computer Sciences, Tunisia in collaboration with the Research Unit of Evaluation of Automated Transport Systems and Safety IFSTTAR, French). The works treat the problem of knowledge management of a critical area, that of security and in particular the railroad accidents. The goal is to provide methodological support and tools to support the capitalization and exploitation of produced knowledge and/or used by domain experts. We have proposed an approach based on domain ontology and Case-based Reasoning CBR. We will present as part of this article, the first realized works in our approach.","PeriodicalId":130476,"journal":{"name":"2012 Sixth International Conference on Research Challenges in Information Science (RCIS)","volume":"23 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124486887","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2012-05-16DOI: 10.1109/RCIS.2012.6240445
Giovanni Giachetti, Francisco Valverde, Beatriz Marín
Nowadays, the emergence of several model-driven development (MDD) proposals that are related to multiple domains requires the definition of proper interoperability mechanisms that facilitate the reuse of knowledge in the MDD community by taking advantage of already defined modeling languages, tools, and standards. However, there are no recent studies that cover the existent interoperability alternatives in the model-driven domain nor is there a common interoperability framework. This paper confronts this situation through a systematic analysis of recent interoperability approaches that provide relevant features for MDD processes. From this analysis, a general interoperability framework is depicted, which is complemented with our contributions to solve specific interoperability issues. Therefore, we present those aspects that are already covered by existent proposals as well as those pending subjects that, from our point of view, are future challenges in order to achieve a suitable interoperability framework for MDD approaches.
{"title":"Interoperability for model-driven development: Current state and future challenges","authors":"Giovanni Giachetti, Francisco Valverde, Beatriz Marín","doi":"10.1109/RCIS.2012.6240445","DOIUrl":"https://doi.org/10.1109/RCIS.2012.6240445","url":null,"abstract":"Nowadays, the emergence of several model-driven development (MDD) proposals that are related to multiple domains requires the definition of proper interoperability mechanisms that facilitate the reuse of knowledge in the MDD community by taking advantage of already defined modeling languages, tools, and standards. However, there are no recent studies that cover the existent interoperability alternatives in the model-driven domain nor is there a common interoperability framework. This paper confronts this situation through a systematic analysis of recent interoperability approaches that provide relevant features for MDD processes. From this analysis, a general interoperability framework is depicted, which is complemented with our contributions to solve specific interoperability issues. Therefore, we present those aspects that are already covered by existent proposals as well as those pending subjects that, from our point of view, are future challenges in order to achieve a suitable interoperability framework for MDD approaches.","PeriodicalId":130476,"journal":{"name":"2012 Sixth International Conference on Research Challenges in Information Science (RCIS)","volume":"38 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115706024","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2012-05-16DOI: 10.1109/RCIS.2012.6240462
V. Coll
The heterogeneity and dispersion of genomic data that currently exists between the multiple genomic databases is a big problem for the geneticists when they look for information for their genomic diagnosis. This problem is especially important when it is referred to genetic data about breast cancer due to the large amount of data available about this disease caused by the high incidence in society. The work in this Doctoral Thesis expects to solve this problem by analyzing all available information about breast cancer in the databases and integrating it into an information system created following the Conceptual Modeling rules. With this information system the data about breast cancer will be stored in one efficient and well-structured database, making it easier for geneticists searching for information about this disease.
{"title":"Design and development of a genomic information system to manage breast cancer data","authors":"V. Coll","doi":"10.1109/RCIS.2012.6240462","DOIUrl":"https://doi.org/10.1109/RCIS.2012.6240462","url":null,"abstract":"The heterogeneity and dispersion of genomic data that currently exists between the multiple genomic databases is a big problem for the geneticists when they look for information for their genomic diagnosis. This problem is especially important when it is referred to genetic data about breast cancer due to the large amount of data available about this disease caused by the high incidence in society. The work in this Doctoral Thesis expects to solve this problem by analyzing all available information about breast cancer in the databases and integrating it into an information system created following the Conceptual Modeling rules. With this information system the data about breast cancer will be stored in one efficient and well-structured database, making it easier for geneticists searching for information about this disease.","PeriodicalId":130476,"journal":{"name":"2012 Sixth International Conference on Research Challenges in Information Science (RCIS)","volume":"143 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124012440","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2012-05-16DOI: 10.1109/RCIS.2012.6240431
Abdelghani Bakhtouchi, Ladjel Bellatreche, Stéphane Jean, Y. A. Ameur
With the increasing needs for the world wide enterprises to integrate, share and visualize data from various heterogeneous, autonomous and distributed sources data and Web data covering a given domain, the development of integration and reconciliation solutions becomes a challenging issue. The existing studies on data integration and reconciliation of results have been developed in an isolated way and did not consider the strong integration between these two processes. On one hand, ontologies were largely used for building automatic integration systems due to their ability to reduce schematic and semantic heterogeneities that may exist among sources. On the other hand, reconciliation of results is performed either by considering that all sources use the same identifier for an instance or by means of statistical methods that identify affinities between concepts. These reconciliation solutions are not usually suitable for real-world sensitive-applications where exact results are required and where each source may use a different identifier for the same concept. In this paper, we propose a methodology that simultaneously integrate source data and reconciliate their instances based on ontologies enriched with functional dependencies (FD) in a mediation architecture. The presence of FD gives more autonomy to sources when choosing their primary keys and facilitates the result reconciliation. This methodology is experimented using the Lehigh University Benchmark (LUBM) dataset to show its scalability and the quality of the reconciliation result phase.
随着全球企业对集成、共享和可视化来自各种异构、自治和分布式源数据和覆盖给定域的Web数据的需求日益增长,集成和协调解决方案的开发成为一个具有挑战性的问题。现有的数据整合与结果协调的研究都是孤立发展的,没有考虑到这两个过程之间的强整合。一方面,本体主要用于构建自动集成系统,因为它们能够减少源之间可能存在的示意图和语义异构性。另一方面,通过考虑所有源对实例使用相同的标识符或通过识别概念之间的亲和力的统计方法来执行结果的协调。这些协调解决方案通常不适合现实世界中的敏感应用程序,这些应用程序需要精确的结果,并且每个源可能对相同的概念使用不同的标识符。在本文中,我们提出了一种方法,该方法可以同时集成源数据并基于中介体系结构中富含功能依赖关系(FD)的本体来协调它们的实例。FD的存在为源在选择主键时提供了更多的自主权,并促进了结果协调。使用Lehigh University Benchmark (LUBM)数据集对该方法进行了实验,以显示其可扩展性和协调结果阶段的质量。
{"title":"Ontologies as a solution for simultaneously integrating and reconciliating data sources","authors":"Abdelghani Bakhtouchi, Ladjel Bellatreche, Stéphane Jean, Y. A. Ameur","doi":"10.1109/RCIS.2012.6240431","DOIUrl":"https://doi.org/10.1109/RCIS.2012.6240431","url":null,"abstract":"With the increasing needs for the world wide enterprises to integrate, share and visualize data from various heterogeneous, autonomous and distributed sources data and Web data covering a given domain, the development of integration and reconciliation solutions becomes a challenging issue. The existing studies on data integration and reconciliation of results have been developed in an isolated way and did not consider the strong integration between these two processes. On one hand, ontologies were largely used for building automatic integration systems due to their ability to reduce schematic and semantic heterogeneities that may exist among sources. On the other hand, reconciliation of results is performed either by considering that all sources use the same identifier for an instance or by means of statistical methods that identify affinities between concepts. These reconciliation solutions are not usually suitable for real-world sensitive-applications where exact results are required and where each source may use a different identifier for the same concept. In this paper, we propose a methodology that simultaneously integrate source data and reconciliate their instances based on ontologies enriched with functional dependencies (FD) in a mediation architecture. The presence of FD gives more autonomy to sources when choosing their primary keys and facilitates the result reconciliation. This methodology is experimented using the Lehigh University Benchmark (LUBM) dataset to show its scalability and the quality of the reconciliation result phase.","PeriodicalId":130476,"journal":{"name":"2012 Sixth International Conference on Research Challenges in Information Science (RCIS)","volume":"14 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128672215","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2012-05-16DOI: 10.1109/RCIS.2012.6240428
Sergio España, M. Ruiz, Arturo González
There is an open challenge in the area of model-driven requirements engineering. Model transformations that allow deriving (platform-independent) conceptual models from (computation-independent) requirements models are being proposed. However, rigorous assessments of the quality of the resulting conceptual models are needed. This paper reports a controlled experiment that compares the performance of subjects applying two different techniques for deriving object-oriented, UML-compliant conceptual models. We compare the quality of the OO-Method conceptual models obtained by applying a text-based derivation technique (which mimics what OO-Method practitioners actually do in real projects) with the quality obtained by applying a novel communication-based derivation technique (which takes as input Communication Analysis requirements models). The results show that there is an interaction between the derivation technique and the OO-Method modelling competence of the subject: the derivation technique has a significant impact on model completeness within the high-competence group. No impact has been observed on model validity. We also discuss new challenges raised by the evaluation.
{"title":"Systematic derivation of conceptual models from requirements models: A controlled experiment","authors":"Sergio España, M. Ruiz, Arturo González","doi":"10.1109/RCIS.2012.6240428","DOIUrl":"https://doi.org/10.1109/RCIS.2012.6240428","url":null,"abstract":"There is an open challenge in the area of model-driven requirements engineering. Model transformations that allow deriving (platform-independent) conceptual models from (computation-independent) requirements models are being proposed. However, rigorous assessments of the quality of the resulting conceptual models are needed. This paper reports a controlled experiment that compares the performance of subjects applying two different techniques for deriving object-oriented, UML-compliant conceptual models. We compare the quality of the OO-Method conceptual models obtained by applying a text-based derivation technique (which mimics what OO-Method practitioners actually do in real projects) with the quality obtained by applying a novel communication-based derivation technique (which takes as input Communication Analysis requirements models). The results show that there is an interaction between the derivation technique and the OO-Method modelling competence of the subject: the derivation technique has a significant impact on model completeness within the high-competence group. No impact has been observed on model validity. We also discuss new challenges raised by the evaluation.","PeriodicalId":130476,"journal":{"name":"2012 Sixth International Conference on Research Challenges in Information Science (RCIS)","volume":"6 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127812200","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2012-05-16DOI: 10.1109/RCIS.2012.6240449
F. Beuvens, J. Vanderdonckt
Several algorithms have been developed for pen-based gesture recognition. Yet, their integration in streamline engineering of interactive systems is bound to several shortcomings: they are hard to compare to each other, determining which one is the most suitable in which situation is a research problem, their performance largely vary depending on contextual parameters that are hard to predict, their fine-tuning in a real interactive application is a challenge. In order to address these shortcomings, we developed UsiGesture, an engineering method and a software support platform that accommodates multiple algorithms for pen-based gesture recognition in order to integrate them in a straightforward way into interactive computing systems. The method is aimed at providing designers and developers with support for the following steps: defining a dataset that is appropriate for an interactive system (e.g., made of commands, symbols, characters), determining the most suitable gesture recognition algorithm depending on contextual variables (e.g., user, platform, environment), fine-tuning the parameters of this algorithm by multi-criteria optimization (e.g., system speed and recognition rate vs. human distinguishability and perception), and incorporation of the fine-tuned algorithm in an integrated development environment for engineering interactive systems.
{"title":"UsiGesture: An environment for integrating pen-based interaction in user interface development","authors":"F. Beuvens, J. Vanderdonckt","doi":"10.1109/RCIS.2012.6240449","DOIUrl":"https://doi.org/10.1109/RCIS.2012.6240449","url":null,"abstract":"Several algorithms have been developed for pen-based gesture recognition. Yet, their integration in streamline engineering of interactive systems is bound to several shortcomings: they are hard to compare to each other, determining which one is the most suitable in which situation is a research problem, their performance largely vary depending on contextual parameters that are hard to predict, their fine-tuning in a real interactive application is a challenge. In order to address these shortcomings, we developed UsiGesture, an engineering method and a software support platform that accommodates multiple algorithms for pen-based gesture recognition in order to integrate them in a straightforward way into interactive computing systems. The method is aimed at providing designers and developers with support for the following steps: defining a dataset that is appropriate for an interactive system (e.g., made of commands, symbols, characters), determining the most suitable gesture recognition algorithm depending on contextual variables (e.g., user, platform, environment), fine-tuning the parameters of this algorithm by multi-criteria optimization (e.g., system speed and recognition rate vs. human distinguishability and perception), and incorporation of the fine-tuned algorithm in an integrated development environment for engineering interactive systems.","PeriodicalId":130476,"journal":{"name":"2012 Sixth International Conference on Research Challenges in Information Science (RCIS)","volume":"143 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131808368","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2012-05-16DOI: 10.1109/RCIS.2012.6240432
Erick Stattner, M. Collard
In the area of the link mining, frequent pattern discovery tasks generally consist in the search for subgraphs frequently found in a network or a set of networks. Very recently, new axes of this problem has been proposed through the search for frequent links. Unlike traditional approaches that focus solely on structural regularities, frequent link mining methods exploit both the network structure and the attributes of nodes for extracting regularities in the links existing between node groups that share common characteristics. However, extracting frequent links is still a particularly challenging and computationally intensive issue since it is much dependent on the number of links. In this study, we propose a solution that is able to reduce the computing time by reducing the search to only a subset of of nodes. Experiments were conducted to understand the effects of different thresholds of the subset size on the loss of patterns and the gain in terms of computation time. Our solution proves to be efficient for a rather wide range of thresholds.
{"title":"How to extract frequent links with frequent itemsets in social networks?","authors":"Erick Stattner, M. Collard","doi":"10.1109/RCIS.2012.6240432","DOIUrl":"https://doi.org/10.1109/RCIS.2012.6240432","url":null,"abstract":"In the area of the link mining, frequent pattern discovery tasks generally consist in the search for subgraphs frequently found in a network or a set of networks. Very recently, new axes of this problem has been proposed through the search for frequent links. Unlike traditional approaches that focus solely on structural regularities, frequent link mining methods exploit both the network structure and the attributes of nodes for extracting regularities in the links existing between node groups that share common characteristics. However, extracting frequent links is still a particularly challenging and computationally intensive issue since it is much dependent on the number of links. In this study, we propose a solution that is able to reduce the computing time by reducing the search to only a subset of of nodes. Experiments were conducted to understand the effects of different thresholds of the subset size on the loss of patterns and the gain in terms of computation time. Our solution proves to be efficient for a rather wide range of thresholds.","PeriodicalId":130476,"journal":{"name":"2012 Sixth International Conference on Research Challenges in Information Science (RCIS)","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132200707","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}