Pub Date : 1998-02-23DOI: 10.1109/ICDE.1998.655765
Evangelos Simoudis
Summary form only given, as follows. Data mining applications deployed in industry are aimed at satisfying two problems organizations face: customer intimacy and better utilization of data assets. These applications can be divided into those that use micro-mining, i.e. single-mining-component desktop systems, and those who use macro-mining, i.e. multi-component server-based systems. The macro-mining applications are usually coupled with data warehouses. The interesting result of this coupling for the data mining community is that the data warehouses cannot be supported by the current data mining offerings delaying the deployment of applications in production environments. The data volumes are too large, the data types too diverse and the data characteristics too incompatible for the existing data mining algorithms. Furthermore, the pure mining operation is a very small part of the entire application life-cycle. The author presents the issues related to the coupling of macro-mining with data warehouses, and proposes issues that must be resolved for large-scale data mining applications to continue being deployed successfully.
{"title":"Industry applications of data mining: challenges and opportunities","authors":"Evangelos Simoudis","doi":"10.1109/ICDE.1998.655765","DOIUrl":"https://doi.org/10.1109/ICDE.1998.655765","url":null,"abstract":"Summary form only given, as follows. Data mining applications deployed in industry are aimed at satisfying two problems organizations face: customer intimacy and better utilization of data assets. These applications can be divided into those that use micro-mining, i.e. single-mining-component desktop systems, and those who use macro-mining, i.e. multi-component server-based systems. The macro-mining applications are usually coupled with data warehouses. The interesting result of this coupling for the data mining community is that the data warehouses cannot be supported by the current data mining offerings delaying the deployment of applications in production environments. The data volumes are too large, the data types too diverse and the data characteristics too incompatible for the existing data mining algorithms. Furthermore, the pure mining operation is a very small part of the entire application life-cycle. The author presents the issues related to the coupling of macro-mining with data warehouses, and proposes issues that must be resolved for large-scale data mining applications to continue being deployed successfully.","PeriodicalId":264926,"journal":{"name":"Proceedings 14th International Conference on Data Engineering","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1998-02-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128868330","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 1998-02-23DOI: 10.1109/ICDE.1998.655815
T. Hara, K. Harumoto, M. Tsukamoto, S. Nishio
Because of the recent development of network technologies such as ATM (Asynchronous Transfer Mode), broader channel bandwidth is becoming available everywhere in the world wide networks. As one of the new technologies to make good use of such broadband channel, dynamic relocation of databases through networks, which we call database migration, will soon become a powerful and basic database operation of practical use. We discuss our proposal of a distributed database system, DB-MAN (distributed database system based on DataBase Migration in ATM Networks), which takes advantage of database migration in virtual LANs (local area networks) of ATM networks. DB-MAN has two notable mechanisms: a mechanism for selecting the transaction processing method and a mechanism for concurrency control with database migration. The former, is a mechanism which chooses the more efficient method between two transaction processing methods: the conventional method based on the two phase commit protocol and our method employing database migration. The latter is a mechanism to prevent the transaction processing throughput from deteriorating in environments where data contention is a significant factor. Then we show simulation results regarding performance comparison between our proposed system and the conventional distributed database system based on the two phase commit protocol. The obtained results demonstrate that effective use of database migration gives higher performance than that of the conventional system.
由于近年来诸如ATM(异步传输模式)等网络技术的发展,更宽的信道带宽在全球网络中无处不在。作为利用这种宽带信道的新技术之一,通过网络对数据库进行动态迁移,即数据库迁移,将很快成为一种实用的强大的基础数据库操作方法。我们提出了一种基于ATM网络数据库迁移的分布式数据库系统DB-MAN (distributed database system based on database Migration in ATM network),它利用了ATM网络虚拟局域网(local area network)中的数据库迁移。DB-MAN有两个值得注意的机制:一个是选择事务处理方法的机制,另一个是通过数据库迁移进行并发控制的机制。前者是在传统的基于两阶段提交协议的事务处理方法和采用数据库迁移的事务处理方法之间选择效率更高的一种机制。后者是一种在数据争用是一个重要因素的环境中防止事务处理吞吐量恶化的机制。在此基础上,给出了系统与基于两阶段提交协议的传统分布式数据库系统的性能对比仿真结果。结果表明,有效地利用数据库迁移可以获得比传统系统更高的性能。
{"title":"DB-MAN: a distributed database system based on database migration in ATM networks","authors":"T. Hara, K. Harumoto, M. Tsukamoto, S. Nishio","doi":"10.1109/ICDE.1998.655815","DOIUrl":"https://doi.org/10.1109/ICDE.1998.655815","url":null,"abstract":"Because of the recent development of network technologies such as ATM (Asynchronous Transfer Mode), broader channel bandwidth is becoming available everywhere in the world wide networks. As one of the new technologies to make good use of such broadband channel, dynamic relocation of databases through networks, which we call database migration, will soon become a powerful and basic database operation of practical use. We discuss our proposal of a distributed database system, DB-MAN (distributed database system based on DataBase Migration in ATM Networks), which takes advantage of database migration in virtual LANs (local area networks) of ATM networks. DB-MAN has two notable mechanisms: a mechanism for selecting the transaction processing method and a mechanism for concurrency control with database migration. The former, is a mechanism which chooses the more efficient method between two transaction processing methods: the conventional method based on the two phase commit protocol and our method employing database migration. The latter is a mechanism to prevent the transaction processing throughput from deteriorating in environments where data contention is a significant factor. Then we show simulation results regarding performance comparison between our proposed system and the conventional distributed database system based on the two phase commit protocol. The obtained results demonstrate that effective use of database migration gives higher performance than that of the conventional system.","PeriodicalId":264926,"journal":{"name":"Proceedings 14th International Conference on Data Engineering","volume":"42 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1998-02-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129467762","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 1998-02-23DOI: 10.1109/ICDE.1998.655771
D. Lomet
We describe how to recover applications after system crashes using database recovery. Earlier efforts, based on frequent application checkpoints and/or logging values read, are very expensive. We treat application state as a cached object and log application execution as operations in the recovery framework of D. Lomet and M. Tuttle (1995). Logging application execution does not require logging the application state. Further logged application reads are mostly logical operations in which only the data source identity is logged. We describe a cache manager that handles the flush order dependencies introduced by these log operations and a recovery process that restores application state by replaying the application.
{"title":"Persistent applications using generalized redo recovery","authors":"D. Lomet","doi":"10.1109/ICDE.1998.655771","DOIUrl":"https://doi.org/10.1109/ICDE.1998.655771","url":null,"abstract":"We describe how to recover applications after system crashes using database recovery. Earlier efforts, based on frequent application checkpoints and/or logging values read, are very expensive. We treat application state as a cached object and log application execution as operations in the recovery framework of D. Lomet and M. Tuttle (1995). Logging application execution does not require logging the application state. Further logged application reads are mostly logical operations in which only the data source identity is logged. We describe a cache manager that handles the flush order dependencies introduced by these log operations and a recovery process that restores application state by replaying the application.","PeriodicalId":264926,"journal":{"name":"Proceedings 14th International Conference on Data Engineering","volume":"25 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1998-02-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125623569","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 1998-02-23DOI: 10.1109/ICDE.1998.655769
L. Lakshmanan, Iyer N. Subramanian, N. Goyal, R. Krishnamurthy
Considers the problem of querying the data in applications such as spreadsheets and word processors. This problem has several motivations from the perspective of data integration, interoperability and OLAP. We provide an architecture for realizing interoperability among such diverse applications and address the challenges that arise specifically in the context of querying data stored in spreadsheet applications. A fundamental challenge is the lack of a well-defined schema. We propose a framework in which the user can specify the layout of data in a spreadsheet, based on his perception of the important concepts underlying that data. Layout specifications can be viewed as the "physical schema" of a spreadsheet. We motivate the concept of an abstract database machine (ADM) that uses the layout specifications to provide a relational view of the data in spreadsheet applications and, similar to a DBMS, supports efficient querying of the spreadsheet data. We develop a methodology for building ADMs for spreadsheets and describe our implementation of an ADM for Microsoft Excel applications, based on the above methodology. Our implementation platform is IBM PCs running Windows NT, Microsoft Office and OLE 2.0. We demonstrate the generality and practicality of our approach by developing a formal characterization of the class of spreadsheets that can be handled in our framework. Our results show that the approach is capable of handling a broad class of naturally occurring spreadsheet applications. This work is part of an office tool integration project.
{"title":"On querying spreadsheets","authors":"L. Lakshmanan, Iyer N. Subramanian, N. Goyal, R. Krishnamurthy","doi":"10.1109/ICDE.1998.655769","DOIUrl":"https://doi.org/10.1109/ICDE.1998.655769","url":null,"abstract":"Considers the problem of querying the data in applications such as spreadsheets and word processors. This problem has several motivations from the perspective of data integration, interoperability and OLAP. We provide an architecture for realizing interoperability among such diverse applications and address the challenges that arise specifically in the context of querying data stored in spreadsheet applications. A fundamental challenge is the lack of a well-defined schema. We propose a framework in which the user can specify the layout of data in a spreadsheet, based on his perception of the important concepts underlying that data. Layout specifications can be viewed as the \"physical schema\" of a spreadsheet. We motivate the concept of an abstract database machine (ADM) that uses the layout specifications to provide a relational view of the data in spreadsheet applications and, similar to a DBMS, supports efficient querying of the spreadsheet data. We develop a methodology for building ADMs for spreadsheets and describe our implementation of an ADM for Microsoft Excel applications, based on the above methodology. Our implementation platform is IBM PCs running Windows NT, Microsoft Office and OLE 2.0. We demonstrate the generality and practicality of our approach by developing a formal characterization of the class of spreadsheets that can be handled in our framework. Our results show that the approach is capable of handling a broad class of naturally occurring spreadsheet applications. This work is part of an office tool integration project.","PeriodicalId":264926,"journal":{"name":"Proceedings 14th International Conference on Data Engineering","volume":"59 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1998-02-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131003815","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 1998-02-23DOI: 10.1109/ICDE.1998.655818
R. Pajarola, T. Ohler, P. Stucki, Kornel Szabo, P. Widmayer
Advocates a desktop virtual reality (VR) interface to a geographic information system (GIS). The navigational capability to explore large topographic scenes is a powerful metaphor and a natural way of interacting with a GIS. VR systems succeed in providing visual realism and real-time navigation and interaction, but fail to cope with very large amounts of data and to provide the general functionality of information systems. We suggest a way to overcome these problems. We describe a prototype system, called ViRGIS (Virtual Reality GIS), that integrates two system platforms: a client that runs the VR component interacts via a (local or wide area) network with a server that runs an object-oriented database containing geographic data. For the purpose of accessing data efficiently, we describe how to integrate a geometric index into the database, and how to perform the operations that are requested in a real-time trip through the virtual world.
{"title":"The Alps at your fingertips: virtual reality and geoinformation systems","authors":"R. Pajarola, T. Ohler, P. Stucki, Kornel Szabo, P. Widmayer","doi":"10.1109/ICDE.1998.655818","DOIUrl":"https://doi.org/10.1109/ICDE.1998.655818","url":null,"abstract":"Advocates a desktop virtual reality (VR) interface to a geographic information system (GIS). The navigational capability to explore large topographic scenes is a powerful metaphor and a natural way of interacting with a GIS. VR systems succeed in providing visual realism and real-time navigation and interaction, but fail to cope with very large amounts of data and to provide the general functionality of information systems. We suggest a way to overcome these problems. We describe a prototype system, called ViRGIS (Virtual Reality GIS), that integrates two system platforms: a client that runs the VR component interacts via a (local or wide area) network with a server that runs an object-oriented database containing geographic data. For the purpose of accessing data efficiently, we describe how to integrate a geometric index into the database, and how to perform the operations that are requested in a real-time trip through the virtual world.","PeriodicalId":264926,"journal":{"name":"Proceedings 14th International Conference on Data Engineering","volume":"18 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1998-02-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123515858","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 1998-02-23DOI: 10.1109/ICDE.1998.655777
Michael H. Böhlen, Renato Busatto, Christian S. Jensen
The association of timestamps with various data items such as tuples or attribute values is fundamental to the management of time varying information. Using intervals in timestamps, as do most data models, leaves a data model with a variety of choices for giving a meaning to timestamps. Specifically, some such data models claim to be point based while other data models claim to be interval based. The meaning chosen for timestamps is important it has a pervasive effect on most aspects of a data model, including database design, a variety of query language properties, and query processing techniques, e.g., the availability of query optimization opportunities. The paper precisely defines the notions of point based and interval based temporal data models, thus providing a new formal basis for characterizing temporal data models and obtaining new insights into the properties of their query languages. Queries in point based models treat snapshot equivalent argument relations identically. This renders point based models insensitive to coalescing. In contrast, queries in interval based models give significance to the actual intervals used in the timestamps, thus generally treating non identical, but possibly snapshot equivalent relations differently. The paper identifies the notion of time fragment preservation as the essential defining property of an interval based data model.
{"title":"Point-versus interval-based temporal data models","authors":"Michael H. Böhlen, Renato Busatto, Christian S. Jensen","doi":"10.1109/ICDE.1998.655777","DOIUrl":"https://doi.org/10.1109/ICDE.1998.655777","url":null,"abstract":"The association of timestamps with various data items such as tuples or attribute values is fundamental to the management of time varying information. Using intervals in timestamps, as do most data models, leaves a data model with a variety of choices for giving a meaning to timestamps. Specifically, some such data models claim to be point based while other data models claim to be interval based. The meaning chosen for timestamps is important it has a pervasive effect on most aspects of a data model, including database design, a variety of query language properties, and query processing techniques, e.g., the availability of query optimization opportunities. The paper precisely defines the notions of point based and interval based temporal data models, thus providing a new formal basis for characterizing temporal data models and obtaining new insights into the properties of their query languages. Queries in point based models treat snapshot equivalent argument relations identically. This renders point based models insensitive to coalescing. In contrast, queries in interval based models give significance to the actual intervals used in the timestamps, thus generally treating non identical, but possibly snapshot equivalent relations differently. The paper identifies the notion of time fragment preservation as the essential defining property of an interval based data model.","PeriodicalId":264926,"journal":{"name":"Proceedings 14th International Conference on Data Engineering","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1998-02-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129580338","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 1998-02-23DOI: 10.1109/ICDE.1998.655761
M. Nakano, H. Imai, M. Kitsuregawa
The distributed shared memory (DSM) architecture is considered to be one of the most likely parallel computing environment candidate for the near future because of its ease of system scalability and facilitation for parallel programming. However, a naive program based on shared memory execution on a DSM machine often deteriorates performance, because of the overhead involved for maintaining cache coherency particularly with frequent remote memory accesses. We show that careful buffer management of parallel join processing on DSM can produce considerable performance improvements in comparison with a naive implementation. We propose four buffer management strategies for parallel hash join processing on the DSM architecture and actually implement them on the HP Exemplar SPP 1600. The basic strategy is to begin with the hash join algorithm for the shared everything architecture and then to consider the memory locality of DSM by distributing the hash table and data pool buffers among the nodes. The results of four buffering strategies are analyzed in detail. Consequently, we can conclude that, in order to achieve high performance on a DSM machine, our buffer management strategy in which the memory access pattern is extracted and buffers are allocated in the local memory of nodes to minimize memory access cost is very efficient.
{"title":"Performance analysis of parallel hash join algorithms on a distributed shared memory machine implementation and evaluation on HP exemplar SPP 1600","authors":"M. Nakano, H. Imai, M. Kitsuregawa","doi":"10.1109/ICDE.1998.655761","DOIUrl":"https://doi.org/10.1109/ICDE.1998.655761","url":null,"abstract":"The distributed shared memory (DSM) architecture is considered to be one of the most likely parallel computing environment candidate for the near future because of its ease of system scalability and facilitation for parallel programming. However, a naive program based on shared memory execution on a DSM machine often deteriorates performance, because of the overhead involved for maintaining cache coherency particularly with frequent remote memory accesses. We show that careful buffer management of parallel join processing on DSM can produce considerable performance improvements in comparison with a naive implementation. We propose four buffer management strategies for parallel hash join processing on the DSM architecture and actually implement them on the HP Exemplar SPP 1600. The basic strategy is to begin with the hash join algorithm for the shared everything architecture and then to consider the memory locality of DSM by distributing the hash table and data pool buffers among the nodes. The results of four buffering strategies are analyzed in detail. Consequently, we can conclude that, in order to achieve high performance on a DSM machine, our buffer management strategy in which the memory access pattern is extracted and buffers are allocated in the local memory of nodes to minimize memory access cost is very efficient.","PeriodicalId":264926,"journal":{"name":"Proceedings 14th International Conference on Data Engineering","volume":"28 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1998-02-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128561130","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 1998-02-23DOI: 10.1109/ICDE.1998.655797
Ling Liu, C. Pu
We describe a family of activity-split and activity-join operations with a notion of validity. The key idea of introducing the set of activity-split and activity-join operations is to allow users to restructure ongoing activities in anticipation of uncertainty so that any significant performance loss due to unexpected unavailablity or delay of shared resources can be avoided or reduced through release of early committed resources or transferring ownership of uncommitted resources. To guarantee the correctness of new activities generated by activity-split or activity-join operations, we define the notion of validity of activity restructuring operations and identify the cases where the correctness is ensured and the cases where activity-split or activity-join are illegal due to the inconsistency incurred.
{"title":"Methodical restructuring of complex workflow activities","authors":"Ling Liu, C. Pu","doi":"10.1109/ICDE.1998.655797","DOIUrl":"https://doi.org/10.1109/ICDE.1998.655797","url":null,"abstract":"We describe a family of activity-split and activity-join operations with a notion of validity. The key idea of introducing the set of activity-split and activity-join operations is to allow users to restructure ongoing activities in anticipation of uncertainty so that any significant performance loss due to unexpected unavailablity or delay of shared resources can be avoided or reduced through release of early committed resources or transferring ownership of uncommitted resources. To guarantee the correctness of new activities generated by activity-split or activity-join operations, we define the notion of validity of activity restructuring operations and identify the cases where the correctness is ensured and the cases where activity-split or activity-join are illegal due to the inconsistency incurred.","PeriodicalId":264926,"journal":{"name":"Proceedings 14th International Conference on Data Engineering","volume":"17 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1998-02-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116749974","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 1998-02-23DOI: 10.1109/ICDE.1998.655752
S. Chawathe, S. Abiteboul, J. Widom
Semistructured data may be irregular and incomplete and does not necessarily conform to a fixed schema. As with structured data, it is often desirable to maintain a history of changes to data, and to query over both the data and the changes. Representing and querying changes in semistructured data is more difficult than in structured data due to the irregularity and lack of schema. We present a model for representing changes in semistructured data and a language for querying over these changes. An important feature of our approach is that we represent and query changes directly as annotations on the affected data, instead of indirectly as the difference between database states. We describe the implementation of our model and query language. We also describe the design and implementation of a query subscription service that permits users to subscribe to changes in semistructured information sources.
{"title":"Representing and querying changes in semistructured data","authors":"S. Chawathe, S. Abiteboul, J. Widom","doi":"10.1109/ICDE.1998.655752","DOIUrl":"https://doi.org/10.1109/ICDE.1998.655752","url":null,"abstract":"Semistructured data may be irregular and incomplete and does not necessarily conform to a fixed schema. As with structured data, it is often desirable to maintain a history of changes to data, and to query over both the data and the changes. Representing and querying changes in semistructured data is more difficult than in structured data due to the irregularity and lack of schema. We present a model for representing changes in semistructured data and a language for querying over these changes. An important feature of our approach is that we represent and query changes directly as annotations on the affected data, instead of indirectly as the difference between database states. We describe the implementation of our model and query language. We also describe the design and implementation of a query subscription service that permits users to subscribe to changes in semistructured information sources.","PeriodicalId":264926,"journal":{"name":"Proceedings 14th International Conference on Data Engineering","volume":"93 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1998-02-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117211514","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 1998-02-23DOI: 10.1109/ICDE.1998.655764
Earl Stahl
Summary form only given, substantially as follows. Compares and contrasts alternative approaches currently in use as methodologies for 'data mining' and knowledge discovery. Specifically, the author focuses on the use of intelligent agents for mining important relationships and patterns, as well as the value of knowledge induction as a basis for quick and accurate discovery of vital information. Beginning with a quick tutorial on how intelligent agent architectures fit within the current n-tier network infrastructure, the author discusses in detail how to build a network or topology of agents representing the important relationships that support the research and discovery efforts.
{"title":"Employing intelligent agents for knowledge discovery","authors":"Earl Stahl","doi":"10.1109/ICDE.1998.655764","DOIUrl":"https://doi.org/10.1109/ICDE.1998.655764","url":null,"abstract":"Summary form only given, substantially as follows. Compares and contrasts alternative approaches currently in use as methodologies for 'data mining' and knowledge discovery. Specifically, the author focuses on the use of intelligent agents for mining important relationships and patterns, as well as the value of knowledge induction as a basis for quick and accurate discovery of vital information. Beginning with a quick tutorial on how intelligent agent architectures fit within the current n-tier network infrastructure, the author discusses in detail how to build a network or topology of agents representing the important relationships that support the research and discovery efforts.","PeriodicalId":264926,"journal":{"name":"Proceedings 14th International Conference on Data Engineering","volume":"6 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1998-02-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121319183","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}