An insurance underwriting expert system is designed and implemented around an existing mainframe program, driven by a sizable old-fashioned database. It is shown how modeling helped not only to identify real user needs but also to untangle many accumulated complications in the existing repository of information. The author also highlights which conceptual modeling mechanisms are most needed to support such tasks.<>
{"title":"Experience in applying conceptual modelling to interface with a real-life business application","authors":"M. Pilote","doi":"10.1109/ICDE.1989.47212","DOIUrl":"https://doi.org/10.1109/ICDE.1989.47212","url":null,"abstract":"An insurance underwriting expert system is designed and implemented around an existing mainframe program, driven by a sizable old-fashioned database. It is shown how modeling helped not only to identify real user needs but also to untangle many accumulated complications in the existing repository of information. The author also highlights which conceptual modeling mechanisms are most needed to support such tasks.<<ETX>>","PeriodicalId":329505,"journal":{"name":"[1989] Proceedings. Fifth International Conference on Data Engineering","volume":"273 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1989-02-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116214190","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Join algorithms on KD-tree indexed relations are proposed. The join algorithms are based on a concept called wave. The wave is a set of pages that is the object of joining and that propagates over the relation space in the direction of the join attribute axis. Four basic join algorithms that determine the wave from one of the relations and one algorithm that determines the wave from both relations are proposed. The algorithms are described and extensively analyzed with analytical formulas and simulation results. Then a garbage collection mechanism is introduced that discards the unnecessary data loaded in the main memory and extends the previous basic algorithms with an efficient memory management. It is shown that the proposed algorithms perform the join of very large relations with one scan.<>
{"title":"Join strategies on KD-tree indexed relations","authors":"M. Kitsuregawa, L. Harada, M. Takagi","doi":"10.1109/ICDE.1989.47203","DOIUrl":"https://doi.org/10.1109/ICDE.1989.47203","url":null,"abstract":"Join algorithms on KD-tree indexed relations are proposed. The join algorithms are based on a concept called wave. The wave is a set of pages that is the object of joining and that propagates over the relation space in the direction of the join attribute axis. Four basic join algorithms that determine the wave from one of the relations and one algorithm that determines the wave from both relations are proposed. The algorithms are described and extensively analyzed with analytical formulas and simulation results. Then a garbage collection mechanism is introduced that discards the unnecessary data loaded in the main memory and extends the previous basic algorithms with an efficient memory management. It is shown that the proposed algorithms perform the join of very large relations with one scan.<<ETX>>","PeriodicalId":329505,"journal":{"name":"[1989] Proceedings. Fifth International Conference on Data Engineering","volume":"106 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1989-02-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131714099","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
A protocol is presented for negotiating access to data in a federated database. This protocol deals with several important aspects of data sharing in an environment where databases belonging to different organizations coexist and cooperate. It is based on the concept of quasicopies. These are cached values that are allowed to deviate from the central value in a controlled fashion. The degree of consistency of a quasicopy is established by the entity where the quasicopy is to reside. Two techniques are presented that can be used for numerical data and for copies that rely on version numbers. It is believed that these two ideas encompass many of the interesting cases in practice. However, it is also possible to fine tune the estimation of this function once the exact semantics of the data involved are known.<>
{"title":"Negotiating data access in federated database systems","authors":"R. Alonso, Daniel Barbará","doi":"10.1109/ICDE.1989.47200","DOIUrl":"https://doi.org/10.1109/ICDE.1989.47200","url":null,"abstract":"A protocol is presented for negotiating access to data in a federated database. This protocol deals with several important aspects of data sharing in an environment where databases belonging to different organizations coexist and cooperate. It is based on the concept of quasicopies. These are cached values that are allowed to deviate from the central value in a controlled fashion. The degree of consistency of a quasicopy is established by the entity where the quasicopy is to reside. Two techniques are presented that can be used for numerical data and for copies that rely on version numbers. It is believed that these two ideas encompass many of the interesting cases in practice. However, it is also possible to fine tune the estimation of this function once the exact semantics of the data involved are known.<<ETX>>","PeriodicalId":329505,"journal":{"name":"[1989] Proceedings. Fifth International Conference on Data Engineering","volume":"99 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1989-02-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131813109","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
The database analyzer and predictor (DBAP) is a CASE (computer-aided software engineering) tool supporting the design and tuning of IDMS (integrated data-management system) DB/DC applications. It is an interactive system, completely integrated with the IDMS online environment and providing the database designer with a comprehensive set of performance analysis models evaluating principal aspects of the database system performance. Close coupling of the workload analysis models with the computer system model provides a means for predicting the response-time behavior of the modeled database system. The performance-oriented database methodology, supported by a menu-driven user interface, guides the users through the database design, as well as existing database system tuning and capacity-planning projects.<>
{"title":"Database analyzer and predictor-an overview","authors":"S. Orlando, V. Perri, S. Scrivano, W. Staniszkis","doi":"10.1109/ICDE.1989.47270","DOIUrl":"https://doi.org/10.1109/ICDE.1989.47270","url":null,"abstract":"The database analyzer and predictor (DBAP) is a CASE (computer-aided software engineering) tool supporting the design and tuning of IDMS (integrated data-management system) DB/DC applications. It is an interactive system, completely integrated with the IDMS online environment and providing the database designer with a comprehensive set of performance analysis models evaluating principal aspects of the database system performance. Close coupling of the workload analysis models with the computer system model provides a means for predicting the response-time behavior of the modeled database system. The performance-oriented database methodology, supported by a menu-driven user interface, guides the users through the database design, as well as existing database system tuning and capacity-planning projects.<<ETX>>","PeriodicalId":329505,"journal":{"name":"[1989] Proceedings. Fifth International Conference on Data Engineering","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1989-02-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130048211","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
A family of practical algorithms is presented to schedule join execution in a shared-memory multiprocessor environment. The algorithms are based on page connectivity graphs and determine when to read each data page into memory and how to schedule page joins on the available processors. The goal is to overlap page reads with parallel join execution in such a way that both the number of processors and total duration of join processing time are minimized. Upper and lower bounds are derived on the number of processors required to complete join execution in optimal time. A description is given of a general strategy for generating read schedules that it is conjectured can be processed in minimal time (over all read schedules on any number of processors) and a family of practical algorithms utilizing an arbitrary number of lookahead steps to approximate this general strategy.<>
{"title":"Processor scheduling for multiprocessor joins","authors":"M. C. Murphy, D. Rotem","doi":"10.1109/ICDE.1989.47209","DOIUrl":"https://doi.org/10.1109/ICDE.1989.47209","url":null,"abstract":"A family of practical algorithms is presented to schedule join execution in a shared-memory multiprocessor environment. The algorithms are based on page connectivity graphs and determine when to read each data page into memory and how to schedule page joins on the available processors. The goal is to overlap page reads with parallel join execution in such a way that both the number of processors and total duration of join processing time are minimized. Upper and lower bounds are derived on the number of processors required to complete join execution in optimal time. A description is given of a general strategy for generating read schedules that it is conjectured can be processed in minimal time (over all read schedules on any number of processors) and a family of practical algorithms utilizing an arbitrary number of lookahead steps to approximate this general strategy.<<ETX>>","PeriodicalId":329505,"journal":{"name":"[1989] Proceedings. Fifth International Conference on Data Engineering","volume":"20 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1989-02-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115130301","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
An approach for evolutionary database design is presented which tries to remedy some of the shortcomings of previous design methods. The approach distinguishes clearly between a conceptual and a logical database design. A conceptual schema models the relevant aspects of reality. A logical schema describes the structure of the database as generic tables, and it reflects the design decisions taken to map the objects of the conceptual schema into the generic tables. To support this strategy with tools, it is necessary to have a version concept and a mechanism for recording design decisions called protocolling. These concepts are realized in Presto, a development environment for an evolutionary design of database applications which is described.<>
{"title":"Evolutionary database design","authors":"Fredy Oertly, G. Schiller","doi":"10.1109/ICDE.1989.47269","DOIUrl":"https://doi.org/10.1109/ICDE.1989.47269","url":null,"abstract":"An approach for evolutionary database design is presented which tries to remedy some of the shortcomings of previous design methods. The approach distinguishes clearly between a conceptual and a logical database design. A conceptual schema models the relevant aspects of reality. A logical schema describes the structure of the database as generic tables, and it reflects the design decisions taken to map the objects of the conceptual schema into the generic tables. To support this strategy with tools, it is necessary to have a version concept and a mechanism for recording design decisions called protocolling. These concepts are realized in Presto, a development environment for an evolutionary design of database applications which is described.<<ETX>>","PeriodicalId":329505,"journal":{"name":"[1989] Proceedings. Fifth International Conference on Data Engineering","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1989-02-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124964719","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
The second version of FDS-R (functional disk system with relational database engine), FDS-RII, which is designed to handle large relations efficiently, is discussed. On FDS-RII, the processing algorithm is selected at run time from two algorithms (nested loop algorithms, grace hash algorithm) by comparing their estimated I/O costs. The processing strategy is discussed in detail. The I/O cost formula is examined by measuring the execution time of a join query on the FDS-RII. With the expanded version of Wisconsin Benchmark, the performance of FDS-RII is measured. FDS-RII attained a high performance level for large relations as compared to other large database systems such as Gamma and Teradata. While FDS uses just one disk and three MC68020s, Teradata uses 40 disks and 20 AMPs and Gamma requires eight disks and 17 VAX 11/750s.<>
讨论了基于关系数据库引擎的功能磁盘系统(functional disk system with relational database engine, FDS-RII)的第二版——FDS-RII。在FDS-RII上,通过比较估计的I/O成本,在运行时从两种算法(嵌套循环算法、grace哈希算法)中选择处理算法。详细讨论了加工策略。通过测量FDS-RII上连接查询的执行时间来检查I/O成本公式。通过威斯康星基准的扩展版本,对FDS-RII的性能进行了测量。与Gamma和Teradata等其他大型数据库系统相比,FDS-RII在处理大型关系方面达到了很高的性能水平。FDS只需要一个磁盘和三个mc68020, Teradata需要40个磁盘和20个amp, Gamma需要8个磁盘和17个VAX 11/750。
{"title":"Query execution for large relations on functional disk system","authors":"M. Kitsuregawa, M. Nakano, M. Takagi","doi":"10.1109/ICDE.1989.47211","DOIUrl":"https://doi.org/10.1109/ICDE.1989.47211","url":null,"abstract":"The second version of FDS-R (functional disk system with relational database engine), FDS-RII, which is designed to handle large relations efficiently, is discussed. On FDS-RII, the processing algorithm is selected at run time from two algorithms (nested loop algorithms, grace hash algorithm) by comparing their estimated I/O costs. The processing strategy is discussed in detail. The I/O cost formula is examined by measuring the execution time of a join query on the FDS-RII. With the expanded version of Wisconsin Benchmark, the performance of FDS-RII is measured. FDS-RII attained a high performance level for large relations as compared to other large database systems such as Gamma and Teradata. While FDS uses just one disk and three MC68020s, Teradata uses 40 disks and 20 AMPs and Gamma requires eight disks and 17 VAX 11/750s.<<ETX>>","PeriodicalId":329505,"journal":{"name":"[1989] Proceedings. Fifth International Conference on Data Engineering","volume":"53 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1989-02-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130480732","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
The problem of efficiently processing recursive path queries in deductive database systems is discussed, and a semimaterialized encoding structure is proposed as an attractive approach that provides a balance between efficiency of retrieval and feasibility of storage. Incremental algorithms are presented that enable the effects of updates to the underlying database to be reflected in the materialized information. Performance simulations indicate that these techniques can significantly speed up the processing of path queries at an acceptable level of storage overhead.<>
{"title":"Materialization and incremental update of path information","authors":"R. Agrawal, H. Jagadish","doi":"10.1109/ICDE.1989.47238","DOIUrl":"https://doi.org/10.1109/ICDE.1989.47238","url":null,"abstract":"The problem of efficiently processing recursive path queries in deductive database systems is discussed, and a semimaterialized encoding structure is proposed as an attractive approach that provides a balance between efficiency of retrieval and feasibility of storage. Incremental algorithms are presented that enable the effects of updates to the underlying database to be reflected in the materialized information. Performance simulations indicate that these techniques can significantly speed up the processing of path queries at an acceptable level of storage overhead.<<ETX>>","PeriodicalId":329505,"journal":{"name":"[1989] Proceedings. Fifth International Conference on Data Engineering","volume":"64 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1989-02-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126455337","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
A nonnumeric analog of statistical methods for proving security is described and implemented. The approach involves analyzing facts and inference rules assumed to be known to a compromiser and deriving all their possible consequences using resolution theorem-proving, a technique which is argued to be far more appropriate to this problem than rule-based expert systems or information-flow analysis. An important contribution of this method is augmentation of resolution to handle associated time intervals and probabilities of statements being true. The augmentation is simple to use by domain experts untrained in computers, and it is believed to provide the first truly practical tool for analysis of indirect logical inferences in information systems. Its capabilities are demonstrated with an example from military security.<>
{"title":"Inference-security analysis using resolution theorem-proving","authors":"N. Rowe","doi":"10.1109/ICDE.1989.47242","DOIUrl":"https://doi.org/10.1109/ICDE.1989.47242","url":null,"abstract":"A nonnumeric analog of statistical methods for proving security is described and implemented. The approach involves analyzing facts and inference rules assumed to be known to a compromiser and deriving all their possible consequences using resolution theorem-proving, a technique which is argued to be far more appropriate to this problem than rule-based expert systems or information-flow analysis. An important contribution of this method is augmentation of resolution to handle associated time intervals and probabilities of statements being true. The augmentation is simple to use by domain experts untrained in computers, and it is believed to provide the first truly practical tool for analysis of indirect logical inferences in information systems. Its capabilities are demonstrated with an example from military security.<<ETX>>","PeriodicalId":329505,"journal":{"name":"[1989] Proceedings. Fifth International Conference on Data Engineering","volume":"70 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1989-02-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126646481","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
A method is proposed to store and efficiently retrieve rules in a deductive database. The rules are compiled into an execution model called a PCN (production compilation network). The PCN is stored as a particular relational database which consists of three relations. For a given end-user query the first step of the inference process consists of searching for the relevant rules by traversing the PCN structure. This traversal is computed by a transitive closure of a part of the PCN. The transitive closure is performed by a loop of joins over the relations storing the PCN. In this context, a transitive closure algorithm that is based on a physical clustering of the relations is proposed. This clustering consists of a double hashing technique. It ensures a linear cost for the join operation with weak conditions on the size of the main memory. The approach reduces the number of I/O operations for semi-naive transitive closure operation. This algorithm will be applied to the search for relevant rules and to the transitive computation of linear recursive relations.<>
{"title":"Relational storage and efficient retrieval of rules in a deductive DBMS","authors":"J. Cheiney, C. D. Maindreville","doi":"10.1109/ICDE.1989.47272","DOIUrl":"https://doi.org/10.1109/ICDE.1989.47272","url":null,"abstract":"A method is proposed to store and efficiently retrieve rules in a deductive database. The rules are compiled into an execution model called a PCN (production compilation network). The PCN is stored as a particular relational database which consists of three relations. For a given end-user query the first step of the inference process consists of searching for the relevant rules by traversing the PCN structure. This traversal is computed by a transitive closure of a part of the PCN. The transitive closure is performed by a loop of joins over the relations storing the PCN. In this context, a transitive closure algorithm that is based on a physical clustering of the relations is proposed. This clustering consists of a double hashing technique. It ensures a linear cost for the join operation with weak conditions on the size of the main memory. The approach reduces the number of I/O operations for semi-naive transitive closure operation. This algorithm will be applied to the search for relevant rules and to the transitive computation of linear recursive relations.<<ETX>>","PeriodicalId":329505,"journal":{"name":"[1989] Proceedings. Fifth International Conference on Data Engineering","volume":"46 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1989-02-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126625596","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}