Pub Date : 1998-01-04DOI: 10.1109/ICVD.1998.646574
S. A. Wadekar, A. C. Parker, C. Ravikumar
We present a novel technique to predict energy and power consumption in an electronic system, given its behavioral specification and library components. The early prediction gives circuit designers the freedom to make numerous high-level choices (such as die size, package type, and latency of the pipeline) with confidence that the final implementation will meet power and energy as well as cost and performance constraints. Our unique statistical estimation technique associates low-level, technology dependent physical and electrical parameters with expected circuit resources and interconnect. Further correlations with switching activity yield accurate results consistent with implementations. All feasible designs are investigated using this technique and the designer may trade off between small size, high speed, low energy and low power. The results for designs of two popular signal processing applications, predicted prior to synthesis, are within 10% accuracy of power estimates performed on synthesized layouts.
{"title":"FREEDOM: statistical behavioral estimation of system energy and power","authors":"S. A. Wadekar, A. C. Parker, C. Ravikumar","doi":"10.1109/ICVD.1998.646574","DOIUrl":"https://doi.org/10.1109/ICVD.1998.646574","url":null,"abstract":"We present a novel technique to predict energy and power consumption in an electronic system, given its behavioral specification and library components. The early prediction gives circuit designers the freedom to make numerous high-level choices (such as die size, package type, and latency of the pipeline) with confidence that the final implementation will meet power and energy as well as cost and performance constraints. Our unique statistical estimation technique associates low-level, technology dependent physical and electrical parameters with expected circuit resources and interconnect. Further correlations with switching activity yield accurate results consistent with implementations. All feasible designs are investigated using this technique and the designer may trade off between small size, high speed, low energy and low power. The results for designs of two popular signal processing applications, predicted prior to synthesis, are within 10% accuracy of power estimates performed on synthesized layouts.","PeriodicalId":139023,"journal":{"name":"Proceedings Eleventh International Conference on VLSI Design","volume":"11 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1998-01-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124359308","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 1998-01-04DOI: 10.1109/ICVD.1998.646600
C. Ravikumar, S. Gupta, Akshay Jajoo
With several commercial tools becoming available, the high-level synthesis of application-specific integrated circuits is finding wide spread acceptance in VLSI industry today. Existing tools for synthesis focus on optimizing cost while meeting performance constraints or vice versa. Yet, verification and testing have emerged as major concerns of IC vendors since the repercussions of chips being recalled are far-reaching. In this paper, we concentrate on the synthesis of testable RTL designs using techniques from Artificial Intelligence. We present an adaptive version of the well known Simulated Annealing algorithm and describe its application to a combinatorial optimization problem arising in the high-level synthesis of digital systems. The conventional annealing algorithm was conceived with a single perturb operator which applies a small modification to the existing solution to derive a new solution. The Metropolis criterion is then used to accept or reject the new solution. In some of the complex optimization problems arising in VLSI design, a set of perturb functions become necessary, leading to the question of how to select a particular function for modifying the current system configuration. The adaptive algorithm described here uses the concept of reward and penalty from the theory of learning automata to "learn" to apply the appropriate perturb function. We have applied both the conventional simulated annealing algorithm and the adaptive simulated annealing algorithm to the problem of testability-oriented datapath synthesis for signal processing applications. Our experimental results indicate that the adaptive algorithm can yield better solutions in shorter time.
{"title":"Synthesis of testable RTL designs","authors":"C. Ravikumar, S. Gupta, Akshay Jajoo","doi":"10.1109/ICVD.1998.646600","DOIUrl":"https://doi.org/10.1109/ICVD.1998.646600","url":null,"abstract":"With several commercial tools becoming available, the high-level synthesis of application-specific integrated circuits is finding wide spread acceptance in VLSI industry today. Existing tools for synthesis focus on optimizing cost while meeting performance constraints or vice versa. Yet, verification and testing have emerged as major concerns of IC vendors since the repercussions of chips being recalled are far-reaching. In this paper, we concentrate on the synthesis of testable RTL designs using techniques from Artificial Intelligence. We present an adaptive version of the well known Simulated Annealing algorithm and describe its application to a combinatorial optimization problem arising in the high-level synthesis of digital systems. The conventional annealing algorithm was conceived with a single perturb operator which applies a small modification to the existing solution to derive a new solution. The Metropolis criterion is then used to accept or reject the new solution. In some of the complex optimization problems arising in VLSI design, a set of perturb functions become necessary, leading to the question of how to select a particular function for modifying the current system configuration. The adaptive algorithm described here uses the concept of reward and penalty from the theory of learning automata to \"learn\" to apply the appropriate perturb function. We have applied both the conventional simulated annealing algorithm and the adaptive simulated annealing algorithm to the problem of testability-oriented datapath synthesis for signal processing applications. Our experimental results indicate that the adaptive algorithm can yield better solutions in shorter time.","PeriodicalId":139023,"journal":{"name":"Proceedings Eleventh International Conference on VLSI Design","volume":"58 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1998-01-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116931694","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 1998-01-04DOI: 10.1109/ICVD.1998.646662
G. Swamy, S. Edwards, R. Brayton
In this paper we solve the problem of identifying a "matching" between two logic circuits or "networks". A matching is a functions that maps each gate or "node" in the new circuit into one in the old circuit (if a matching does not exist it maps it to null). We present both an exact and a heuristic way to solve the maximal matching problem. The matching problem does not require any input correspondences. The purpose is to identify structurally identical regions in the networks, and exploit the commonality between them for more efficient verification and synthesis. Synthesis and verification tools that recognize commonalities both between two versions of the same design, as well within a single design, may be able to outperform their counterparts that do not utilize these commonalities. This work is concerned with detecting structural "matchings" that may be re-utilized.
{"title":"Efficient verification and synthesis using design commonalities","authors":"G. Swamy, S. Edwards, R. Brayton","doi":"10.1109/ICVD.1998.646662","DOIUrl":"https://doi.org/10.1109/ICVD.1998.646662","url":null,"abstract":"In this paper we solve the problem of identifying a \"matching\" between two logic circuits or \"networks\". A matching is a functions that maps each gate or \"node\" in the new circuit into one in the old circuit (if a matching does not exist it maps it to null). We present both an exact and a heuristic way to solve the maximal matching problem. The matching problem does not require any input correspondences. The purpose is to identify structurally identical regions in the networks, and exploit the commonality between them for more efficient verification and synthesis. Synthesis and verification tools that recognize commonalities both between two versions of the same design, as well within a single design, may be able to outperform their counterparts that do not utilize these commonalities. This work is concerned with detecting structural \"matchings\" that may be re-utilized.","PeriodicalId":139023,"journal":{"name":"Proceedings Eleventh International Conference on VLSI Design","volume":"24 3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1998-01-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128455404","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 1998-01-04DOI: 10.1109/ICVD.1998.646632
Johnny Öberg, A. Jantsch, Anshul Kumar
The increasing complexity of electronic systems has led to a core based design methodology where large, predesigned blocks can be reused and configured to compose huge systems with several million transistors on a single chip. In a core based methodology, the designer, or the tool, selects an appropriate core from a library and adapts it to the surrounding environment. However, a specific core configuration in a specific environment very often opens up the potential for performance and cost optimizations which today cannot be utilized due to the inflexibility of cores. In this paper we propose an object oriented concept for a library of cores, which not only contains design models of the cores but also rule based design transformations to optimize the core in the context of a specific configuration and a specific environment.
{"title":"An object-oriented concept for intelligent library functions","authors":"Johnny Öberg, A. Jantsch, Anshul Kumar","doi":"10.1109/ICVD.1998.646632","DOIUrl":"https://doi.org/10.1109/ICVD.1998.646632","url":null,"abstract":"The increasing complexity of electronic systems has led to a core based design methodology where large, predesigned blocks can be reused and configured to compose huge systems with several million transistors on a single chip. In a core based methodology, the designer, or the tool, selects an appropriate core from a library and adapts it to the surrounding environment. However, a specific core configuration in a specific environment very often opens up the potential for performance and cost optimizations which today cannot be utilized due to the inflexibility of cores. In this paper we propose an object oriented concept for a library of cores, which not only contains design models of the cores but also rule based design transformations to optimize the core in the context of a specific configuration and a specific environment.","PeriodicalId":139023,"journal":{"name":"Proceedings Eleventh International Conference on VLSI Design","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1998-01-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129574489","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 1998-01-04DOI: 10.1109/ICVD.1998.646648
D. Saha, A. Chandrakasan
Emerging "systems-on-a-chip" will require a design environment that allows distributed access to libraries, models and design tools. In this paper we present a framework using the Object-Web technologies to implement Web-based CAD. The framework includes the infrastructure to store and manipulate design objects, protocols for tool communication and WebTop, a Java hierarchical schematic/block editor with interfaces to distributed Web tools and cell libraries.
{"title":"Web-based distributed VLSI design","authors":"D. Saha, A. Chandrakasan","doi":"10.1109/ICVD.1998.646648","DOIUrl":"https://doi.org/10.1109/ICVD.1998.646648","url":null,"abstract":"Emerging \"systems-on-a-chip\" will require a design environment that allows distributed access to libraries, models and design tools. In this paper we present a framework using the Object-Web technologies to implement Web-based CAD. The framework includes the infrastructure to store and manipulate design objects, protocols for tool communication and WebTop, a Java hierarchical schematic/block editor with interfaces to distributed Web tools and cell libraries.","PeriodicalId":139023,"journal":{"name":"Proceedings Eleventh International Conference on VLSI Design","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1998-01-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129599502","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 1998-01-04DOI: 10.1109/ICVD.1998.646628
R. Pillai, D. Al-Khalili, A. Al-Khalili
In CMOS logic implementations, the architectural/algorithmic power/delay/area implications of functional units are crucial as far as design economies of the target application are concerned. This paper addresses the architectural design of a low power floating point accumulator by using a transition activity scaled triple data path floating point adder core. The proposed scheme offers a worst case power reduction of 50% in comparison to schemes that use conventional floating point adders. The reduction in power delay product is better than 3X.
{"title":"A low power floating point accumulator","authors":"R. Pillai, D. Al-Khalili, A. Al-Khalili","doi":"10.1109/ICVD.1998.646628","DOIUrl":"https://doi.org/10.1109/ICVD.1998.646628","url":null,"abstract":"In CMOS logic implementations, the architectural/algorithmic power/delay/area implications of functional units are crucial as far as design economies of the target application are concerned. This paper addresses the architectural design of a low power floating point accumulator by using a transition activity scaled triple data path floating point adder core. The proposed scheme offers a worst case power reduction of 50% in comparison to schemes that use conventional floating point adders. The reduction in power delay product is better than 3X.","PeriodicalId":139023,"journal":{"name":"Proceedings Eleventh International Conference on VLSI Design","volume":"30 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1998-01-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133603812","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 1998-01-04DOI: 10.1109/ICVD.1998.646603
D. K. Das, Indrajit Chaudhuri, B. Bhattacharya
A novel design of a test pattern generator (TPG) for built-in self-testing (BIST) of path delay faults, is proposed. For an n-input CUT, the TPG generates a sequence of length (n.2/sup n/+1), that includes all n.2/sup n/ single-input-change (SIC) test pairs, and hence optimal. The generation of such a sequence of minimum length (i.e., n.2/sup n/+1) was an open problem. A simple iterative circuit of the TPG is then constructed. This provides minimum test application time for testing path delay faults, and compares favorably with the earlier BIST designs.
{"title":"Design of an optimal test pattern generator for built-in self testing of path delay faults","authors":"D. K. Das, Indrajit Chaudhuri, B. Bhattacharya","doi":"10.1109/ICVD.1998.646603","DOIUrl":"https://doi.org/10.1109/ICVD.1998.646603","url":null,"abstract":"A novel design of a test pattern generator (TPG) for built-in self-testing (BIST) of path delay faults, is proposed. For an n-input CUT, the TPG generates a sequence of length (n.2/sup n/+1), that includes all n.2/sup n/ single-input-change (SIC) test pairs, and hence optimal. The generation of such a sequence of minimum length (i.e., n.2/sup n/+1) was an open problem. A simple iterative circuit of the TPG is then constructed. This provides minimum test application time for testing path delay faults, and compares favorably with the earlier BIST designs.","PeriodicalId":139023,"journal":{"name":"Proceedings Eleventh International Conference on VLSI Design","volume":"32 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1998-01-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133981245","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 1998-01-04DOI: 10.1109/ICVD.1998.646633
Dong-Hyun Heo, A. C. Parker, C. Ravikumar
This paper considers the system redesign problem where an existing digital electronic system must be redesigned for reasons such as better performance, technology upgrade, lower power, and so on. The designer may wish to retain certain design for purely, technical reasons or reasons of propriety. We presented an algorithm based on the evolutionary paradigm for optimally redesigning an existing design. A tool called GRAFT which implements the evolutionary algorithm is described.
{"title":"An evolutionary approach to system redesign","authors":"Dong-Hyun Heo, A. C. Parker, C. Ravikumar","doi":"10.1109/ICVD.1998.646633","DOIUrl":"https://doi.org/10.1109/ICVD.1998.646633","url":null,"abstract":"This paper considers the system redesign problem where an existing digital electronic system must be redesigned for reasons such as better performance, technology upgrade, lower power, and so on. The designer may wish to retain certain design for purely, technical reasons or reasons of propriety. We presented an algorithm based on the evolutionary paradigm for optimally redesigning an existing design. A tool called GRAFT which implements the evolutionary algorithm is described.","PeriodicalId":139023,"journal":{"name":"Proceedings Eleventh International Conference on VLSI Design","volume":"120 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1998-01-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123108259","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 1998-01-04DOI: 10.1109/ICVD.1998.646604
Subhasish Subhasish, P. Banerjee, M. Sarrafzadeh
A popular approach to reduce power consumption is to identify self-loops in a state transition graph (STG) of a finite state machine (FSM) followed by gating the clock with a suitable function to power down the circuit (implementation of the FSM) during the self-loop cycles. Although this approach is effective in circuits with plenty of self-loops, it fails for FSMs without self-loops. Since self loops may not be inherently present in a given FSM, we decompose it into interacting FSMs such that they have plenty of self-loops. In this paper we present a novel partitioning algorithm to decompose a given finite state machine. By using this approach, we could save upto 71% of the total power on circuits like fetch, where other techniques could not save any power.
{"title":"Partitioning sequential circuits for low power","authors":"Subhasish Subhasish, P. Banerjee, M. Sarrafzadeh","doi":"10.1109/ICVD.1998.646604","DOIUrl":"https://doi.org/10.1109/ICVD.1998.646604","url":null,"abstract":"A popular approach to reduce power consumption is to identify self-loops in a state transition graph (STG) of a finite state machine (FSM) followed by gating the clock with a suitable function to power down the circuit (implementation of the FSM) during the self-loop cycles. Although this approach is effective in circuits with plenty of self-loops, it fails for FSMs without self-loops. Since self loops may not be inherently present in a given FSM, we decompose it into interacting FSMs such that they have plenty of self-loops. In this paper we present a novel partitioning algorithm to decompose a given finite state machine. By using this approach, we could save upto 71% of the total power on circuits like fetch, where other techniques could not save any power.","PeriodicalId":139023,"journal":{"name":"Proceedings Eleventh International Conference on VLSI Design","volume":"191 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1998-01-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124277202","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 1998-01-04DOI: 10.1109/ICVD.1998.646639
H. Selvaraj, M. Nowicka, T. Luba
Existing FPGA-oriented algorithms can be divided into two categories: minimising the number of LUTs in the solution (MIS-pga, Trade, and ASYL); and minimising the delay in the solution (DAG-Map, SWEEP, and Flow-map). Several algorithms have been implemented with both area and delay minimisation versions, for example MIS-pga and ASYL. Two collaborating groups from Warsaw University of Technology, Poland and Monash University, Australia have developed decomposition theory and procedures for single and multiple-output Boolean functions. These include a balanced decomposition algorithm which applies either parallel or serial decomposition at each phase of the synthesis process. The algorithm has been implemented in an experimental logic synthesis tool DEMAIN. Recent tests on MCNC and industrial benchmarks show that DEMAIN produces much more economical designs than major FPGA vendors software.
{"title":"Decomposition strategies and their performance in FPGA-based technology mapping","authors":"H. Selvaraj, M. Nowicka, T. Luba","doi":"10.1109/ICVD.1998.646639","DOIUrl":"https://doi.org/10.1109/ICVD.1998.646639","url":null,"abstract":"Existing FPGA-oriented algorithms can be divided into two categories: minimising the number of LUTs in the solution (MIS-pga, Trade, and ASYL); and minimising the delay in the solution (DAG-Map, SWEEP, and Flow-map). Several algorithms have been implemented with both area and delay minimisation versions, for example MIS-pga and ASYL. Two collaborating groups from Warsaw University of Technology, Poland and Monash University, Australia have developed decomposition theory and procedures for single and multiple-output Boolean functions. These include a balanced decomposition algorithm which applies either parallel or serial decomposition at each phase of the synthesis process. The algorithm has been implemented in an experimental logic synthesis tool DEMAIN. Recent tests on MCNC and industrial benchmarks show that DEMAIN produces much more economical designs than major FPGA vendors software.","PeriodicalId":139023,"journal":{"name":"Proceedings Eleventh International Conference on VLSI Design","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1998-01-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130127530","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}