About cookies on this site Our websites require some cookies to function properly (required). In addition, other cookies may be used with your consent to analyze site usage, improve the user experience and for advertising. For more information, please review your options. By visiting our website, you agree to our processing of information as described in IBM’sprivacy statement. To provide a smooth navigation, your cookie preferences will be shared across the IBM web domains listed here.
Publication
IEEE Topical Meeting EPEPS 2003
Conference paper
The necesity and consequences of modeling driver and load nonlinearity in on-chip global interconnect noise verification
Abstract
The verification of noise in an-chip global interconnect is performed through Simulation of an electrical circuit comprised of a network of coupled transmission lines, terminated by appropriate models for drivers (transmitters) and loads (receivers). The current methodology utilizes linearid models of the terminations, thus requiring only linear circuit simulations. In this study, we show that while a linear noise analysis methodology that relies on the termination model linearization is very efficient and convenient, it may result in significant loss of accuracy and/or in excessively conservative designs. We identdy the situations where modeling the nonlinearity of the termination becomes a determining factor in the accuracy of the analysis. We also study the implications of adopting a fully nonlinear analysis methodology, and propme a practical compromise.