IISc Logo    Title

etd AT Indian Institute of Science >
Division of Physical and Mathematical Sciences  >
Instrumentation (isu) >

Please use this identifier to cite or link to this item: http://hdl.handle.net/2005/692

Title: A New Method To Determine Optimal Time-Delays Between Switching Of Digital VLSI Circuits To Minimize Power Supply Noise
Authors: Srinivasan, G
Advisors: Mathias, P C
Keywords: Electric Power System
VLSI Circuits
Power Distribution Network (PDN)
Integrated Circuits - Very Large Scale Integration
VLSI Systems - Noise Reduction
Power Supply Noise
Submitted Date: Jun-2007
Series/Report no.: G21109
Abstract: Power supply noise, which is the variation in the supply voltage across the on-die supply terminals of VLSI circuits, is a serious performance degrader in digital circuits and mixed analog-digital circuits. In digital VLSI systems, power supply noise causes timing errors such as delays, jitter, and false switching. In microprocessors, power supply noise reduces the maximum operating frequency (FMAX) of the CPU. In mixed analog-digital circuits, power supply noise manifests as the substrate noise and impairs the performance of the analog portion. The decrease in the available noise margin with the decrease in the feature size of transistors in CMOS systems makes the power supply noise a very serious issue, and demands new methods to reduce the power supply noise in sub-micron CMOS systems. In this thesis, we develop a new method to determine optimal time-delays between the switching of input/output (I/O) data buffers in digital VLSI systems that realizes maximum reduction of the power supply noise. We first discuss methods to characterize the distributed nature of the Power Delivery Network (PDN) in the frequency-domain. We then develop an analytical method to determine the optimal delays using the frequency-domain response of the PDN and the supply current spectrum of the buffer units. We explain the mechanism behind the cancellation of the power supply noise by the introduction of optimal buffer-to-buffer delays. We also develop a numerical method to determine the optimal delays and compare it with the analytical method. We illustrate the reduction in the power supply noise by applying the optimal time-delays determined using our methods to two examples of PDN. Our method has great potential to realize maximum reduction of power supply noise in digital VLSI circuits and substrate noise in mixed analog-digital VLSI circuits. Lower power supply noise translates into lower cost and improved performance of the circuit.
URI: http://hdl.handle.net/2005/692
Appears in Collections:Instrumentation (isu)

Files in This Item:

File Description SizeFormat
G21109.pdf945.88 kBAdobe PDFView/Open

Items in etd@IISc are protected by copyright, with all rights reserved, unless otherwise indicated.

 

etd@IISc is a joint service of NCSI & IISc Library ||
Feedback
|| Powered by DSpace || Compliant to OAI-PMH V 2.0 and ETD-MS V 1.01