Large-Scale Performance Modeling of Analog and
Download
Report
Transcript Large-Scale Performance Modeling of Analog and
Bayesian Model Fusion: Large-Scale
Performance Modeling of Analog and MixedSignal Circuits by Reusing Early-Stage Data
Fa Wang*, Wangyang Zhang*, Shupeng Sun*, Xin Li*, Chenjie Gu┼
*ECE Dept. Carnegie Mellon University, Pittsburgh, PA 15213
┼ Intel Corp. Hillsboro, OR 97124
Slide 1
Outline
Background
Bayesian Model Fusion
Experiment Results
Conclusion
Slide 2
Process Variations and Performance Modeling
65nm
45nm
32nm
Small Size
Large Variation
Statistical performance modeling: approximate circuit
performance as an analytical function of process variations
f X 1 g1 X 2 g 2 X N g N X
f:
circuit performance of interest (e.g. read delay of SRAM)
∆X:
a vector of random variables to model process variations
gi(∆X): basis functions (e.g., linear or quadratic polynomials)
αi:
model coefficients
Performance model is a powerful tool for efficient circuit
analysis:
Yield estimation
Corner extraction
Sensitivity analysis
Slide 3
Solving Performance Model: Least Squares Fitting (LSF)
Determine performance model
LSF
1
Performance f (X ) am g m (X ) g1 (X ), g 2 (X ),...g M (X )
m
M
Basis functions
Model
coefficients
A set of sampling points are collected
Model coefficients are solved from the following linear equation
Total of M basis
f (1) g1 (X (1) ) g 2 (X (1) )
( 2)
( 2)
( 2)
g
(
X
)
g
(
X
)
f
1
2
(K )
(K )
(K )
f g1 (X ) g 2 (X )
Basis 1
Basis 2
g M (X (1) ) 1
g M (X ( 2) ) 2
(K )
g M (X ) M
Basis M
Total of K
MC samples
The problem is required to be over-determined in order to be
solvable (i.e. K > M)
Slide 4
Challenge: High Dimensionality
High dimensionality becomes a challenge in performance
modeling
Large number of independent random variables must be used to
describe variations in each transistor
Increased number of transistors in circuits
Example: a commercial 32nm CMOS process
~40 random variables to model mismatches of a single transistor
Circuit
Transistor #
Random variable #
Operational amplifier
~ 50
~ 2000
SRAM critical path
~ 10K
~ 400K
Due to high dimensionality (i.e. large # of basis functions), it’s
unrealistic to apply LSF (which requires # of MC samples> #
of basis functions)
Slide 5
Sparsity
To handle the high dimensionality problem, sparsity feature of
circuits has been explored[1]
Sparsity means that the circuit performance variability is only
dominated by a few random variables
Example: In SRAM critical path, many Vth mismatches of
transistors are not important
M
Performance model f (X ) am g m (X ) has a sparse profile:
m 1
Most of coefficients an are zero or close to zero
a1
a
f (X ) [ g1 (X ), g 2 (X ),..., g M (X )] 2
aM
Performance
Model coefficients
Basis functions
[1] X. Li, "Finding deterministic solution from underdetermined equation: large-scale performance modeling of
Slide 6
analog/RF circuits," TCAD, vol. 29, no. 11, pp. 1661-1668, Nov. 2010
Sparse Regression
Sparse regression algorithm is an efficient performance
modeling algorithm that utilizes the sparsity feature
Sparse regression is better than LSF because it requires less
number of samples by using sparsity feature
Efficiency of performance modeling can be further improved,
by considering additional information from design flow (will
be discussed in detail later)
Slide 7
Outline
Background
Bayesian Model Fusion
Experiment Results
Conclusion
Slide 8
Bayesian Model Fusion (BMF): Overview
Key idea: BMF facilitates late stage performance modeling by
reusing data collected in the early stage
Traditional
Performance modeling
Performance modeling
Early stage
data
Late stage
data
Performance modeling
BMF
Proposed
Slide 9
AMS Circuit Design Flow
Analog and mixed-signal (AMS) circuit design spans multiple stages
Design cycle for analog and mixed-signal circuits
Early stage
Schematic
design stage
Performance
modeling
…
Late stage
Layout
design stage
…
Performance
modeling
…
Circuit modeling
Slide 10
Correlation in AMS Design Flow
One important fact in AMS design flow is that different stages
share the same circuit topology and functionality
Comparator:
Schematic stage
Layout stage
Leads to correlation among different stages
Slide 11
Correlation in Performance Models
f E X E1 g1 X E 2 g 2 X EN g N X
f L X L1 g1 X L 2 g 2 X LN g N X
fE(∆X):
fL(∆X):
αEi, αLi :
gi(∆X):
early-stage performance model
late-stage performance model
model coefficients
basis functions
Correlation: fE(∆X) and fL(∆X) are “likely” to be similar
fE(∆X)
g1(∆X)
g2(∆X)
g3(∆X)
g4(∆X)
αE1
αE2
αE3
αE4
…
fL(∆X) αL1 αE1 αL2 αE2 αL3 αE3 αL4 αE4 …
Slide 12
The Proposed Algorithm Flow
Early stage data
Very few late stage data
Prior
Early stage
performance model
f E x
M
E ,m
Bayesian inference
(Proposed)
f L x
g m x
Likelihood
M
L, m
g m x
m 1
m 1
Late stage performance model
Early stage
Late stage
Slide 13
Prior
Prior is a distribution that describes the uncertainty of
parameters based on early stage data, before late stage data
is taken into account
In our work, information in early design stage is encoded in
prior, which describes the uncertainty of late stage model
coefficients
pdf(αL,m)
Prior distribution
αL,m1
αL,m2
Higher
Probability
Lower
Probability
Slide 14
Prior
Magnitude information of early-stage model coefficients is
encoded in prior
Magnitude information here describes whether the absolute
value of coefficient is relatively large or small
Small (or zero) coefficients information represents sparsity
profile, which is essential for performance model[1]
Define prior distribution as a zero-mean Gaussian distribution
Key idea of encoding: the shape of prior is related to magnitude
information
PDF
pdf(αL,1) ~ N(0, 12)
Prior distribution
pdf(αL,2) ~ N(0, 22)
0
αL,1 or αL,2
[1] X. Li, "Finding deterministic solution from underdetermined equation: large-scale performance modeling of
Slide 15
analog/RF circuits," TCAD, vol. 29, no. 11, pp. 1661-1668, Nov. 2010
Likelihood
Likelihood is a function of parameters, which evaluates how
parameters fit with data
Late stage information is encoded in likelihood function
Specifically, late stage performance function information is
encoded in likelihood function
In our work, likelihood function describes how well model
coefficients fit with late stage data
likelihood(αL,m)
Likelihood
αL,m1
αL,m2
Better fit
Worse fit
Slide 16
Maximum-A-Posteriori Estimation
However, if we determine model coefficients solely based on
likelihood, we may have over-fitting problem
In our case, # of samples in late stage is smaller than # of model
coefficients in late stage
Bayesian’s theorem
p( L | FL ) p( L ) p( FL | L )
Posterior
Prior
Likelihood
p( L | FL )
Maximum-a-posteriori (MAP) estimation: max
aL
pdf(αL)
MAP estimation of αL
likelihood(αL)
Prior distribution
Likelihood
Posterior
Slide 17
Outline
Background
Bayesian Model Fusion
Experiment Results
Conclusion
Slide 18
SRAM Example
Example 1: CMOS SRAM
Designed in a commercial 32nm SOI
61572 independent random process parameters are considered
Read delay is considered as performance
Linear performance model is fitted
Experiments run on a 2.5GHz Linux server with 16GB memory
Cell array
Timing logic
WL
Sense amp
Out
Slide 19
Modeling Error
Two different methods are compared:
The proposed method (BMF)
Orthogonal Matching Pursuit (OMP)
Modeling error
Modeling Error (%)
4
OMP (Traditional)
BMF (Proposed)
3
2
4x
1
0
200
400
600
800
Number of Post-Layout Samples
Slide 20
Modeling Time Speed-up
BMF requires 4x less samples to achieve similar accuracy as
OMP in SRAM
4x runtime speed-up to build performance model
OMP
(Traditional)
400
BMF
(Proposed)
100
Read delay error
1.02%
0.99%
Simulation cost (Hour)
38.77
9.69
Fitting cost (Second)
3.56
2.11
Total modeling cost (Hour)
38.77
9.69
Post-layout samples
Slide 21
RO Example
Example 2: CMOS ring oscillator
Designed in a commercial 32nm SOI
7177 independent random process parameters are considered
Power, frequency and phase noise are considered as performance
Linear performance model is fitted
Experiments run on a 2.5GHz Linux server with 16GB memory
Slide 22
Modeling error is measured
for power, frequency and
phase noise
Modeling Error (%)
Modeling Error
3
OMP (Traditional)
BMF (Proposed)
2
Power
9x
1
0
2
1.5
1
OMP (Traditional)
BMF (Proposed)
Frequency
9x
0.5
0
200 400 600 800
Number of Post-Layout Samples
Modeling Error (%)
Modeling Error (%)
200
400
600
800
Number of Post-Layout Samples
0.3
0.2
OMP (Traditional)
BMF (Proposed)
9x
Phase noise
0.1
0
200 400 600 800
Number of Post-Layout Samples
Slide 23
Modeling Time Speed-up
BMF requires 9x less samples to achieve similar accuracy as
OMP in RO
9x runtime speed-up to build performance model
OMP
(Traditional)
900
BMF
(Proposed)
100
Power error
0.77%
0.72%
Frequency error
0.65%
0.54%
Phase noise error
0.12%
0.12%
Simulation cost (Hour)
12.58
1.40
Fitting cost (Second)
5.75
1.69
Total modeling cost (Hour)
12.58
1.40
Post-layout samples
Slide 24
Conclusion
The proposed BMF method facilitates efficient highdimensional performance modeling at late stage by reusing
early stage data
BMF achieves more than 4x runtime speedup over traditional
OMP method on SRAM and RO test cases
BMF can be used for commercial applications such as macromodeling based verification
Slide 25