A Machine-Learning Approach to Analog/RF Circuit Testing* Yiorgos - - PowerPoint PPT Presentation

a machine learning approach to analog rf circuit testing
SMART_READER_LITE
LIVE PREVIEW

A Machine-Learning Approach to Analog/RF Circuit Testing* Yiorgos - - PowerPoint PPT Presentation

A Machine-Learning Approach to Analog/RF Circuit Testing* Yiorgos Makris Departments of Electrical Engineering & Computer Science YALE UNIVERSITY * Joint work with Dr. Haralampos Stratigopoulos (TIMA Lab, Grenoble, France), Prof. Petros


slide-1
SLIDE 1

A Machine-Learning Approach to Analog/RF Circuit Testing*

Yiorgos Makris

Departments of Electrical Engineering & Computer Science

YALE UNIVERSITY

* Joint work with Dr. Haralampos Stratigopoulos (TIMA Lab, Grenoble, France),

  • Prof. Petros Drineas (RPI) and Dr. Mustapha Slamani (IBM).

Partially funded by NSF, SRC/IBM, TI, and the European Commission.

slide-2
SLIDE 2

Test and Reliability @ YALE

http://eng.yale.edu/trela

slide-3
SLIDE 3
  • Machine learning-based testing
  • Correlation mining for post-fabrication tuning and yield enhancement
  • Design of on-chip checkers and on-line test methods
  • Hardware Trojan detection in wireless cryptographic circuits

Research Areas

  • Analog/RF circuits
  • Workload-driven error impact analysis in modern microprocessors
  • Logic transformations for improved soft-error immunity
  • Concurrent error detection / correction methods for FSMs
  • Digital Circuits
  • Fault simulation & test generation for speed-independent circuits
  • Test methods for high-speed pipelines (e.g. Mousetrap)
  • Error detection and soft-error mitigation in burst-mode controllers
  • Asynchronous circuits
slide-4
SLIDE 4

Presentation Outline

  • Testing analog/RF circuits
  • Summary
  • Testing via a non-linear neural classifier
  • Construction and training
  • Selection of measurements
  • Test quality vs. test time trade-off
  • Machine learning-based testing
  • Experimental results
  • Analog/RF specification test compaction
  • Stand-alone Built-in Self-Test (BIST)
  • Yield enhancement
  • Performance calibration via post-fabrication tuning
slide-5
SLIDE 5

Definition of Analog/RF Functionality

Transistor-Level Symbol Specifications

slide-6
SLIDE 6

Design

Analog/RF IC Testing - Problem Definition

Chip

Layout

Actual silicon performances

Measurement

  • Targets manufacturing defects
  • Once per chip

Testing:

Pre-layout or post-layout performances

Simulation

  • Targets design errors
  • Once per design

Verification:

Comparison

Fabrication

slide-7
SLIDE 7

Analog/RF IC Test - Industrial Practice

Interface Board Wafer

  • Post-silicon production flow

Automatic Test Equipment (ATE) Die

design specifications performance parameters compare

pass/fail

  • Current practice is specification testing

chip

test configurations ATE

slide-8
SLIDE 8

Limitations

  • Expensive ATE (multi-million dollar equipment)
  • Specialized circuitry for stimuli generation and

response measurement

Test Cost:

  • Multiple measurements and test configurations
  • Switching and settling times

Test Time:

  • Fault-model based test – Never really caught on
  • Machine learning-based (a.k.a. “alternate”) testing

– Regression (Variyam et al., TCAD’02) – Classification (Pan et al., TCAS-II’99, Lindermeir et al., TCAD’99)

Alternatives?

slide-9
SLIDE 9

Machine Learning-Based Testing

  • Determine if a chip meets its specifications without

explicitly computing the performance parameters and without assuming a prescribed fault model

General idea:

  • Infer whether the specifications are violated through a

few simpler/cheaper measurements and information that is “learned” from a set of fully tested chips

How does it work?

  • Since chips are produced by the same manufacturing

process, the relation between measurements and performance parameters can be statistically learned

Underlying assumption:

slide-10
SLIDE 10

Regression vs. Classification

Problem Definition:

simple functions performance parameters π specification tests (T1, … , Tk) design specifications compare pass/fail unknown,complex, non-linear functions (no closed-form) alternate tests (x1, … , xn)

Use machine-learning to approximate these functions

  • Explicitly learn

these functions (i.e. approximate f:x → π)

Regression:

  • Implicitly learn

these functions (i.e. approximate f:x→Y,Y={pass/fail})

Classification:

slide-11
SLIDE 11

Overview of Classification Approach

specification tests pass/fail labels training set of chips projection on measurement space

  • 0.025
  • 0.02
  • 0.015
  • 0.01
  • 0.005

0.005 0.01 0.015 0.0

  • 0.015
  • 0.01
  • 0.005

0.005 0.01 0.015 0.02

x1 x2

Nominal patterns Faulty patterns

simpler measurements acquisition of measurement patterns trained classifier new (untested) chip measurement pattern pass/fail label

TESTING

learn boundary

LEARNING

slide-12
SLIDE 12

Using a Non-Linear Neural Classifier

  • Allocates a single boundary of arbitrary order
  • No prior knowledge of boundary order is required
  • The topology is not fixed, but it grows (ontogeny)

until it matches the intrinsic complexity of the separation problem

  • Constructed using linear perceptrons only
slide-13
SLIDE 13

Linear Perceptron

1

x

d

x

i

y 1 =

  • x

i

w

1

i

w

d

i

w

perceptron connectivity synapses

= d j j

x

j

i

w

i

y (x)

1

  • 1

perceptron output threshold activation function geometric interpretation

1

x

2

x

= d j j

x

j

i

w = 0

1 Nominal patterns Faulty patterns

i

y (x) =

for nominal

  • 1

i

y (x) =

for faulty

adjusts

j

i

w

training to minimize error

slide-14
SLIDE 14

Topology of Neural Classifier

1

1

x

d

x

1

y =

  • x

i

w

1

i

w

d

i

w

  • Every newly added

layer assumes the role

  • f the network output

i

y

, y i − i

y

2 − i − i

y

1 2 −

w

,y i

w w

1 =

  • x

1

y

1

y

1

y

3 − i

y

2

w

,

i

y i

1 − i

i

1

i

w

d

i

w

1 + d

i

w

1 =

  • x

y

= + = d i i d

x x

1 2 1 2

1 =

  • x
  • Pyramid structure

– First perceptron receives the pattern x ∈ Rd – Successive perceptrons also receive inputs from preceding perceptrons and a parabolic pattern xd+1=∑xi

2, i=1…d

  • Non-linear boundary by

training a sequence of linear perceptrons

slide-15
SLIDE 15

Training and Outcome

  • Weights of added layer are adjusted through the

thermal perceptron training algorithm

  • Allocated boundary is non-linear in the original space x ∈ Rd
  • Each perceptron separates its input space linearly
  • Weights of preceding layers do not change

Theorem:

  • The sequence of boundaries allocated by the neurons

converges to a boundary that perfectly separates the two populations in the training set

slide-16
SLIDE 16

Boundary Evolution Example (layer 0)

Faulty patterns erroneously classified Faulty patterns correctly classified Nominal patterns erroneously classified Nominal patterns correctly classified

x1 x2

  • 2
  • 1.5
  • 1
  • 0.5

0.5 1 1.5

  • 2
  • 1.5
  • 1
  • 0.5

0.5 1 1.5 2 2.5

slide-17
SLIDE 17

Boundary Evolution Example (layer 1)

Faulty patterns erroneously classified Faulty patterns correctly classified Nominal patterns erroneously classified Nominal patterns correctly classified

  • 2
  • 1.5
  • 1
  • 0.5

0.5 1 1.5

  • 2
  • 1.5
  • 1
  • 0.5

0.5 1 1.5 2 2.5

x1 x2

slide-18
SLIDE 18

Boundary Evolution Example (layer 2)

Faulty patterns erroneously classified Faulty patterns correctly classified Nominal patterns erroneously classified Nominal patterns correctly classified

  • 2
  • 1.5
  • 1
  • 0.5

0.5 1 1.5

  • 2
  • 1.5
  • 1
  • 0.5

0.5 1 1.5 2 2.5

x1 x2

slide-19
SLIDE 19

Boundary Evolution Example (output layer)

Faulty patterns erroneously classified Faulty patterns correctly classified Nominal patterns erroneously classified Nominal patterns correctly classified

  • 2
  • 1.5
  • 1
  • 0.5

0.5 1 1.5

  • 2
  • 1.5
  • 1
  • 0.5

0.5 1 1.5 2 2.5

x1 x2

slide-20
SLIDE 20

Matching the Inherent Boundary Order

  • Monitor classification on validation set
  • Prune down network to layer that achieves

the best generalization on validation set

  • Evaluate generalization on test set

Finding The Trade-Off Point (Early Stopping)

  • No! The goal is to generalize
  • Inflexible and over-fitting

boundaries

Is Higher Order Always Better?

2 4 6 8 10 12 14 70 75 80 85 90 95 100 number of layers rate (%) training set validation set

slide-21
SLIDE 21

Are All Measurements Useful?

  • 3
  • 2.5
  • 2
  • 1.5
  • 1
  • 0.5

0.5 1 1.5 2

  • 3
  • 2
  • 1

1 2 3 4 Nominal patterns Faulty patterns

Non-Discriminatory

x1 x2

  • 4
  • 3
  • 2
  • 1

1 2 3 4

  • 4
  • 3
  • 2
  • 1

1 2 3 4 Nominal patterns Faulty patterns

Linearly Dependent

x1 x2

slide-22
SLIDE 22

Curse of Dimensionality

  • Several possible boundaries exist – choice is random
  • By increasing the dimensionality we may reach

a point where the distributions are very sparse

New patterns

  • Random label assignment to new patterns

Nominal training patterns Faulty training patterns Nominal training patterns Faulty training patterns Nominal training patterns Faulty training patterns

slide-23
SLIDE 23

Genetic Measurement Selection

  • Encode measurements in a bit string, with the k-th bit

denoting the inclusion (1) or exclusion (0) of the k-th measurement

0 1 1 0 1 1 1 0 0 1 0 0 0 0 1 1 1 0 0 1 1 1 1 0 1 1 0 0 0 0 0 1 1 1 1 0 1 0 0 0 1 0 1 0 1 1 1 0 0 0 0 1 0 0 1 0 1 0 1 1 0 0 1 0 1 1 1 0 1 1 0 0 1 0 1 1 1 0 0 1 0 1 1 0 1 1 0 0 0 0 0 1 1 1 1 1 0 0 0 0 1 0 1 0 1 1 1 0 0 0 0 1 0 0 1 0 1 0 0 0 Reproduction Crossover & Mutation GENERATION

t

GENERATION

t+1

NSGA II: Genetic algorithm with multi-objective fitness function reporting pareto-front for error rate (gr) and number of re-tested circuits (nr) Fitness function:

slide-24
SLIDE 24

Two-Tier Test Strategy

Low-Cost Tester

All Chips Simple Alternative Measurements

Neural Classifier

Most Chips

Highly Accurate Pass/Fail Decision Measurement Pattern in Guard-band

Few Chips

High-Cost Tester

Specification Test Measurements

Compare

Highly Accurate Pass/Fail Decision

Design Specs

Inexpensive Machine Learning Test Expensive Specification Test

slide-25
SLIDE 25

Guard-bands

2.25 2.3 2.35 2.4 2.45 2.5 2.55 1.65 1.7 1.75 1.8 1.85 1.9

nominal guard-band faulty guard-band

x1 x2

Nominal patterns Faulty patterns test hypersurface Guard-banded region

  • Introduce a trichotomy in the measurement space

in order to assess the confidence of the decision

slide-26
SLIDE 26

Testing Using Guard-bands

  • Examine measurement pattern with respect to the guard-bands

2.25 2.3 2.35 2.4 2.45 2.5 2.55 1.65 1.7 1.75 1.8 1.85 1.9

nominal guard-band faulty guard-band

x1 x2

Nominal patterns Faulty patterns test hypersurface Guard-banded region

Classify chip to dominant class Re-test via specification test

  • Identify circuits that need to be re-tested via specification tests
slide-27
SLIDE 27

Guard-band Allocation

D

  • Identify the overlapping regions
  • Guard-bands are allocated separately
  • The ontogenic classifier allocates the guard-band

D

  • Clear the overlapping regions

D

Nominal patterns Faulty patterns

slide-28
SLIDE 28

Guard-band Allocation

D D D D

  • The guard-banded region can been varied by controlling D

Nominal patterns Faulty patterns

slide-29
SLIDE 29

Experiment 1: Switch-Capacitor Filter

C1A C2A C2B C4A C4B C5A C1 C2 C3 C4 C5 C1C C3A C3B C5B C5C C3D C1D C3C C1B

Vin Vout

1

φ

2

φ

2

φ

2

φ

1

φ

1

φ

1

φ

1

φ

2

φ

2

φ

2

φ

2

φ

1

φ

1

φ

1

φ

1

φ

1

φ

1

φ

2

φ

2

φ

2

φ

2

φ

Specifications considered

  • Ripples in stop- and

pass-bands

  • Gain errors
  • Group delay
  • Phase response
  • THD

Experiment Setup

  • N=2000 instances
  • N/2 assigned to the

training set, N/4 to the validation set and N/4 to the test set

slide-30
SLIDE 30

White-Noise Test Stimulus

CLK Din Qn Qm LP filter Type-1 LFSR

DUT signature

DUT digitizer

pseudo-random bit sequence white noise

Time (ms)

  • 2.0

2.0 4.0

Voltage (V)

2 4 6 8 10 12 14 16 18

DUT signature pseudo-random bit sequence white noise

t1 t2 t30 tr ts

slide-31
SLIDE 31

Test Time vs. Test Accuracy Trade-Off

(0%,4%) (8.1%,2%) (15.6%,0.6%) (20.7%,0%)

5 10 15 20 25 30 35 40 45 1 2 3 4 5 6 7 8 retested circuits (%) test error (%) f2 f1 f3

guard-band test time=15ms specification test time >0.5s

slide-32
SLIDE 32

Experiment 2: RFMD2411

Gain 3rd Order Intercept Noise Figure Other S Parameters

LNA

x x x x x x

Mixer

x x x x

  • LNA + Mixer

(Cascade)

x x x

  • 541 devices (training set 341, validation set 100, test set 100)
  • Considered 13 specs @850MHz (7 different test configurations)
slide-33
SLIDE 33

Test Configuration

Stimulus DUT Response

1 2 3 4 5 6 7 8 9 10 x 10

7

50 100 150 200 250 300 350 400 450 frequency FFT of DUT response

noise floor 28 tones

FFT DUT Test Signal xt(t) f1 f2 Mixer DUT signature xs(t) LPF

slide-34
SLIDE 34

5 10 15 20 25 30 35 40 45 50 1 2 3 4 5 6 7 8 retested circuits (%) test error (%) f2 (gr<2) f2 (gr<4) f3 f1

(0%,3.92%) (11.67%,1.96%) (27.62%,0%)

Test Time vs. Test Accuracy Trade-Off

guard-band test time=5ms specification test time ~1s

slide-35
SLIDE 35

Analog/RF Specification Test Compaction

  • Instead of “other, alternate measurements” use

existing inexpensive specification tests to predict pass/fail for the chip and eliminate expensive ones

Non-disruptive application:

  • RFCMOS zero-IF down-converter for cell-phones
  • Fabricated at IBM, in production until late 2008
  • Data from 944 devices (870 pass – 74 fail)
  • 136 performances (65 non-RF – 71 RF)
  • Non-RF measurements assumed much cheaper than

RF, aim at minimizing number of measurements

Case-Study:

slide-36
SLIDE 36

Setup and Questions Asked

  • How accurately can we predict pass/fail of a device

using only a subset of non-RF measurements?

Question #1:

  • How does this accuracy improve by selectively adding

a few RF measurements to this subset?

Question #2:

  • Split 944 devices into training and test set (472 devices

each, chosen uniformly at random.

  • Repeat 200 times and average to obtain statistically

significant results

Setup:

slide-37
SLIDE 37

Results

Only non-RF performances:

  • 16 measurement suffice to

predict correctly over 99%

  • f the devices (in our case

4 out of 472 are on average mispredicted Adding RF performances:

  • By adding to these 16 non-

RF performances 12 RF performances, error drops to 0.38% (i.e. 1 out of 472 devices mispredicted)

slide-38
SLIDE 38

Continuation of Case-Study

  • Given appropriate test cost information (test time per

performance, cost per test configuration, groups of performances sharing test configuration, per second cost of non-RF vs RF ATE, etc.) select subsets of performances that minimize cost instead of cardinality

Cost-driven compaction (TVLSI’09 – in press):

  • Assess the effectiveness of our guard-banding method in

enabling exploration of test-cost / test quality trade-off

Guard-banding:

  • Experiment repeated for data from 4450 devices with

very similar results

Larger Data Set:

slide-39
SLIDE 39

Current Research Activities (TI/IBM)

  • Investigate whether process variations over the lifetime
  • f production affect the accuracy of the models
  • Devise a method for adapting to process variations

(periodic, event-driven, or monitoring-based retraining)

Dealing with process variations, shifts, and drifts:

  • Investigate whether variations across sites of an ATE or

across ATE affect the accuracy of the models

  • Devise a method for adapting to ATE/site variations

(training per site or ATE, using calibration filters, etc.)

Dealing with variations across ATE and across sites:

  • Use correlations between PCM, wafer sort, and final test

measurements to support adaptive test

Measurement correlations:

slide-40
SLIDE 40

What’s Next? Stand-alone Built-in Self-Test

  • On-chip generation of simple test stimulus
  • On-chip acquisition of simple measurements
  • On-chip implementation of trainable neural classifier (w. floating gates)

A stand-alone BIST method for mixed-signal/RF circuits:

On-Chip Stimulus Generator

Analog Mux Mixed- Signal / RF Circuit

Programmable Response Acquisition Trainable On-Chip Neural Classifier

Input Output BIST Input BIST Output BIST Control

slide-41
SLIDE 41

Summary

  • A non-linear neural classifier supporting a

novel paradigm for testing analog/RF circuits

– Achieves significant reduction in test cost without sacrificing accuracy of specification test – Enables exploration of trade-off between test cost and test accuracy via guard-bands Contribution:

  • Disruptive: Machine learning-based test
  • Non-disruptive: Specification test compaction
  • Futuristic: Stand-alone Built-in Self-Test

Applications:

slide-42
SLIDE 42

Knob-Tuning for Performance Calibration

  • Low-cost machine-learning based testing
  • Iterative, correlation-based knob tuning

“Healing” of failing chips:

slide-43
SLIDE 43

Correlation Mining for Yield Improvement

  • Guiding modifications in potential design re-spin
  • Assisting in fabrication process quality control

Improve yield by:

slide-44
SLIDE 44

More Information

  • E-mail: yiorgos.makris@yale.edu

Contact:

  • H-G. D. Stratigopoulos, Y. Makris, “Constructive Derivation of Analog

Specification Test Criteria,” Proceedings of the IEEE VLSI Test Symposium (VTS), pp. 245-251, 2005

  • H-G. D. Stratigopoulos, Y. Makris, “Non-Linear Decision Boundaries for

Testing Analog Circuits,” IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems (T.CAD), pp. 1360-1373, Nov. 2005

  • H-G. D. Stratigopoulos, Y. Makris, “Bridging the Accuracy of Functional and

Machine-Learning-Based Mixed-Signal Testing,” Proceedings of the IEEE VLSI Test Symposium (VTS), pp. 406-411, 2006

  • H-G. D. Stratigopoulos, P. Drineas, M. Slamani, Y. Makris, “Non-RF to RF Test

Correlation Using Learning Machines: A Case Study,” Proceedings of the IEEE VLSI Test Symposium (VTS), pp. 9-14, 2007

  • H-G. D. Stratigopoulos, Y. Makris, “Error Moderation in Low-Cost Machine

Learning-Based Analog/RF Testing ,” IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems (T.CAD), pp. 339-351, Feb. 2008

Publications:

slide-45
SLIDE 45

Questions?