Bayesian Inference Technique for Data mining for Yield Enhancement in Semiconductor Manufacturing Data Presenter: M. Khakifirooz Co-authors: C-F Chien, Y-J Chen National Tsing Hua University
ISMI 2015, 16th -18th Oct. KAIST, Daejeon, Korea
1
Co-authors: C-F Chien, Y-J Chen National Tsing Hua University ISMI - - PowerPoint PPT Presentation
Bayesian Inference Technique for Data mining for Yield Enhancement in Semiconductor Manufacturing Data Presenter: M. Khakifirooz Co-authors: C-F Chien, Y-J Chen National Tsing Hua University ISMI 2015, 16 th -18 th Oct. KAIST, Daejeon, Korea 1
1
The Purpose
Inference Data Structure provided by Data Model Data Analysis Approach
Selection (BVS)
Conclusive Research Framework Final Decision Table
Conclusion & Path Forward
2
Bayesian Inference Naรฏve Bayesian Classifier Gaussian Bayesian Classifier โฆ Bayesian Networks
3
Learning Curve
4
Human Experience Human Experience + System Analysis
Yield Learning Curve of Semiconductor Manufacturing
Yield Learning Curve of Semiconductor Manufacturing: In addition to data analytics, Cumulative
significantly enhanced yield improvement
Effron(1996), Tobin et al. (1999)
5
๐๐๐ โค ๐๐๐
๐๐๐
๐=1 ๐๐ ๐=1 ๐๐๐๐
๐=1 ๐ ๐=1 ๐๐ ๐=1 ๐๐๐๐
Obs. ๐ฐ๐๐ฌ๐ ๐ฐ๐๐ฌ๐ ๐1 ๐1 ๐2 ๐2 ๐1 ๐2 ๐3 ๐1 Na Obs. ๐ฐ๐๐ฌ๐-๐๐ ๐ฐ๐๐ฌ๐โ๐1 ๐ฐ๐๐ฌ๐-๐๐ ๐ฐ๐๐ฌ๐-๐2 ๐1 1 1 ๐2 1 1 ๐3 1
Nominal Variables Dummy Variables
6
Yield ๐๐๐๐๐ ๐ ๐๐๐๐๐ ๐
๐๐๐๐ 1 ๐๐๐๐ 2
๐๐๐๐ 1 ๐๐๐๐ 1
๐๐๐๐ 2 Tool 2 Yield ๐๐๐๐๐ ๐ ๐๐๐๐๐ ๐
Chamber 1 Chamber 2
Chamber 2 Chamber 1
Chamber 1 Chamber 2 Yield ๐๐๐๐๐ ๐ ๐๐๐๐๐ ๐
๐๐๐๐ 1. Chamber 1 ๐๐๐๐ 2. Chamber 2
๐๐๐๐ 1. Chamber 2 ๐๐๐๐ 1. Chamber 1
๐๐๐๐ 2. Chamber 1 ๐๐๐๐ 2. Chamber 2 Yield ๐๐๐๐๐ ๐ ๐๐๐๐๐ ๐
๐ธ๐๐ข๐ 1.1 ๐ธ๐๐ข๐ 1.2
๐ธ๐๐ข๐ 2.1 ๐ธ๐๐ข๐ 2.2
๐ธ๐๐ข๐ 3.1 Date 3.2
Yield ๐ ๐. ๐ผ ๐. ๐ซ๐ ๐ ๐ ๐. ๐ผ ๐. ๐ซ๐ ๐ ๐ ๐. ๐ผ ๐. ๐ซ๐ ๐ ๐ ๐. ๐ผ ๐. ๐ซ๐ ๐ ๐ ๐. ๐ผ ๐. ๐ซ๐ ๐
1 1
1 1
1 1
Yield ๐ ๐. ๐ผ ๐. ๐ซ๐ ๐ ๐ ๐. ๐ผ ๐. ๐ซ๐ ๐ ๐ ๐. ๐ผ ๐. ๐ซ๐ ๐ ๐ ๐. ๐ผ ๐. ๐ซ๐ ๐ ๐ ๐. ๐ผ ๐. ๐ซ๐ ๐
๐ธ๐๐ข๐ 1.1 ๐ธ๐๐ข๐ 1.2
๐ธ๐๐ข๐ 2.1 ๐ธ๐๐ข๐ 2.2
๐ธ๐๐ข๐ 3.1 ๐ธ๐๐ข๐ 2.3
7
Obs. ๐ฐ๐๐ฌ๐-๐๐ ๐ฐ๐๐ฌ๐โ๐1 ๐ฐ๐๐ฌ๐-๐ ๐ ๐1 1 ๐2 1 ๐3 1 Pr(ith variable sellected) 1 3 1 3 1 3
๐ Multinomial 1
3 , 1 3 , 1 3 1,0,0 0,0,1 0,1,0 ๐ฐ๐๐ฌ๐-๐๐ ๐ฐ๐๐ฌ๐-๐ ๐ ๐ฐ๐๐ฌ๐โ๐1 To randomly pick a point in this space, we need a continues distribution Distribution over Multinomial (posterior distribution): Dirichlet Distribution
selection probability based on engineer experience
๏ฎ Critical Phenomena: i.
ii.
iii.
Use random sampling (MCMC techniques: Gibbs sampler, Metropolis-Hastings,โฆ) to approximate the distribution and selecting significant explanatories
8
9
Suppose ๐๐, ๐๐~๐๐ฌ ๐ฆ, ๐ฆ2 Beginning with initial value ๐๐
๐ , ๐๐ ๐
Sampling at iteration t as follow:
Iteration Sample ๐ฒ๐ Sample ๐ฒ๐ k x๐
๐ข ~๐๐ฌ x๐|x๐ tโ1
x๐
๐ข ~๐๐ฌ x๐|x๐ ๐ข
Iterating the above step until the sample values have the same distribution as if they where sampled from the true posterior joint distribution
10
Yield value Representative var. Bad Yield 53.12 < 1 Middle Yield 53.12 โค and โค 57.51 ignore Good Yield >57.51
11
Level a Level b Level c fc๐ fc๐ Level d fd๐ fd๐
Variable I Variable II
๐ง < 0, "No agreement" 0 โค ๐ง < 0.2, โSlight agreementโ 0.2 โค ๐ง < 0.4, "Fair agreement" 0.4 โค ๐ง < 0.6, "Moderate agreement" 0.6 โค ๐ง < 0.8, "Substantial agreement" 0.8 โค ๐ง โค 1, "Almost perfect agreement"
12
Data Preparation Data Mining & Key Factor Screening Problem Definition Data Integration Dummy Variable Construction for Integrated Variables (1460 var.) Wrap the associate variables Cohenโs Kappa Statistics for each pairs of input variables
Agreement
Assign Cutting Point & Bad/Middle/Good Wafers
No Agreement
A Bayesian Framework for Semiconductor Manufacturing Data
Almost perfect agreement Substantial agreement Moderate agreement 3 109 1,764 Fair agreement Slight agreement No agreement 24,539 280,081 758,574 THE CLASS DISTRIBUTION FOR THE KAPPA TEST FOR EACH PAIR OF INPUT VARIABLES
13
BVS via Gibbs Sampler Data Clearance ๐ง โค 0.2
No Agreement Agreement
GLM Construction with Gaussian distribution & Repeated Random Sub-sampling Validation A Comparison to the Wrapped Variables Define Abnormal Devices & Time Model Construction, Evaluation & Interpretation Cohenโs Kappa Statistics for each pairs of X & Y Data Mining & Key Factor Screening
Model RMSE Adjusted R-squared Min Median Max Min Median Max Gibbs + GLM 1.842 2.653 2.841 0.046 0.371 0.711 GBM + GLM 2.534 3.051 3.332 0.000 0.053 0.337 RF + GLM 2.268 2.838 3.660 0.016 0.293 0.507 GLM 7.951 34.60 139.8 0.000 0.029 0.214
Number of resamples 20, Number of iterations 2
14
15
๏ฎ
๏ฎ
๏ฎ
๏ฎ
16
17