SHAHZAD 2012 Archivage
SHAHZAD 2012 Archivage
SHAHZAD 2012 Archivage
Présentée par
Permission to make digital or hard copies of all or part of this thesis for the academic use is granted for free;
however all the copies must bear this notice and the full citation on the first page. To copy otherwise or to
republish requires prior specific permission.
iv
iv
The Dissertation Committee for Muhammad Kashif Shahzad Certifies that this is the
approved version of the following dissertation:
<< Exploitation dynamique des données de production pour améliorer les méthodes DFM
dans l’industrie Microélectronique >>
Thesis Committee:
xi
where I was able to industrialize the research carried out through the course of this PhD experienced an
exceptionally advanced R&D environment.
Finally I would like to express my words of gratitude for my academic supervisor Dr. Michel
Tollenaere, co-supervisor Dr. Ali Siadat and my industrial research director Mr. Stéphane Hubac for their
guidance in developing the present work and also for being so supportive, encouraging, and understanding in
good and bad times. It is because of their support and guidance that I was able to publish 7 international
conferences along with 4 journal submissions currently under review. My special thanks for my colleagues at
STMicroelectronics and G-SCOP with home I participated in different case studies and important discussions
that helped me in understanding the complex semiconductor manufacturing processes.
This is not the end but the beginning and there is a long way to go...
xii
xii
Acknowledgement
I feel necessary to personally thank all those who have helped and guided me throughout my PhD.
This research and thesis could not have been completed without the help and guidance from many
people. First and foremost I have to thank my academic supervisors Dr. Michel Tollenaere and Dr. Ali
Siadat. They were always there to guide me in my research as well as in solving problems in general. Under
their guidance I was able to grow from an under graduate student to a Doctor of Philosophy. Although I
spent 30% of my time at the research lab, but still my supervisors were always there for me when needed. It
is really an honor for me to work at the advanced R&D center Crolles, of the 7th largest semiconductor
manufacturer (2012) STMicroelectronics. I owe my special gratitude for Mr. Stéphane Hubac, my industrial
director, who helped me in integration and getting familiar with the complex semiconductor manufacturing
and R&D processes in the most simple and beautiful way. I found him as an intelligent and highly competent
person with exceptional managerial skills and his training and guidance shall help all the way throughout my
career.
I worked with different R&D teams and I would like to acknowledge their support, guidance and
participation. First of all I am grateful to Renaud Nicolas, Lidwine Chaize and Jean-Claude Marin from T2D
(technology to design) Aples team for their participation and we together proposed MAM (mapping and
alignment) model for die/site level correlations to capture spatial correlations. It is a significant contribution
towards more effective DFM methods. I cannot forget my experience with the yield management group
particularly Celine Julien and Roberto Gonella where we together proposed SPM (spatial positioning) model
to accurately capture the spatial variations. This proposed solution is a significant contribution towards
accurately capturing the spatial variations. My special thanks go to IT teams at STMicroelectronics because
without their support it would have not been possible to analyze the production database issues and propose
the ROM2I2 (referential ontology Meta model for information integration). I am grateful to Hugues
Duverneuil, Guillaume Chezaud, Brigitte Boulloud, David Rozier, Stephane Coquio and Bruno Sauvage for
giving me access to the manufacturing database and data storage architecture.
I appreciate the contribution and support of Jerome Altieri, Thomas Chaillou and Hugues Kejikian
from the equipment engineering team at STMicroelectronics. We together developed a 3-step yield aware
sampling strategy (YASS) to enhance the inspection capacities for more R&D measurements. I nust
appreciate the team lead by Mr. Herve Jaouen in the development of SMA (spice model alignment) tool for
the process integration team. I personally thank all the team members in the process integration team Raul-
andres Bianchi, Gregory Bidal, Claire Gallon, Dominique Golanski, Guillaume Ribes, Jerome Bonnouvrier
and Nicolas Planes for their support and help.
Finally I want to thank my colleagues at STMicroelectronics: Francois Pasqualini, Philippe
Vialletelle, Alain Delporte and Aymen Mili with whom I did not work in a team but their support has
helped me a lot. I also thank all of my colleagues at G-SCOP (Mhamed Sahnoun, Mohammed Farouk
Bouaziz) and LCFC labs for their encouragement and support.
xiii
xiv
xiv
Exploitation dynamique des données de production pour améliore les méthodes
DFM dans l’industrie Microélectronique
Abstract: The DFM (design for manufacturing) methods are used during technology alignment and
adoption processes in the semiconductor industry (SI) for manufacturability and yield assessments. These
methods have worked well till 250nm technology for the transformation of systematic variations into rules
and/or models based on the single-source data analyses, but beyond this technology they have turned into
ineffective R&D efforts. The reason for this is our inability to capture newly emerging spatial variations. It
has led an exponential increase in technology lead times and costs that must be addressed; hence, objectively
in this thesis we are focused on identifying and removing causes associated with the DFM ineffectiveness.
The fabless, foundry and traditional integrated device manufacturer (IDM) business models are first analyzed
to see coherence against a recent shift in business objectives from time-to-market (T2M) and time-to-volume
towards (T2V) towards ramp-up rate. The increasing technology lead times and costs are identified as a big
challenge in achieving quick ramp-up rates; hence, an extended IDM (e-IDM) business model is proposed to
support quick ramp-up rates which is based on improving the DFM ineffectiveness followed by its smooth
integration. We have found (i) single-source analyses and (ii) inability to exploit huge manufacturing data
volumes as core limiting factors (failure modes) towards DFM ineffectiveness during technology alignment
and adoption efforts within an IDM. The causes for single-source root cause analysis are identified as the (i)
varying metrology reference frames and (ii) test structures orientations that require wafer rotation prior to the
measurements, resulting in varying metrology coordinates (die/site level mismatches). A generic coordinates
mapping and alignment model (MAM) is proposed to remove these die/site level mismatches, however to
accurately capture the emerging spatial variations, we have proposed a spatial positioning model (SPM) to
perform multi-source parametric correlation based on the shortest distance between respective test structures
used to measure the parameters. The (i) unstructured model evolution, (ii) ontology issues and (iii) missing
links among production databases are found as causes towards our inability to exploit huge manufacturing
data volumes. The ROMMII (referential ontology Meta model for information integration) framework is then
proposed to remove these issues and enable the dynamic and efficient multi-source root cause analyses. An
interdisciplinary failure mode effect analysis (i-FMEA) methodology is also proposed to find cyclic failure
modes and causes across the business functions which require generic solutions rather than operational fixes
for improvement. The proposed e-IDM, MAM, SPM, and ROMMII framework results in accurate analysis
and modeling of emerging spatial variations based on dynamic exploitation of the huge manufacturing data
volumes.
Keywords: design for manufacturing, effective root cause analysis, information integration, metrology
coordinates mapping and alignment
15
16
16
Table of Contents
List of Tables ......................................................................................................................................21
List of Figures ....................................................................................................................................23
List of Acronyms ................................................................................................................................27
Chapter 1: Introduction ....................................................................................................................31
1.1 Introduction to Semiconductor Industry (SI) ...................................................................33
1.2 Role of DFM and Economic Benefits ..............................................................................35
1.3 DFM Challenges and Limitations ....................................................................................36
1.4 Research Questions ..........................................................................................................40
1.5 Research Methodology and Schematic ............................................................................41
1.6 Major Contributions .........................................................................................................44
1.6.1 Analysis of Overall System and Industrial Contributions .......................................44
1.6.2 Scientific Contributions ...........................................................................................45
1.7 Thesis Organization ..........................................................................................................46
1.8 Typographic Conventions ................................................................................................47
Chapter 2: Literature Review ...........................................................................................................49
2.1 Semiconductor Industry (SI): Background and Challenges .............................................51
2.1.1 Historical Background .............................................................................................51
2.1.2 Evolution of Semiconductor Industry .....................................................................52
2.1.3 Role of Moore’s Law in Semiconductor Industry ...................................................53
2.1.4 Nanometer vs. Micrometer Semiconductor Technologies ......................................55
2.1.5 Semiconductor Business Model and its Evolution ..................................................56
2.1.6 Trends in Semiconductor Industry ..........................................................................57
2.1.7 Challenges faced by Semiconductor Industry .........................................................58
2.2 Role of DFM Methods in Semiconductor Industry and Evolution ..................................61
2.2.1 SI Challenges and Rise of Interest in DFM .............................................................62
2.2.2 A Comparison of DFM Efforts in SI and Manufacturing Industries ......................62
2.2.3 DFM Techniques (pre-1980 era) .............................................................................64
2.2.4 Adaption and Diversification of DFM to SI (post-1980 era) ..................................66
2.2.5 DFM Challenges and ECAD/TCAD Tools .............................................................69
2.2.6 Increasing Design Size and DFM Realization Challenges ......................................70
2.2.7 Role of SI Business Models in DFM Evolution and Adaption ...............................71
2.2.8 Industry wide Understanding of the DFM Concept ................................................72
2.3 Information Integration Challenges towards more Effective DFM methods ...................74
2.3.1 Data/Information Integration Issues ........................................................................75
2.3.2 Ontology from Philosophy to Computer Science....................................................75
2.3.3 Data/Information Integration ...................................................................................75
2.3.3.1 Metadata based data/information Integration approaches ......................76
2.3.3.2 Ontology based data/information Integration approaches ......................76
2.3.4 Ontology based Database-Integration Approaches .................................................77
2.3.5 RDB Integration based on Schema Matching .........................................................78
17
2.3.6 RDB Schema to Ontology Mapping Approaches ................................................... 79
2.3.7 Ontology Driven Data Extraction Tools ................................................................. 79
2.4 Summary and Conclusions .............................................................................................. 80
Chapter 3: An Extended IDM (e-IDM) Business Model .................................................................. 83
3.1 Introduction...................................................................................................................... 85
3.1.1 Strategic Planning and Analysis ............................................................................. 85
3.1.2 SCAN Analysis ....................................................................................................... 86
3.1.2.1 Ranking Business Objectives (Step-1)..................................................... 86
3.1.2.2 SWOT Analysis (Step-2) .......................................................................... 87
3.2 SCAN Analysis: Part-1 (Top Ranked Business Objectives in SI)................................... 87
3.3 Key Improvement Areas in IDM-fablite Business Model ............................................... 88
3.4 Technology Derivative/Improvement Process Analysis ................................................. 90
3.5 Key Challenges in Technology Derivative/Improvement Process .................................. 91
3.5.1 Data Extraction Issues ............................................................................................ 91
3.5.2 Variance Analysis Challenges ................................................................................ 92
3.5.3 Silicon Based Correlation Limitations .................................................................... 92
3.6 SWOT Analysis on IDM-fablite Business Model ........................................................... 92
3.7 Proposed Extended IDM (e-IDM) Business Model ........................................................ 94
3.8 Research Schematic and Advancements (e-IDM Model)............................................ 95
3.9 Summary and Conclusions .............................................................................................. 97
Chapter 4: I-FMEA Methodology for True DFM Challenges ........................................................... 99
4.1 Introduction.................................................................................................................... 101
4.2 Historical Evolution of FMEA Methodology ................................................................ 101
4.2.1 FMEA Process and Evolution .............................................................................. 101
4.2.2 Basic FMEA Vocabulary ...................................................................................... 102
4.2.3 Benefits and Limitation of Traditional FMEA Approach .................................... 102
4.3 Proposed Interdisciplinary FMEA (i-FMEA) Methodology ......................................... 103
4.3.1 Comparison of i-FMEA with Traditional FMEA Approach ................................ 103
4.3.2 i-FMEA Methodology and Thesis Schematic ...................................................... 104
4.4 i-FMEA Methodology Results ...................................................................................... 104
4.4.1 Step-2: Initial Failure Modes and Root Causes .................................................... 104
4.4.1.1 Technology Derivative/Improvement Initiative ..................................... 106
4.4.1.2 Fast Technology Transfer ...................................................................... 108
4.4.2 Operational Fixes through Joint Projects .............................................................. 111
4.4.3 Step-3: Cyclic Failure Modes and Root Causes ................................................... 113
4.4.4 Generic R&D Solutions ........................................................................................ 115
4.5 Research Schematic and Advancements (i-FMEA Methodology) ................................ 116
4.6 Summary and Conclusions ............................................................................................ 118
Chapter 5: Measurement Coordinates Mapping, Alignment and Positioning ................................. 119
5.1 Introduction to Device/Interconnect Modeling.............................................................. 121
5.2 Challenges in Multi-Source Data Analysis.................................................................... 124
5.3 Site/Die Level Mismatch Problem................................................................................. 126
18
5.4 Proposed Die/Site Level Mapping, Alignment and Qualification (MAM) Model ........128
5.4.1 Site/Site or Die/Die Level Mapping and Alignment .............................................128
5.4.2 Die/Site level Qualification ...................................................................................131
5.5 Test Structure Position Based Mapping and Alignment (SPM) Model .........................134
5.5.1 SPM (spatial positioning) Problem (Source/Target 1*1) Context ...................135
5.5.2 Step-Circle based Basic Algorithm [Source/Target (1*1)] for Mapping ..............135
5.5.3 Example for Basic Step Circle Algorithm [Source/Target (1*1)] .........................136
5.5.4 SPM (spatial positioning) Problem (Source/Target 1*n) Context ...................138
5.5.5 Optimized Step-Circle Based Algorithms (i*n) Problem......................................139
5.5.6 Example for Optimized Step Circle Algorithm [Source/Target (1*n)] .................139
5.6 Data Model for Position Based Site/Site Mapping ........................................................140
5.7 Research Schematic and Advancements (MAM and SPM Models) ...............................142
5.8 Summary and Conclusions .............................................................................................144
Chapter 6: ROMMII and R&D Data Model for Information Integration ........................................145
6.1 Introduction ....................................................................................................................147
6.2 Historical Evolution from Unstructured towards Structured Data Storage ....................147
6.2.1 Flat Files Database Era (1890 till 1968) ................................................................147
6.2.2 Non-Relational Database Era (1968-1980) ...........................................................148
6.2.3 Relational Database Era (1970 till present) ...........................................................148
6.2.4 Dimensional Database Era (1990 till present) .......................................................148
6.3 Existing Data/Information Integration Systems .............................................................149
6.4 DWH-DM: Information Integration and Business Intelligence Platform ......................150
6.4.1 Basic Definitions and Concepts ............................................................................150
6.4.2 The DWH Architectures, Models and Schemas ....................................................151
6.4.3 Inmon and Kimbell DWH Philosophies ................................................................154
6.4.4 The DWH Challenges ...........................................................................................155
6.5 Proposed R&D DWH Model .........................................................................................156
6.6 Problem Context and Current Challenges ......................................................................162
6.7 Proposed ROMMII Framework .....................................................................................162
6.7.1 Use Case Diagram for ROMMII Platform ............................................................163
6.7.2 Meta Model for ROMMII Platform ......................................................................164
6.7.3 Activity and Sequence Diagrams against Use Cases ............................................165
6.8 The Big Picture of ROMMII Platform ...........................................................................176
6.9 Research Schematic and Advancements (ROMMII and R&D Data Model) .................177
6.10 Summary and Conclusions .............................................................................................179
Chapter 7: Yield Aware Sampling Strategy (YASS) for Tool Capacity Optimization....................181
7.1 Introduction ....................................................................................................................183
7.2 Metrology/Inspection and Production Tools Capacities Issues .....................................183
7.2.1 Why do we need 100% inspection? ......................................................................183
7.2.2 Why additional capacities? ....................................................................................184
7.2.3 What is wrong with the sampling strategies? ........................................................184
7.3 Proposed 3-Step Yield Aware Sampling Strategy (YASS)............................................185
19
7.3.1 Heuristic Algorithm for [PAM, PSM] Models [Step-1] ....................................... 186
7.3.2 Example for [PAM, PSM] predictions ................................................................. 187
7.3.3 Clustering and priority queue allocation [Step-2 and Step-3] .............................. 188
7.4 Data model to Support [PAM, PSM] Models ................................................................ 190
7.5 Research Schematic and Advancements (YASS Strategy) ............................................. 191
7.6 Summary and Conclusions ....................................................................................... 193
Discussions and Conclusions ........................................................................................................... 195
Appendix A: List of Publications .................................................................................................... 199
Appendix B: Semiconductor Design, Mask and Manufacturing Processes ................................... 203
Appendix C: CMOS Inverter Design and Manufacturing (An Example) ...................................... 227
Appendix D: SMA (Spice Model Alignment) Tool ...................................................................... 237
Appendix E: BEOL (back-end-of-line) Variance Analysis Tool .................................................. 247
Appendix F: KLA-Ace Recipe for PT-Inline Correlation ............................................................. 253
Appendix G: EPP (Equipment, Product, Process) Life Cycle Tool .............................................. 255
Appendix H: ACM (Alarm Control Management) Tool ............................................................... 257
References ........................................................................................................................................ 265
20
List of Tables
Table 2.1 – CMOS technology scaling and characteristics ..............................................................................55
Table 2.2 – Semiconductor business models ....................................................................................................56
Table 3.1 – List of SI objectives .......................................................................................................................87
Table 4.1 – Device/Interconnect modeling FMEA result ...............................................................................107
Table 4.2 – Data extraction, alignment and pre-processing FMEA results ....................................................109
Table 4.3 – Fast technology transfer FMEA results .......................................................................................110
Table 4.4 – FMEA results on cyclic failure modes and root causes...............................................................114
Table 5.1 – Description of the MAM model variables ...................................................................................130
Table 5.2 – Description of the Die/Site qualification variables......................................................................131
Table 5.3 – Description of the Step-Circle (B) variables ...............................................................................135
Table 5.4 – Description of Step-Circle (O) algorithms ..................................................................................139
Table 6.1 – OLAP vs. OLTP systems ............................................................................................................149
Table 6.2 – Comparison of DWH data models ..............................................................................................152
Table 7.1 - Description of [PAM, PSM] models variables.............................................................................186
Table 7.2 - Alarm matrix for equipment Ei [Good Yield]..............................................................................187
Table 7.3 - Alarm matrix for equipment Ei [Bad Yield] ................................................................................188
Table 7.4 - Alarm matrix for equipment Ei [Confusion] ................................................................................188
Table 7.5 - Alarm matrix for equipment Ei, Wafer Wj ..................................................................................188
Table 7.6 - Local and Global support for wafer Wj [54%, Good].................................................................188
21
22
22
List of Figures
Figure 1.1 - Global sales revenues of SI...........................................................................................................33
Figure 1.2 - Global product and technology costs for SI ..................................................................................34
Figure 1.3 - Global sales revenues of SI...........................................................................................................34
Figure 1.4 - The role of DFM and economic benefits ......................................................................................35
Figure 1.5 - Cross section of an electronic chip (transistors and interconnects) ..............................................36
Figure 1.6 - Structure of the product wafer ......................................................................................................37
Figure 1.7 - The role of DFM in SI ..................................................................................................................37
Figure 1.8 - Historical evolution of the DFM ...................................................................................................38
Figure 1.9 - Drifts in drawn features and printed images .................................................................................38
Figure 1.10 - Lithography and feature size.......................................................................................................39
Figure 1.11 - The role of DFM in technology alignment and adoption processes ...........................................40
Figure 1.12 - The Research schematic and contributions at a glance ...............................................................42
Figure 2.1 - Benchmarks in semiconductor technology evolution ...................................................................52
Figure 2.2 - Product based market structure .....................................................................................................52
Figure 2.3 - Application segment based market structure ................................................................................53
Figure 2.4 - Geographical position based market structure ..............................................................................53
Figure 2.5 - Circuit integration eras and reducing costs ...................................................................................54
Figure 2.6 - Diversification in Moore’s law [more Moore and more than Moore] ..........................................55
Figure 2.7 - Technology size scale, how small is small? .................................................................................56
Figure 2.8 - Trends in semiconductor industry within last 50 years ................................................................58
Figure 2.9 - Major design challenges faced by semiconductor industry ..........................................................60
Figure 2.10 - Comparison of the design flows in manufacturing industries and SI .........................................63
Figure 2.11 - DFMA (design for manufacturability and assembly) ................................................................65
Figure 2.12 - Early semiconductor design flows with loop back .....................................................................66
Figure 2.13 - Typical design flow within SI .....................................................................................................67
Figure 2.14 - Manufacturability criteria for IC designs ....................................................................................68
Figure 2.15 - Integrated product development framework ...............................................................................68
Figure 2.16 Scope of design rules, DFM rules and DFM models ....................................................................69
Figure 2.17 - 4-dimensional innovation framework for unified agile DFM system.........................................73
Figure 2.18 - Data-method-stat triangule .........................................................................................................74
Figure 3.1 - Ranking of SI Business Objectives ...............................................................................................88
Figure 3.2 - IDM-fablite Business Model Operations ......................................................................................89
Figure 3.3 - Technology derivative/improvement process ...............................................................................90
Figure 3.4 - SWOT Analysis Results ...............................................................................................................93
Figure 3.5 - Proposed extended IDM-fablite business (e-IDM) model ............................................................94
Figure 3.6 - The Research schematic and advancement with e-IDM business model .....................................96
Figure 4.1 - Traditional 5-step FMEA process ...............................................................................................102
23
Figure 4.2 - 4-step i-FMEA methodology ..................................................................................................... 103
Figure 4.3 - 4-Step i-FMEA methodology and research thesis schematic .................................................... 105
Figure 4.4 - The Research schematic and advancement with i-FMEA methodology ................................... 117
Figure 5.1 - BEOL-Interconnect modeling process ....................................................................................... 121
Figure 5.2 - The BEOL-Interconnect modeling process (tentative model) ................................................... 122
Figure 5.3 - The BEOL-Interconnect modeling process (preliminary/pre-production models) .................... 122
Figure 5.4 - The BEOL-Interconnect modeling process (production models) .............................................. 123
Figure 5.5 - Proposed BEOL-Interconnect modeling process ....................................................................... 123
Figure 5.6 - Multi-Source data analysis challenges ....................................................................................... 124
Figure 5.7 - The BEOL – process for interconnections ................................................................................. 125
Figure 5.8 - The BEOL – geometric computations in BEOL process ........................................................... 126
Figure 5.9 - Metrology reference frames ....................................................................................................... 126
Figure 5.10 - Site/Site level mismatches due to notch position ..................................................................... 127
Figure 5.11 - Die/Die levels mismatches due to reference frames ................................................................ 127
Figure 5.12 - Die/Site level qualifications and varying reference frames ..................................................... 128
Figure 5.13 - Polar coordinates formulation and rotation .............................................................................. 128
Figure 5.14 - Reference frame and notch position rotation and translation................................................... 129
Figure 5.15 - Generic formulation for reference frame and notch position rotation ..................................... 129
Figure 5.16 - Reference frame and notch position rotation with translation example ................................... 131
Figure 5.17 - Full map transformations ......................................................................................................... 131
Figure 5.18 - Die/Site qualification, mask and wafer center and site counts................................................. 132
Figure 5.19 - Die/Site qualification example ................................................................................................. 133
Figure 5.20 - Full map Die/Site qualification example ................................................................................. 134
Figure 5.21 - Structure f wafer, sites and dies ............................................................................................... 134
Figure 5.22 - Structure of wafer, sites and dies ............................................................................................. 135
Figure 5.23 - The example of initialization variables (step-1) ...................................................................... 136
Figure 5.24 - The example of Basic Step-Circle algorithm (step-2).............................................................. 136
Figure 5.25 - The example of Basic Step-Circle algorithm (step-3).............................................................. 137
Figure 5.26 - The SPM problem Source/Target (1*n) context ...................................................................... 138
Figure 5.27 - Computational costs with increasing target parameters ........................................................... 138
Figure 5.28 - Computational costs with increasing target parameters ........................................................... 140
Figure 5.29(a) - The data model for SPM model ........................................................................................... 141
Figure 5.29(b) - The data model for SPM model .......................................................................................... 141
Figure 5.30 - The Research schematic and advancement with MAM and SPM models ............................... 143
Figure 6.1 - Existing data extraction and analysis challenges ....................................................................... 147
Figure 6.2 - OLAP cube architecture ............................................................................................................. 151
Figure 6.3 - Slicing and dicing operations on OLAP cube ............................................................................ 151
Figure 6.4 - Principle DWH architectures and frameworks .......................................................................... 152
Figure 6.5 - DWH data models ...................................................................................................................... 152
24
24
Figure 6.6 - DWH schemas ............................................................................................................................153
Figure 6.7 - Data warehouse Meta model for business intelligence ...............................................................154
Figure 6.8(a) - Proposed R&D data warehouse for knowledge capitalization ...............................................158
Figure 6.8(b) - Proposed R&D data warehouse for knowledge capitalization ...............................................159
Figure 6.9 - Meta model for knowledge capitalization...................................................................................160
Figure 6.10 - Flower schema for knowledge capitalization ...........................................................................160
Figure 6.11 - Current data extraction and analysis challenges .......................................................................162
Figure 6.12 - Use Case diagram for ROMMII platform.................................................................................163
Figure 6.13 - Meta model for ROMMII platform ..........................................................................................164
Figure 6.14 - Learning Meta model for new database, table and/or fields .....................................................165
Figure 6.15 - Sequence diagram to learn Meta model use case......................................................................166
Figure 6.16(a) - Modification and synchronization activity diagram .............................................................168
Figure 6.16(b) - Modification and synchronization activity diagram.............................................................169
Figure 6.17 - Sequence diagram to modify and synchronize Meta model .....................................................170
Figure 6.18 - Query validation and optimization Activity diagram ...............................................................172
Figure 6.19 - Query validation and optimization sequence diagram ..............................................................173
Figure 6.20 - Log File parsing and user statistics computation activity diagram ...........................................174
Figure 6.21 - Pre-Failure assessment activity diagram...................................................................................176
Figure 6.22 - The Big picture with ROMMII platform ..................................................................................177
Figure 6.23 - The Research schematic and advancement with ROMMII and R&D model ...........................178
Figure 7.1 - Generic production process with existing sampling strategies ...................................................184
Figure 7.2 - Methodology with predictive state (PSM) and alarm (PAM) models ........................................185
Figure 7.3 - Flow chart for clustering and queue allocations .........................................................................189
Figure 7.4 - Data model to support [PAM, PSM] models ..............................................................................190
Figure 7.5: The Research schematic and advancement with YASS sampling strategy .................................192
25
26
26
List of Acronyms
AAP Average Application Probability
a.k.a. also known as
BEOL Back-End-Of-Line
BPR Business Process Reengineering
CAP Current Application Probability
CMOS Complementary Metal Oxide Semiconductor
C&E Cost and Effect Matrix
CTP Common Technology Platform
DBMS Database Management Systems
DFA Design for Assembly
DFD Design for Design
DFF Design for Fabrication Facilities
DFM Design for Manufacturing
DFMA Design for Manufacturability and Assembly
DFMEA Design Failure Mode Effect Analysis
DFP Design for Product
DFR Design for Reliability
DFSM Design for Semiconductor Manufacturing
DFT Design for Test
DFX Design for All
DFY Design for Yield
DL Description Logic
DM Data Mart
DMS Data-Method-Stat Triangle
DOE Design of Experiment
DRC Design Rule Check
DRM Design Rule Manual
DSP Digital Signal Processing
DTD Document Type Definition
DWH Data Warehouse
EBR Edge Bevel Removal
ECAD Electrical Computer Aided Design
EDA Engineering Data Analysis
EEPROM Electronically Erasable Programmable Read Only Memory
EHF Equipment Health Factor
ER Entity Relationship
EWS Electrical Wafer Sort
27
FDC Fault Detection and Classification
FE Focus Exposure
FEOL Front-End-Of-Line
FIS Federated Information System
FMEA Failure Mode Effect Analysis
FMECA Failure Mode Effect and Criticality Analysis
GDS Graphic Design System
HOLAP Hybrid Online Analytical Processing
IA Impact Analysis
IDM Integrated Device Manufacturer
IO Input Output
IP Intellectual Property
IPD Integrated Product Development
LIFO Last In First Out
LSI Large Scale Integration
LSL Lower Specification Limit
LVS Layout versus Schematic
MCU Micro Controller Unit
MDD Manufacturing Driven Design
MOLAP Multidimensional Online Analytical Processing
MPU Micro Processing Unit
MSI Medium Scale Integration
ODS Operational Data Stores
OEM Original Equipment Manufacturer
OLAP Online Analytical Processing
OLTP Online Transactional Processing
OOC Out of Control
OPC Optical Proximity Correction
OWL Ontology Web Language
PCM Process Control Monitor
PCS Process Control Structures
PDA Personal Digital Assistant
PLC Product Life Cycle
PMB Process Monitoring Box
POP Package on Package
PFMEA Process Failure Model Effect Analysis
PLM Product Life Cycle Management
PSM Phase Shift Masking
QA Quality Assurance
28
QFD Quality Function Deployment
RAM Random Access Memory
RDF Resource Description Framework
ROLAP Relational Online Analytical Processing
ROM Read Only Memory
RPN Risk Priority Number
RTL Register Transfer Level
SAE Society of Automotive Engineers
SCAN Strategic Creative Analysis
SI Semiconductor Industry
SIP System in Package
SIPOC Supplier, Input, Process, Output and Customer
SOC System on Chip
SPICE Simulation Program with Integrated Circuit Emphasis
SSI Small Scale Integration
SWFMEA Software Failure Mode Effect Analysis
SWOT Strengths, Weakness, Opportunity and Threat Analysis
TCAD Technology Computer Aided Design
TRO Top Ranked Objective
TSMC Taiwan Semiconductor Manufacturing Company
T2D Technology to Design
T2M Time to Market
T2V Time to Volume
T2Q Time to Quality
ULSI Ultra Large Scale Integration
UPT Unified Process Technology
USL Upper Specification Limit
VLSI Very Large Scale Integration
WIP Work in Process
29
30
30
Chapter 1: Introduction
The objective of this chapter is to introduce readers with the problem background, research questions and
methodology along with a brief description of the industrial and scientific contributions. We start with an
introduction to the semiconductor industry (SI), current trends and challenges. The biggest challenge faced today is
the increasing technology lead times and costs due to ineffective R&D efforts against newly emerging
manufacturability and yield loss mechanisms. The design for manufacturing (DFM) approach being used against
these issues is presented and discussed for its economic benefits and limitations. The single-source wafer/lot level
analysis is found as the core limiting factor towards more effective DFM methods, resulting in extended lead times
and costs. Based on this discussion, four research questions are formulated followed by a graphical representation of
research methodology and timeline. The industrial and scientific contributions are also briefly presented to complete
the improvement cycle and finally we conclude the chapter with typographic convention used throughout this thesis
and the thesis organization.
Contents
1.1 Introduction to Semiconductor Industry (SI) ..................................................................... 33
1.2 Role of DFM and Economic Benefits ................................................................................ 35
1.3 DFM Challenges and Limitations ...................................................................................... 36
1.4 Research Questions. ........................................................................................................... 40
1.5 Research Methodology and Schematic .............................................................................. 41
1.6 Major Contributions. .......................................................................................................... 44
1.6.1 Analysis of Overall System and Industrial Contributions .................................... 44
1.6.2 Scientific Contributions ....................................................................................... 45
1.7 Thesis Organization ........................................................................................................... 46
1.8 Typographic Conventions .................................................................................................. 47
31
32
32
1.1 INTRODUCTION TO SEMICONDUCTOR INDUSTRY (SI)
The history of the semiconductor can be traced back to 1947 with the invention of first transistor by John Bardeen,
Walter Brattain and William Shockley. The transistor acts like an on/off switch and is the basic building block for an
electronic circuit where thousands of transistors are manufactured and interconnected on the silicon wafer to form an
integrated circuit (IC) chip. The SI is responsible for an efficient and effective manufacturing of the IC chips and was
borne with the invention of first integrated circuit in 1959. The first IC was developed by Jack Kilby at the Texas
Instruments (TI) in 1959; however the Fairchild lab is credited for the invention of first commercial IC logic gate in
1961 by Robert Noyce [Brinkman et al., 1997]. The first highly integrated circuit had 2300 transistors and it was
manufactured by the Intel (4bit microprocessor 4004) in 1971 [Kumar, 2008]. Since then, the transistor count has
exponentially increased till today. This potential growth was first predicted by Gordon E. Moore, cofounder of Intel,
in 1965 who postulated that the transistor count shall double every 18-24 months at reduced cost, area and power
[Moore, 1998]. This prediction along with the industrial slogan “smaller, faster and cheaper” has been accepted by
the SI as a standard. The “smaller” means more transistors in the same area resulting in faster flow of current at
reduced power and “cheaper” refers to continuously decreasing costs due to increasing yield.
The semiconductor industry (SI) has revolutionized our daily lives with IC chips and on the average we are
using more than 250 chips and 1 billion transistors per day per person. These chips are installed in almost all the
equipments around us ranging from dish washers, microwave ovens and flat screens to office equipments. The use of
semiconductors in cars, trains, aircraft and ships is constantly expanding and PCs, servers and pocket calculators also
owe their existence to these chips. The global electronics business is 1.04T$ industry [Dummer, 1997] including
300+ B$ share of the semiconductor manufacturing (2012) during last 50 years. The global semiconductor
manufacturing (1998 till 2011) sales revenues characterize SI with the cyclic demand patterns and a +ve cumulative
annual growth rate (CAGR) of 8.72% (Figure 1.1). This positive CAGR ensures that the demand driven downfalls
follow a cumulative demand growth; hence, it motivates the SI to continuously invest in the R&D for a new
technology to capture the maximum market share from cumulative growths. The R&D investment in SI is expected
to reach 18% of the total revenues by the year 2012. It is evident that the success of SI lies in the effectiveness of
these R&D efforts which has a strong impact on the lead times and costs associated with the new technology
development, its derivative or simply the improvement initiatives.
The SI has kept its pace as per Moore’s law with the continuous introduction of new technologies, every 2
to 3 years. The increase in transistor count for new technologies has led exponential and linear increases in the
technology and product development costs (Figure 1.2) respectively. As a matter of fact, we need new technology
every 2 years to keep up the pace with the Moore’s law but doubling transistors count add manufacturing
complexities that result in new manufacturability and yield loss mechanisms. The existing DFM methods do not
1 The data is collected from the well known technology research centers (i) Gartner {www.gartner.com} and (ii) isuppli {www.isuppli.com}
33
provide solution to these new yield loss mechanisms; hence, extensive R&D efforts along with innovation in the
material, process and equipment are must for new technologies. It often results in increasing (i) technology costs and
(ii) technology lead times.
2 The data is collected from the well known technology research centers (i) Gartner {www.gartner.com}, (ii) isuppli {www.isuppli.com} and (iii)
International Business Strategies, Inc., report: Key trends in technology and supply for advanced features within IC industry (October 1, 2009)
34
34
technology alignment and adoption lead times (e.g. 30nm, Figure 1.3) and costs [Shahzad et al., 2011b]. The
ineffective DFM methods are a big challenge for an IDM-fablite model and in this thesis we are focused on
putting DFM back on track by removing the causes for ineffectiveness.
In this thesis we are focused on (i) fast technology transfer from an alliance to the IDM-fablite business
model and (ii) continuous technology alignment and adoption for its subsequent derivatives and improvements. It is
based on the fact that an IDM-fablite business model provides coherent platform for knowledge capitalization based
on production data exploitation and analysis (effective root cause analysis) for technology alignment and adoption
lead times and costs improvements. The technology is generally defined by the minimum feature size that can be
manufactured with it e.g. 250nm, 180nm, 130nm, 90nm, 65nm, 45nm, 32nm and 22nm. It comprises of the design
flow, Simulation Programs with Integrated Circuit Emphasis (SPICE) models, Design Rules Manual (DRM),
Computer Aided Design (CAD) tools, process flow, equipments, recipes, Statistical Process Control (SPC) and Run
to Run (R2R) feed forward or feed backward strategies. The DRM includes the rules and/or models to address the
manufacturability and yield loss mechanisms for a given technology. These rules and/or models are programmed in
the form of process and DFM kits which are used during the CAD simulations to assess the final printed features and
associated drifts. These drifts in the geometric shapes of the features are then analyzed with Critical Area (CAA) and
hot spot Analyses to find manufacturability and yield limitations. It helps us in solving the yield issues early in design
phase to avoid design respins and waste of resources during prototyping, which result in lead times and costs
improvements.
The technology alignment is a process to validate design, process and DFM kits and is equally applicable
for new technology, its derivative or simple improvement initiatives. It has a strong impact on the reusability concept
as all the design libraries (pre-designed circuits) need to be requalified which is not trivial. The technology adoption
refers to the customization of base technology for each product to achieve target yield levels and it uses Advance
Process (APC) and Equipment (AEC) Control techniques. The DFM methods are objectively focused on the
manufacturability and yield but subjectively they are focused on finding root causes against deviations/drifts
encountered during the technology alignment and adoption efforts. It can be concluded that the success of SI lies in
the effective R&D efforts to identify root causes; hence, effective R&D is a key to improve existing ineffective DFM
methods resulting in improved lead times and costs.
1.2 ROLE OF DFM AND ECONOMIC BENEFITS
The economic benefits associated with the DFM methods are presented in Figure 1.4, which are divided in
technology alignment and/or adoption efforts and economic benefits. The green and red curves represent cumulative
cash flows with and without DFM efforts. The DFM methods help in improving design, development and ramp-up
periods (lead times) by quickly finding the root causes against newly emerging manufacturability and yield loss
mechanisms. It results in cost reduction and provides an early penetration into the market with higher profit margins;
however the biggest gain is the long selling period in the SI against the continuously shortening product life cycles.
35
It is evident that in this competitive environment with continuously shortening product life cycles, we must
focus and improve the technology lead times. This improvement results in the maximum market share in an equal
opportunity 300+ B$ industry along with long selling periods. It is important to note that we are introducing new
technologies at every 2-3 years at an exponentially increasing technology costs (Figure 1.2) and lead times where
DFM ineffectiveness is attributed towards ineffective R&D efforts. The ineffective DFM methods are compensated
with the innovation in design, process, material and equipment which result in extended technology lead times and
costs. These innovations are of course necessary for the changing requirements and to establish leader position,
however local R&D efforts within IDM-fablite models for technology derivative and improvement alignment and
adoption must be improved to achieve true DFM economic benefits.
1.3 DFM CHALLENGES AND LIMITATIONS
It is very important to start with the basic understanding of semiconductor design and manufacturing processes prior
to discuss the DFM role and challenges in SI. Let us start with the design side where an IC is characterized by
electrical parameters (functions) where it undergoes a complex manufacturing process with approximately 200+
operations, 1100+ steps and 8 weeks of processing, however the number of steps and operations vary with the
selected technology. A chip is designed using CAD tools and design libraries (reusable blocks of circuits). It follows
the design simulation steps where design rules are validated and electrical and parasitic (unwanted) parameters are
extracted using SPICE models and technology files. The drifts and variations are adjusted through layout
optimization. The design is further simulated using DFM rules and/or models (CAA and hotspot analyses) to find out
potential drifts in drawn features and printed images that could result in manufacturability and yield losses. These
potential failures are addressed by either changing the design or layout optimization. Upon validation, the design
moves to the mask preparation step. The masks are glass plates with an opaque layer of chrome carrying the target
chip layout. For one product the mask set consists of 15 to 35 individual masks depending on the technology. They
are used to fabricate thousands of transistors and a network of interconnected wires to form an electronic chip
(Figure 1.5) on silicon wafer.
36
36
shapes are drifted based on the DFM rules and/or models for a given technology. The electronic design automation
tools are then used to characterize and extract the electrical and parasitic parameters. The design and wire layout is
optimized until these parameters comply with the technology and product specification. To ensure volume
production, the DFM methods are used to optimize product masks where geometric shapes are compensated against
process drifts to keep printed layout as close as possible to the optimized layout that passed the CAD simulations.
A chip is manufactured on a wafer (Figure 1.6) made from silicon (Si). The wafer is divided in
horizontal/vertical lines crossing each other known as the scribe lines. These scribe lines serve dual purpose: (i) they
contain test structures used for metrology and/or inspection and (ii) they are used to cut the wafer and separate
individual dies (chips). The notch is a cut in the wafer and it is used to describe the crystal orientation and wafer
position during process and metrology operations. A site (field) is composed of individual dies whereas the number
of dies in a site is characterized by the product mask. In order to monitor intra-die variations the test structures could
be placed within the field at different positions. The electronic product undergoes metrology and inspection steps to
ensure product quality where decisions are made based on the parametric yield to either scrap or move the wafer to
next steps. These measurements are done on the test structures in scribe lines or fields; however, the product itself
goes for a functional test at the end of the manufacturing process to sort the bad and good chips.
10
5
9
8
Site Die 7
4
6
3
5
4
A 3
2
2
1
1
0
0
-1
-2
B -3
-1
-4
-2
-5
-6 -3
Notch Position -7
-8 -4
-9
-10 -5
-11
-13-12 -11-10 -9 -8 -7 -6 -5 -4 -3 -2 -1 0 1 2 3 4 5 6 7 8 9 10 11 12
-6 -5 -4 -3 -2 -1 0 1 2 3 4 5 6
c c
Design Manufacturing
Models Rules
Respin Variability
CAD Simulations
37
The concept equivalent to “DFM” was coined by Mr. LeBlanc (French) and Mr. Eli Whitney (American) in
1778 and 1788 respectively [Dummer, 1997] by proposing a system for the production of musket. It received an
industry wide recognition as “producibility through interchangeable parts”. Mr. Roger W. Bolz is credited for
organized DFM methodology [Bolz, 1958] as an alternative term for “producibility”, introduced in his book “The
producibility handbook”; however Design for Manufacturability “DFM” received industry wide acceptance around
1960 [Boothroyd, 1968]. Let us formally extend the DFM concept. In the SI it is defined as the ability to reliably
predict downstream life cycle needs and issues during early phases of design [Herrmann et al., 2004]. It is focused on
economic benefits from the volume production by trading off cost-quality-time triangle [Raina, 2006] and is
classified as [Mehrabi et al., 2002] product DFM (producing manufacturable designs within defined processes) and
process DFM (developing processes with less rework and high manufacturability). The most appropriate
classification of DFM methods in the SI is physical and electrical DFM [Appello et al., 2004]. The physical DFM
refers to the process variations that result in geometric shape drifts during manufacturing whereas electrical DFM is
focused on the characterization of parametric and functional product yields. The parametric characterization refers to
the extraction of key electrical and parasitic parameters whereas functional characterization mainly refers to the
signal timing and delays that result in the faulty products to ensure product functionality.
38
38
In 1980 the DFM concept was adapted as a yield enhancement strategy in the SI (Figure 1.8). It went very
well till 250nm technology [Cliff, 2003 and Radojcic et al., 2009] but after that the increasing complexity of the
circuit layout and shrinking sub wavelength lithography resulted in multiple respins and yield losses with 193nm
stepper and 130nm node (Figure 1.10). The introduction of the compensation techniques like Optical Proximity
Correction (OPC) and Resolution Enhancement Technique (RET) emerged as an extended design flow (DFM flow)
to mitigate the yield loss mechanisms. These methods are used during the mask data preparation to ensure
manufacturability and yield. Beyond this point the manufacturability and yield losses can only be controlled through
process control and recipe adjustments, however new systematic drifts patterns in the printed design layouts resulting
in parametric and functional yield losses can be modeled for the technology improvements.
39
targets, however the feedback link to improve the technology do not exist because design kits are frozen at this stage
and any proposed change shall need requalification of all design libraries which is costly and time consuming.
Equipment
Recipes
5 5
10 SPICE 10 Process Alignment Process
Technology Alignment
20 20
30
Models 30 Control Plans
Feedback
Technology Prototypes
Feedback
Interconnect Models Single Source Data
Multi Source Data
(CMP, SSTA, CAA)
(OPC, Litho, RET)
Design Rules
(VIA, Routing) Root cause analysis
DFM Models
DFM Rules
Weak Link
Limiters
Yield
DRM DFM
Model Alignment
DFM
DFM Model to Hardware Gaps
Contraints Strong Link
Lot/Wafer/Site
Site/Die/Position
Libraires
Limiters
Yield
MFD
DFM
Model to Hardware Gaps
Feedback
Technology Adaption
Feedback
Feedback
Design (CAD) Single Source Data
Multi Source Data
Simulations Product Prototypes Process Plan
Equipment
Recipes
Process
Process Alignment
Control Plans
No link Possible link
Model Alignment
Figure 1.11 - The role of DFM in technology alignment and adoption processes
It is evident that the effectiveness of R&D efforts depends on the input data which at present is the single-
source (lot/wafer/site); hence, our engineers are not able to exploit the huge data volumes collected across the
production line. The goal of this thesis is to find and remove the limiting factors associated with the single-source
root cause analysis and provide generic scientific contributions to enable multi-source (lot/wafer/site/die) and test
structure position based dynamic data exploitation. It shall result in efficient root cause analysis and effective DFM
methods. If the analyses results are quickly transformed into rules and/or models followed by its inclusion in the
technology models, the designs shall result in higher yield and manufacturability. It shall also result improvements in
technology alignment and adoption lead times and costs. So in order to put DFM back on track we suggest (i) a shift
from MFD to DFM efforts and (ii) single-source to multi-source site/die/position based root cause analysis within the
technology alignment and adoption processes.
40
40
Q1: What are the top ranked business objectives in the SI, today? What is the best strategy to achieve
these top ranked objectives? Do we have a business model coherent to achieve these objectives, if no
then what it should be?
We have seen that new technology, which is developed in an alliance, is transferred into the business model
where it follows technology alignment and adoption processes. The DFM methods can play a significant role in the
alignment and adoption that has become ineffective resulting in increased lead times and costs. It depicts our
inabilities to model the emerging drifts and variations, besides the availability of huge data volumes and dimensions
(multi-source). We believe that if we can find and fix the limitations in multi-source data exploitation, it shall result
in more effective DFM methods. So, the second research question is to identify the key limitations that result in our
inability to dynamically exploit the production data sources for R&D purposes. The second research question is
formally presented as under:
Q2: What are the true DFM challenges (limiting factors and failure modes) and respective root causes
within technology alignment and adoption processes?
The identified limitations are further investigated to find relevance with an organization or a domain; hence,
the third research question is about generic solutions (scientific contributions) to solve the problems associated with
dynamic exploitation of production data sources. The third research question is formally presented below:
Q3: What are the generic solutions to remove these root causes and put DFM back on track for the
technology lead time and costs improvements?
Finally, it is important to assess the potential industrialization and post industrialization challenges of proposed
generic solutions. The proposed solutions provide an opportunity to exploit multi-source data; hence, it raises need
for more metrology and inspection data for R&D purposes. It is likely to reduce the metrology/inspection capacities
reserved for normal production. We cannot buy new tools to increase the capacities (fixed cost); hence, we need an
intelligent way to spare metrology/inspection capacities for R&D purposes. The fourth research question is formally
presented below:
Q4: What are the consequence of proposed solutions industrialization and how we can resolve it to allow
smooth integration of the proposed solutions?
41
Figure 1.12 - The Research schematic and contributions at a glance
42
We started with brainstorming sessions (A0) with engineers and managers at STMicroelectronics and
highlighted potential business (T2M, T2V, ramp-up-rate) and technical (area, power, timing, leakage) objectives.
The SCAN analysis (A1) resulted in leadership position and ramp-up-rate as the top ranked objectives and traditional
IDM business model analysis (A2) resulted in the (i) fast technology transfer, (ii) manufacturing databases and (iii)
effective root cause analysis (R&D) as key improvement areas. These findings were used as input in SWOT analysis
(A3) and we agreed on the strategy to remove weaknesses and capitalize the knowledge. As we know that the
technology lead times and costs are the key factors for success in SI; hence, we analyzed technology derivative and
improvement processes (A4) to find limitations within the scope of key improvement areas as identified in A2. The
key failure modes were found as (i) data extraction, (ii) alignment and (iii) pre-processing due to ontology issues and
missing database links. This analysis resulted in the proposition of an extended IDM-fablite (e-IDM) business model
with integrated DFM methodology for quick ramp-up-rate and leadership position. The analyses blocks A0A4
shows that existing IDM, IDM-fablite, fabless and foundry business models do not support recent shift in the
business objectives (ramp-up rate); hence, we propose an extended IDM-fablite (e-IDM) business model to achieve
quick ram-up rate. It also improves the internal manufacturing window while keeping intact its backward
compatibility with technology, initially developed in an alliance. It provides an answer to the first research question.
We have used a well known Failure Mode Effect Analysis (FMEA) methodology to find potential root
causes against initially identified failure modes (A4). The initial root causes identified are ontology issues, missing
links between databases, missing values and varying measurement reference coordinates that restrict our engineers to
single-source root cause analysis. It is because of the fact that wafer is often rotated prior to site/die level
measurements due to test structure orientation in the scribe lines and fields. This rotation changes the x, y coordinates
resulting in varying measurement coordinates which requires an accurate alignment prior to perform multi-source
site/die level analyses. We developed 5 tools (industrial contributions: ICs) (i) Spice Model Alignment (SMA) for
fast technology transfer (IC1), (ii) BEOL-variance analysis to analyze parametric drifts in back-end-of-line
interconnect modeling process (IC2), (iii) KLA-Ace recipe for the mapping and alignment of coordinates to enable
multi-source site/die level analysis (IC3), (iv) Equipment, Product, Process (EPP) life cycle extraction tool to extract
multi-source contextual data to support drift/excursion analysis (IC4) and (v) Alarm Control and Management
(ACM) tool to extract, analyze and manage alarm data for tool capacity optimization (IC5). These tools were
developed and provided at the disposition of Technology to Design (T2D), Equipment Engineering (EE) and
European project IMPROVE teams, being crucial for the successful and quick technology derivative improvements
with local DFM efforts. The objective was to find new failure modes which are not evident at this stage.
While using these tools, the teams identified new failure modes that were taken into account and the tools
were rectified and updated accordingly. This hybrid approach helped us to find the core failure modes: (i)
unstructured data model evolutions, (ii) missing data dimensions and (iii) wrong correlation due to test structure
positions. These failure modes are grouped as (i) ineffective root cause analysis (infield and scribe line test structure
positions) and (ii) data extraction, mapping and alignment, were identified to propose generic solutions (A6 and A7).
We also performed FMEA on the technology transfer step and found issues between SPICE model and measured
electrical parameter relationship. The SPICE models are the mathematical equations used to extract the electrical
parameters based on the transistors and interconnect geometries and process technology selected for the
manufacturing. The name of the parameters in these SPICE models are generic whereas the names used by the test
engineers while writing test programs for the test structure on the wafer are different; hence, the mismatch creates a
significant problem and delays the model validation resulting in extended lead times. The initial root causes that were
found are incorrect and error prone SPICE-PT parameter alignment due to manual data alignment and pre-
processing; hence, we initially developed SMA (spice model alignment) tool (IC1) that resulted in automation and
removed the associated issues. The further use of this software tool identified unstructured naming conventions for
the metrology parameters and varying formats of the CAD simulation results as the key root causes. These industrial
contributions IC1 to IC6 are added in appendices for reference. The analysis blocks A6-A7 and industrial
contributions IC1 IC5 partially answers the second question, established in section 1.4.
A detailed FMEA analysis is performed on the (i) ineffective root cause analysis and (ii) data extraction,
43
mapping and alignment failure modes. The resulting root causes are used to propose generic solutions a.k.a. scientific
contributions (SCs). The key root causes linked with “ineffective root cause analysis” failure modes are (i) missing
and varying coordinate due to varying metrology reference frames, (ii) rotation of wafer prior to the measurement
steps and (iii) the infield/scribe line test structure positions. A generic MAM (mapping and alignment) model (SC1)
is proposed to remove die/site level mismatches and provide means for an accurate correlation. The SPM (spatial
positioning) model (SC2) is also proposed to enable correlation between parameters, based on the shortest distance
between test structures, used for metrology. The ROMMII (relational/referential ontology Metal model for
information integration) framework (SC3) and R&D Data Warehouse (DWH) model (SC4) are proposed to address
the causes identified against data extraction, mapping and alignment failure modes. The word mapping in this failure
mode refers to the mapping between multiple databases (links). It provides the ability to perform pre-failure analysis
on the potential impact of any structural change in data model over existing users and applications. These scientific
contributions (SC1 to SC4) refer to the question-3 whereas the identified failure modes and root causes complement
the answer to question-2. It is very important to note that root causes identified fall in different business functions and
in our case it is Information Technologies (IT) and Engineering Data Analysis (EDA), so based on the experience
during PhD, we have also proposed a 4-step i-FMEA (interdisciplinary failure mode effect analysis) approach (SC5)
to capture true challenges that might fall in other business functions. The i-FMEA methodology completes our
answer to question-2.
At the end, a brain storming session on the industrialization of proposed scientific contributions highlighted
potential impacts. It was found that it shall result in extended analysis and metrology demands by R&D engineers for
an effective root cause analyses. It is likely to reduce inspection/metrology tools capacities for normal production
lots; hence, to address this issue a yield aware sampling (YASS) strategy (SC6) is proposed which is based on
information fusion (alarms, states and meters data) to intelligently predict the production lots with likely yield loss.
So the objective is to inspect bad or suspected lots and allow good lots to move to the next production steps. The
additional metrology and inspection capacities can be then used for the R&D purposes. This YASS strategy provides
answer to the question-5.
b) BEOL Variance Analysis Tool (IC2): This tool is developed for Technology to Design (T2D)
team to support the device/interconnect modeling process by quickly pre-processing data followed by
parasitic variance analysis that result in the parametric and functional yield losses. The significant
variations are further analyzed using KLA-ACE recipe for the PT-Inline site level correlations to find
root causes against these drifts.
c) KLA-Ace Recipe for PT-Inline Correlation (IC3): In this recipe we enable multi-source
correlation using PT-Inline data sources by mapping the data at site levels. Both of the data sources,
Parametric Tests (PT) and Inline measurements are captured at Site level. The PT is a type of electrical
44
measurements whereas Inline corresponds to the physical measurements. The PT database has site
numbers and x,y coordinates against each measurement; however, the Inline data has only site number.
These site numbers cannot be used for site level mapping because during measurements the Wafer is
rotated based on the position of the test structure; hence, every measurement has different coordinates.
It means that until and unless we have wafer position defined by notch (a cut on the wafer edge used
for alignment) and its x,y coordinates, site level mapping cannot be performed; hence, multi-source
root cause analysis is not possible. In this recipe we use the mask level information to normalize the
site level identifiers so that quick mapping can be performed to provide engineers an ability to perform
multi-source root cause analysis.
d) EPP (Equipment, Product, Process) Life Cycle Tool (IC4): The results obtained from the BEOL
and KLA-Ace recipe justify the parasitic drifts against geometric specification variations; however, it does not
provide an answer if the drift was caused due to process or equipment variation. In order to perform an in
depth analysis we developed this tool that exploits the manufacturing data sources and generates product and
equipment life cycles. The EPP tool directly connects with the maintenance (TGV) and out of control (OOC)
databases for the equipment related data extraction whereas product and process data is extracted from the
process database using KLA-Ace recipes. All these data are input to the EPP tool that perform consistency
checks and populates them into a multidimensional database. It provides a user friendly interface to extract
customized equipment and product life cycles.
e) ACM (Alarm Control and Management) Tool (IC5): This tool is developed for the lithography
equipment engineering team. At present engineers do not have information about alarms categorization based
on machine states; hence, we have linked INGENIO (equipment alarms) database with maintenance and
process databases to classify alarms based on the machine state. The extracted data is further used to develop
predictive models for the likely yield loss across the production processes with an objective to skip good lots
and inspect the lost with suspicion of bad yield. It is used in the YASS (yield aware sampling strategy) as
presented in section 1.6.2e.
These tools have helped us in finding true DFM challenges (failure modes) and associated root causes as a
part of our proposed i-FMEA methodology. The brief description for these tools is presented in the appendices D to
H.
a) MAM (mapping and alignment) Model (SC1): It is a generic model for site/site and die/die
mapping of metrology data along with die/site qualification. The objective is to enable our R&D
engineers to perform multi-source root cause analysis to find root cause against newly emerging
spatial drifts and variations. It is presented and discussed in detail in Chapter-5.
b) SPM (spatial positioning) Model (SC2): It is a fact that all metrology measurements are
performed on test structure and these test structures are located in scribe lines and/or in the fields. To
capture and better understand newly emerging spatial variations, it is likely to perform analysis using
measurements coming from the test structures with shortest distance. This model performs mapping
on different data sources based on the test structure position.
45
models to include new data dimensions. The R&D data model is proposed to support the need of
R&D engineers to ensure that 1 year R&D data is always available for analysis purposes.
e) YASS (yield aware sampling strategy) Strategy (SC6): This intelligent sampling strategy
predicts the production lots as good, bad or suspected lots based on the predictive state and alarms
model. These models are learned from data extracted using ACM tool. The objective is to increase the
metrology capacities so that it can be spared for the R&D purposes. It is important to support the
industrialization of our proposed generic scientific contributions because of the fact that ability to
model abnormal drifts shall result in huge demand for more metrology for R&D analysis.
46
model for information integration) framework is proposed that allows the data model evolution and enables us
to exploit huge data volumes as well as inclusion of new data dimensions. The proposed R&D data model
removes different data retention issues with multiple production databases.
Chapter-7 [YASS (yield aware sampling strategy) for tool capacity optimization]: The
industrialization of proposed contributions (chapter-3, chapter-5 and chapter-6) does result in new demands for
the additional metrology to capture and model variations. It is quite difficult in an IDM business model to spare
metrology/inspection capacities for R&D purposes; hence, the YASS strategy spares tool capacities for R&D.
The proposed strategy advocates and facilitates industrializations of e-IDM fablite, MAM, SPM and ROMMII
contributions.
We conclude our thesis with discussion and critical analysis on the proposed solutions against alternative solutions.
We also present some industrial recommendations for SI on the potential industrialization of the proposed scientific
contributions and end up with future research directions to improve the DFM methods.
Appendix A (Publications): provides a list of publications in international conferences and journals along
with abstracts and keywords.
Appendix B (Semiconductor Design, Mask and Manufacturing Processes): provides a brief review
on design, mask preparation and manufacturing processes along with key challenges and limitations and most
common manufacturing and yield loss mechanisms.
Appendix C (CMOS Inverter Design and Manufacturing: practical example): simulates the
CMOS inverter design, masks and manufacturing steps.
Appendix D (SMA-Tool): presents a brief description of SMA tool used by the PI team to support fast
technology transfer efforts.
Appendix E (BEOL-Tool): presents a brief description of the BEOL-Tool for T2D team to support quick
variance analysis during interconnects modeling process.
Appendix F (KLA-Ace Recipe): presents a KLA-Ace Recipe that uses the mask data to complement the
missing x, y coordinates for inline data and enables PT/Inline site level correlation analysis.
Appendix G (EPP-Tool): presents brief description of the tool that provide product and equipment life cycle
data extraction, used by IMPROVE project team for the computation of Equipment Health Factor (EHF).
Appendix H (ACM-Tool): presents brief description of the tool that provides extraction of alarms and states
data to be used for PAM and PSM prediction models. It is used by the equipment engineers to analyze and
manage the alarm controls to identify the alarms to be controlled on priority.
This thesis ends with a list of references used during research and an index of important terms. To help the readers,
the first paragraph of each chapter provides a synopsis of that chapter’s contents.
Contents
2.1 Semiconductor Industry (SI): Background and Challenges ............................................ 51
2.1.1 Historical Background ....................................................................................... 51
2.1.2 Evolution of Semiconductor Industry ................................................................. 52
2.1.3 Role of Moore’s Law in Semiconductor Industry ............................................... 53
2.1.4 Nanometer vs. Micrometer Semiconductor Technologies .................................. 55
2.1.5 Semiconductor Business Model and Evolution .......................................................... 56
2.1.6 Trends in Semiconductor Industry .......................................................................... 57
2.1.7 Challenges Faced By Semiconductor Industry ........................................................... 58
2.2 Role of DFM Methods in the Semiconductor Industry and Evolution. ........................ 61
2.2.1 SI Challenges and Rise of Interest in DFM. ..................................................... 62
2.2.2 A Comparison of DFM Efforts in SI and Manufacturing Industries. ........................... 62
2.2.3 DFM Techniques (Pre-1980 Era). ...................................................................................64
2.2.4 Adaption and Diversification of DFM to SI (Post-1980 Era). .................................... 66
2.2.5 DFM Challenges and ECAD/TCAD Tools. ............................................................. 69
2.2.6 Increasing Design Size and DFM Realization Challenges......................................... 70
2.2.7 Role of SI Business Models in DFM Evolution and Adaption. ................................... 71
2.2.8 Industry Wide Understanding of the DFM Concept.................................................. 72
2.3 Information Integration Challenges Towards more Effective DFM Methods ............... 74
2.3.1 Data/Information Integration Issues ........................................................................ 75
2.3.2 Ontology from Philosophy to Computer Science ........................................................ 75
2.3.3 Data/Information Integration .................................................................................. 75
2.3.4 Ontology Based Database-Integration Approaches .................................................... 77
2.3.5 RDB Integration Based on Schema Matching ............................................................ 78
2.3.6 RDB Schema to Ontology Mapping Approaches ........................................................ 79
2.3.7 Ontology Driven Data Extraction Tools ................................................................... 79
2.4 Summary and Conclusions ........................................................................................... 80
49
50
2.1 SEMICONDUCTOR INDUSTRY (SI): BACKGROUND AND CHALLENGES
The Semiconductor industry (SI) is characterized by the fastest change in smallest period of time and has
become a 300+ B$ industry in less than 60 years [Stamford, 2012 and Dale, 2012]. It has followed the
industrial slogan smaller, faster and cheaper in compliance with Moore’s law that predicted the doubling of
transistor every 18 to 24 months [Moore, 1998]. This miniaturization has led serious design and
manufacturing challenges which are resolved with the continuous introduction of new technologies,
processes and production equipments at the cost of huge R&D investments. The increasing R&D costs have
been compensated by increasing wafer sizes and reducing technology alignment and adoption lead times.
The SI has seen many circuit integration eras and series of technology nodes to manufacture more complex
integrated circuits. The DFM methods were initially introduced around 1980 as a yield enhancement strategy
that worked very well up until the 250nm technology node but beyond this it became a high cost R&D
activity. The newly emerging variations have resulted in increasing the technology alignment and adoption
lead times and associated costs. We need to reduce these lead times so that associated R&D costs are reduced
and high profit gains with early penetration into the market.
The objective of this section is to briefly review the history of the semiconductor industry, evolution,
business models, current trends and major challenges. It shall provide us with an overview of the market and
its influence on the semiconductor industry, further we shall also see the evolution of technical and business
challenges that led the structural transformation of the IDM structure into fablite and fabless business
models. This section shall highlight the need to analyze the existing evolutions and find or propose a
business model (see chapter-1, section 1.4) to achieve the new shift in objectives from T2M and T2V
towards ramp-up rate a.k.a. Time-to-Quality (T2Q), essential for success [Carrillo and Franza, 2006].
51
a) First Transistor (1947) b) First IC Device (1959) c) Silicon IC Chip (1961)
Bell Labs Texas Instruments Fairchild Lab
Figure 2.1 - Benchmarks in semiconductor technology evolution [Quirk and Serda, 2000]
2.1.2 Evolution of Semiconductor Industry
The demand for semiconductors is mainly driven by end-user markets: data processing, consumer
electronics, communications, automotive industry and industrial sector [Ballhaus et al., 2009]. The SI forms
a part of this complex interaction among multiple industrial sectors [Yoon et al., 2010 and Kumar, 2008]. In
general, the semiconductors demands do not generate directly from end users, but it is determined by the
related end-customer market. We can analyze the market supply and demand by dividing it on product type,
application segments and geographical regions.
Understanding market structure based on products is better understood as two categories as (i) non
memory products and (ii) memory products (Figure 2.2). It is evident that memory is the biggest market in
semiconductor till today. The Microprocessor (MPU) and Microcontroller (MCU) Units have lowest relative
market demand but these products are classified as high value products; hence return on investment is also
high on these products.
100,000
50,000
0
2000 2001 2002 2003 2004 2005 2006 2007 2008 2009 2010 2011
1 The data is used from International Business Strategies, Inc., report: Key trends in technology and supply for advanced features
within IC industry (October 1, 2009), Gartner {www.gartner.com} and isuppli {www.isuppli.com}
52
350,000
Other (M$)
100,000
50,000
0
2000 2001 2002 2003 2004 2005 2006 2007 2008 2009 2010 2011
120,000
North America (M$)
Japan (M$)
100,000
Europe (M$)
Others (M$)
80,000
China (M$)
60,000
40,000
20,000
0
2000 2001 2002 2003 2004 2005 2006 2007 2008 2009 2010 2011
4,2 The data is used from International Business Strategies, Inc., report: Key trends in technology and supply for advanced features
within IC industry (October 1, 2009) , Gartner {www.gartner.com} and isuppli {www.isuppli.com}
53
We saw that this projection proved wrong and Moore’s 2nd law still prevails; however the cost associated with new
technologies is exponentially increasing.
During last 50 years, the SI has followed the industrial slogan smaller, faster and cheaper that resulted in
shrinking the device geometries, increasing chip sizes (Figure 2.5) and rapid improvements at reduced costs. The
period from 1960 to 1970 is credited for the emergence of new technologies that lead the Small and Medium Scale
Integrations (SSI, MSI) but actual competition started around 1970 with Large and Very Large Scale Integrations
(LSI, VLSI) to capture maximum market share with the support of automation technologies that were introduced
around 1980 [Quirk and Serda, 2000]. The circuit integration era beyond 1999 is marked as Ultra Large Scale
Integration (ULSI) and is attributed to the start of volume production efforts [Kumar, 2008]. These circuit
integration eras are attributed to the Moore’s law that has pushed the SI towards smaller, cheaper and faster devices.
Figure 2.5 – Circuit integration eras and reducing costs (Chang and Sze, 1996)6
The well-known Moore’s law has diversified into (i) equivalent scaling and (ii) functional scaling as
presented in Figure 2.6 [Kahng, 2010]. The equivalent scaling is based on the principle of miniaturization that has led
the emergence of System On Chip (SoC) where functionalities performed by different chips e.g. memory, IO (input
output), processing unit etc. are put on the same chip to get higher performance at reduced area and cost. The major
SoC application areas are information and digital contents processing e.g. memories (ROM, RAM, EEPROM,
Flash) microcontrollers, microprocessors, Digital Signal Processors (DSP) etc. It has resulted in the design
complexity and requires new technologies to be regularly introduced to cope up with the pace defined by Moore’s
law. The functional scaling is an important area for the advancement of semiconductor technology. It is focused on
grouping multiple functionalities (circuits systems) in a single chip (hetero-integration of digital and non-digital
contents). These systems are also known as System in Package (SiP). The current trends include functional and
equivalent scaling at the same time which has resulted in the emergence of new type of systems called Package on
Package (PoP). Multiple chips are stacked on one another in the PoP which give rise to more complex design and
manufacturing interface complexities (variations). The most common examples for such systems are the mobile
phones, digital cameras and Personal Digital Assistants (PDA). These systems require innovation in terms of new
devices as well as materials; hence results in the most complex manufacturing environment.
6 Redrawn from C. Chang and S. Size, McGraw-Hill, ULSI Technology, (New York: McGraw-Hill, 1996).
54
Figure 2.6 – Diversification in Moore’s law [more Moore and more than Moore]
55
Figure 2.7 – Technology size scale, how small is small? [Jones, 2012]
Fabless Yes
Foundry Yes
7 The acronym EDA stands for engineering data analysis whereas electronic design automation is always written in full.
56
The production capacities are also outsourced to the technology alliance if required. It is a modern IDM model
which is still holding the maximum market share e.g. Samsung, Toshiba, STMicroelectronics.
Fabless and Foundry Model: This model is being promoted by ITRS based on the fact that the new
emerging technology challenges require strong collaboration in this competitive environment to reduce the
lead times and costs. The fabless companies do not carry out the product development; they simply design the
products based on the technology which is developed in an alliance. The designs produced are equally
manufacturable in the technology alliance partners manufacturing facilities. These companies have been
successful as there is no capital investment required but the success of these companies relies on IP circuits. In
comparison to these fabless, the foundries mostly operate large and modern production facilities with high
levels of capacity utilization e.g. TSMC, UMC etc.
EDA and IP Companies: The EDA companies are involved in the design phase of the value chain and
they operate in strong collaboration with IDM, fabless and foundries. These companies devote themselves
exclusively to licensing (intellectual property or IP companies) and they specialize in the design of certain
modules and license the resulting intellectual property to their customers. Unlike fabless companies, IP
companies do not have sales operations and license their design and development services exclusively to third
parties. There are also companies that focus on electronic design automation. Compared with the other
business models, the volume of sales generated by IP and EDA companies is a small part of the overall market
but without any initial investment.
The reasons behind structural transformations have been reported as potentially increasing technical and
business challenges that have resulted in increasing technology lead times and costs. The objective is to capture the
maximum market share; hence the shift in business objective from T2M and T2V towards ramp-up rate a.k.a. T2Q is
attributed to be the key driver in this transformation. The fabless and foundry business models are being promoted
since 1999 [Kumar, 2008] where design, manufacturing and electronic design automation are separated as the core
business functions and emphasis is put on the collaboration and alliances to address the exponentially increasing
costs. IDM business model has been demonstrated as the best business model in this competitive environment
because it is coherent to reduce the technology alignment and adoption lead times [Shahzad et al., 2011a] and
benefit from the cyclic demand patterns to capture maximum market share. We believe that success lies in our ability
to quickly ramp-up the production that requires a mature technology and continuous improvements. The IDM
platform provides R&D engineers with all the data to be used for the multi-source analysis. The problem today is
that engineers - even with the availability of large data sets - are not able to exploit this data, resulting in ineffective
DFM methods. Removing these challenges to utilize the data while performing the technology alignment and
adoption activities will support the lead times and cost reduction efforts.
57
the technology scaling and the increasing wafer sizes. It is usually carried out in the equipment alliances to
share R&D costs and provide the partners an early penetration in the market with new technologies and
wafer sizes. It is evident that such high R&D costs even if shared, result in higher entry costs (Figure 19-f)
for smaller chips but it is compensated with the volume production resulting from the increasing wafer sizes
[Kumar, 2008 and Quirk and Serda, 2000].
58
Besides the cyclic downfalls, [International Business Strategies, 2009] investment in R&D within
the SI has continuously increased and is expected to reach 18% of total revenues by the year 2012 (Figure
2.9a). Major portion of these R&D costs is focused on the development of new technology platforms (2-7
billion USD from 65nm to 22nm) and IP/library qualifications for predictable design/manufacturing
interfaces. Major designs respin reasons (Figure 2.9b) faced by the SI are leakage (87%), I/O functionality
(50%), verification bugs (27%), design rule violations (12%) and testability issues (8%). One of the
important challenges faced by SI is exponentially increasing costs during technology derivative
improvements (alignment) and adoption (product design/development) efforts (Figure 1.2). The
design/debug periods for the 90nm, 65nm and 45/40nm are 44/9, 48/23 and 60/40 weeks respectively and
prototype/volume production takes 8/10 weeks respectively for all nodes (Figure 2.9c).
59
c) Design, debug and prototype periods
Figure 2.9 – Major design challenges faced by semiconductor industry [International
Business Strategies, 2009]
The IC industry is in a highly challenging phase and IC business continues to have many
opportunities with IP (intellectual proprietary) and cell libraries (reusable circuit components) qualifications.
These IPs and reusable components reduces the product design and development time subject to the ability to
quickly adopt the given technology against new products. The key for success lies in the ability to quickly
design, develop and ramp-up the production which at present is challenged by the manufacturing variations.
The DFM (design for manufacturing) methods are the best tools to quickly model these manufacturing
variations to reduce the technology alignment and adoption lead times for economic benefits (Figure 1.4).
Let us conclude this section and formally propose the research question-1 (chapter-1, section 1.4).
The market is characterized by the demand that depends on the population and economy. The world
population and demands are continuously increasing whereas the demand patterns depend on economic
cycles. The CAGR (+8.72%) guarantees a cumulative demand; hence the SI is obliged to respond to the
growing market with new, faster and high value but low cost products. It has led a shift in business objective
from T2M and T2V towards T2Q and the emergence of design and manufacturing interface complexities.
The SI can only respond the market growth by introducing new technologies every 2 to 3 years, but the R&D
cost for the development of new technologies has exponentially increased. In order to address the increased
technology lead times and R&D costs, SI model (IDM) has transformed into a fablite and fabless business
models. But still, IDM fablite model is reported to be the best models in terms of revenue generation. The
IDM fablite model provides a coherent platform for the knowledge capitalization from production line and
our ability to improve the challenges faced by the R&D engineers shall result in the continuous improvement
efforts for the technology alignment or adoption within the business model. It shall reduce the lead times,
associated costs along with an opportunity for early penetration in the market with higher profits.
The SI is a 300+ B$ industry with an equal opportunity to capture maximum market for its stake
holders. In order to maximize maximum share, the SI objectives need to be revisited and ranked to formulate
a strategy that is truly in line with the top ranked objective. This can be accomplished by further analyzing
the existing business models (IDM, Fablite and Fabless) to either find or propose an extension in the existing
60
models to get maximum market share in line with the top ranked business objectives. These arguments
formulate the first research question, which is addressed in chapter-3.
Readers who are not familiar with the SI design and manufacturing flows should read the appendices
B and C before moving to section 2.2. These appendices provide a brief review on the design, mask
preparation and manufacturing flows in the SI along with a CMOS inverter example that simulates these
steps. The objective is to better understand manufacturability and yield limitations prior to investigate the
DFM ineffectiveness beyond the 250nm node. It is to remember that the DFM ineffectiveness has resulted in
an exponential increase in the technology development costs followed by their extended alignment and
adoption lead times. It is a challenge towards accomplishing quick ramp-up rates with the existing SI
business models. The next section reviews literature about DFM and its evolution in the SI to find reasons
for the DFM ineffectiveness.
61
2.2.1 SI Challenges and Rise of Interest in DFM
Initially DFM efforts were based on rough estimates of the downstream effects and rest was expected to be
controlled through advanced process control (APC) and advanced equipment control (AEC). This worked
well until 250nm but after that increasing complexity of circuit layout and shrinking sub wavelength
lithography eventually resulted multiple respins and yield losses. The 130nm node is considered the cut-off
point where the need for DFM was felt to tackle increasing feature and design limited yield losses [Cliff,
2003]. From the designer's perspective, things are getting more difficult because the process windows they
are getting back from manufacturing are so tight that they are having a hard time getting the design
methodologies to work [Peters 2005 and Dauzere-Peres et al., 2011].
The concept of DFM has emerged in a diversification of terms like DFY, DFV and DFT etc. but all
terms come under the umbrella of DFM along the product life cycle (PLC) as DFX having similar objectives
of cost, quality, yield and time-to-market [Anderson, 2006] where X refers to the various stages in Product
Life Cycle (PLC). As the new Design For All (DFX) methods are being explored, the definition of DFM has
become synonymous with DFX and [Herrmann et al., 2004] concurrent engineering (simultaneous
development of a design and the supporting life cycle processes). These methods could be written rules or
simulation tools for cost/performance estimation. The DFX tools are focused to provide designer,
predictability information on multiple issues within down side of the product life cycle. These tools are
applied directly on the CAD designs and the provide advice on the product performance (qualitative,
quantitative and binary).
The design and manufacturing interface modeling has been straightforward till 250nm node [Cliff,
2003 and Radojcic et al., 2009]; however hardware to model gaps (variability) started increasing with the
industrial shift towards 193nm stepper to manufacture features less than the wavelength of light source. The
compensation techniques like OPC and RET emerged as an extended flow (DFM flow) to mitigate these
manufacturability and yield loss mechanisms. The efforts to move towards 157nm wavelength light source
have been abandoned; hence we have to live with 193nm stepper for the smaller process nodes e.g. 45nm,
32nm, 22nm etc. It is one of the major causes for manufacturability and yield related issues and we have seen
the rise of interest in DFM methods followed by aggressive use of compensation techniques like OPC, RET,
Phase Shift Masking (PSM) and immersion lithography [Venkataraman, 2007] during design flow.
From the above facts, it is evident that stretching CMOS technology for the smaller process nodes
without extending the traditional DFM methods is not possible. It is important that we put DFM back on
track because of the investments made in the equipment, material, process and design (IP/libraries) for
economical benefits. The DFM methods in last two decades has evolved from design rules to DFM rules like
(layout/routing) and DFM models like CAA, CMP, Shape, Yield, Leakage and SSTA to mitigate yield losses
[McGregor, 2007]. The new business objective of the SI “ramp-up rate” links our success with the first time
correct design or a design that could be ramped-up without yield loss. Increasing complexities and newly
emerging spatial variations along with the associated costs has been the core reason for the structural
transformation of the SI business models. We need to put DFM back on track for the future smaller nodes,
otherwise the Moore’s 2nd which was predicted to end in 2005 [Ross, 2003] shall become true now (see
section 2.1.3).
62
sensitive are evaluated against yield loss limiters and are strongly impacted by the Computer Aided /
Integrated Manufacturing (CAM/CIM) tools and the manufacturing system (DML, FMS, RMS).
63
processing; hence it is critical to find out the potential issues as early as possible in design flow to save the
production and inspection capacities. The process based DFM initiatives in SI can be referred as a
technology; however its maturity has a strong impact on the potential yield and manufacturability. A
common approach in DFM philosophy as stated by [Das and Kanchanapiboon, 2009] is the use of parametric
models and knowledgebase library which is similar to device and interconnects Simulation Program with
Integrated Circuit Emphasis (SPICE) models and design rules.
In manufacturing industry we are focused on integrating DFM efforts within the design flow to
improve effectiveness of DFM philosophy [Vliet and Luttervelt, 2004]. It is exactly the same objective that
we are following in SI to put DFM back on track against its emerging ineffectiveness, hence
manufacturability and yield related issues can be attributed to incapable or ineffective DFM methods. The SI
is focused on building design libraries (pre-simulated circuit reusable components) that are used by the
designers during the design process. It is assumed that all the manufacturing data is available [Myint and
Tabucanon, 1998] for its use in the automated DFM evaluations but at present besides the huge data volume
our engineers are not able to exploit it for the root cause analysis against newly emerging drifts or variation
phenomenon. The biggest differences we can find between the semiconductor and manufacturing industries
in the evolution of DFM methods are (i) the change of pace for the new technologies and in SI it is
characterized by the fastest change in the smallest period of time and (ii) the inability to exploit the huge data
volume and dimensions in SI that result in increased technology lead times.
We can see that objectively both industries use the DFM philosophy to find manufacturing and yield
related issues as early as possible. The SI is different than other manufacturing industries in the sense that it
is characterized by the fastest change in the smallest period of time, hence our success lies in our ability to
quickly analyze the emerging spatial variations and transform them into rules and/or models. It can be seen
that DFM evaluations in both industries are supported with the automated CAD tools but in the SI, R&D
engineers are not able to exploit huge data volume and dimensions to find root causes against new variations
and drifts. This ineffectiveness in the DFM methods has resulted from our inability to exploit these sources
and is resulting in high costs and extended technology lead times.
64
[Anderson, 2006]. The Sequential design flows are not appropriate for quality improvements and cost
reduction, Design For Manufacturability and Assembly (DFMA) methodology with loopback option is a
solution guaranteed for the reduction in cost, assembly time, parts in assembly and time to market. The term
DFMA was coined by Boothroyd as a methodology [Boothroyd, 1994] and implemented as a simulation tool
that proved beneficial for manufacturing industries in terms of cost, design spin and cycle time reduction.
The generic DFMA flow (Figure 2.11) defines the scope of initial DFM efforts limited to simplify the
product structure followed by material and process selection from the existing knowledge. It is important to
note that accurate technical and economical prediction is not possible until the geometry and specification of
the products are complete [Herrmann et al., 2004]; however during DFMA cycle, prototype validate such
predictions and ensures smooth transition toward volume production.
These facts are equally applicable to the SI as well, for example if we look at the IC design the CAA,
Hotspot and SSTA analyses are applied only after the transistor level synthesis where the geometric shapes
of the devices and interconnects are well defined. The cycle time reduction by simplifying the product
structure is less applicable in SI, however layout optimization is performed in order to avoid manufacturbility
and yield issues. We have seen that technology plays a critical role in lead times and costs and success
depends on its matruity which can be attributed to DFM effectiveness.
Figure 2.11 - DFMA (design for manufacturability and assembly) [Herrmann et al., 2004]
The product, process and cost models are proposed for a detailed manufacturability analysis
[Ramana and Rao, 2002]. The product and process models describe product as design parameters (geometric
and non geometric information) and process as the capabilities associated with geometric and non geometric
parameters respectively. The cost models are classified as the scaling, activity-based and statistical models
which provide precise cost estimation based on the detailed design. The measures for the manufacturability
are classified as binary (0/1: if one or more DFM guidelines or rules are violated [Zuozhi and Jami, 2005],
design is not manufacturable), qualitative (difficult, easy or moderate with an ordinal ranking among
candidate designs) and quantitative (cycle time and cost against allocated budget and time). The
manufacturability assessment is performed at two stages, high level (parametric matching) and detailed level
(interpolation/extrapolation). The parametric matching system uses a large database of processes
characterized by capabilities whereas matching is done while designers provide limits to the design attributes
for appropriate candidate processes.
65
The manufacturing system palys the most important role in achieveing DFM objectives [Koren et
al., 1999 and Mehrabi et al., 2002]. In production, the DML is a fixed automated line used to produce the
core technology parts in volume with low cost per unit and is successful until demand exceeds supply. The
FMS is focused on product variety with changeable volume/mix whereas RMS suport rapid change in system
structure adjust for the production capacity in response to product mix. The DML and FMS are systems
whereas RMS is a dynamic system that combines DML’s high throughput and FMS’s flexibility. In SI, all
the equipments used fall in RMS category; however they are limited by the size of the wafer being used for
production e.g. production lines for 200mm wafer and 300 mm wafers. In these production lines we can use
multiple technologies by process and equipment adjustments. The production database are the key source for
first hand knowledge to provide an accurate feeback to design, process and product engineers. It provides
means to find the root causes against abnormal behaviors. It can be concluded from the above discussion that
DFM effectiveness depends on the ability to access and exploit production databases. Integrated DFM tools
have also been proposed to benefit from the available manufacturing data because the relevant information
extraction and availability to apply DFM is a challenging part [Eskelinen, 2001]. A universal data model
with an open access technology is needed as design flows have turned up data-centric with the inclusion of
extreeme RET [Cottrell and Grebinski, 2003] techniques.
The above proposed solutions are industrialized and open access technologies are developed for the
extraction of relevant information. These systems worked well and have led the emergence of EDA
(engineering data analysis) companies to develop these technologies. The problem is that these tools become
obselete quickly and the customers are required to purchase the new upgraded versions. New equipments and
metrology techniques provide additional information on production processes, but propriatry nature of
production data sources, extraction and analysis tools restricts our ability to include additional information in
existing data analysis framework which results in ineffective DFM methods.
Rel/Qal Rel/Qal
Figure 2.12 - Early semiconductor design flows with loop back [Reid, and Sanders, 2005]
66
Let us discuss the general IC design [Reddy, 2000] flow (Figure 2.13). It starts by transforming
customers’ basic requirements into design specifications followed by behavioral description (interface and
functional block diagrams) which is simulated to ensure target functionality. The most commonly used
languages at this stage are Hardware Description Language (HDL) and Verilog that offers the advantage of
concurrency over other high level programming languages. The design is further detailed out at functional
and schematic levels by gate level synthesis using Register Transfer Level (RTL) language at
structural/schematic levels. The netlist is generated in step-5 using automated EDA tools. It follows the
timing and delay simulations on netlist where process, timing and delay models are taken from intended
processes. The objective of the timing and delay simulations is to assess functionalities as well as
manufacturability with Layout Versus Schematic (LVS) checks. The Design Rule Check (DRC) is
performed before LVS to ensure the designs compliance with Design Rules (DR) for a given technology. It
verifies spacing, shapes etc. of devices and interconnections to ensure manufacturability of a design during
production. These steps are repeated until all DR violations are removed and the parametric performance is
achieved. The final design is extracted in Graphic Database System (GDS) format, also known as tape-out. It
is similar to netlist with additional information and it is generic enough to be used within any manufacturing
facility for prototyping or production.
67
base, expert system and concurrent system design elements; are developed with an objective to trade-off
cost, quality and yield for the integrated product design and process selection [Yuyin et al., 1996].
Assembly cost
Maintenance cost
Economic
Productivity
criteria Labor in
Maintenance
Labor in
Labor
Manufacturing
Material Std.
Machinability Material
Shape Complexity
IPD Planning
Manufacturing
process / Equipment
Wafer Fabrication
Device Design
Figure 2.15 - Integrated product development framework [Cho and Hsu, 1997]
Ultimate diversification of the DFM within IC industry has emerged as a yield improvement strategy
focused on an ideal objective ‘first time correct design’ achieved by proactively mitigating design driven
(sub-micron lithographic printability issues) and defect driven (process or technology issues) yield losses
[Redford 2009 and Raina, 2006]. We need to proactively solve systematic defect driven yield related issues
using DFM methodology in an efficient and effective manner from cell to chip level. These methodologies
are based on data mining and are applied [Ouyang et al., 2000 and Schuermyer et al., 2005 and Appello et
al., 2004] to learn yield limiters but the extent to which such methods can be applied on the production lines
is still a question; however they highlighted features with high correlation and those which were predicted to
have no correlation, turned up with strong correlation requiring correction in the models. The DFY elements
68
so far taken into consideration are grouped under physical DFM (via and wire optimization, metal density
uniformity and cell swapping) whereas areas which require to be integrated are the electrical DFM (model
based guidelines) to continue CMOS scaling [Raina, 2006].
Before we proceed further, let us summarize the discussion by defining the scope of DFM rules and
or models [DFM Dictionary, 2004] within design and manufacturing flows (Figure 2.16). The presented flow
clearly demonstrates that once the design is ready, it moves to the mask data preparation step which is the
second line of defense against the manufacturability and yield loss mechanisms. Both scribe and device level
Boolean operations are performed along with aggressive RET compensation. The design moves to the
production where it is controlled using the advanced process and equipment control methods (MFD). The
design rules, DFM rules and model improvements are based on the data collected from the production lines.
The analysis against the significant drift updates the rules and models to ensure manufacturability and yield
gains. It can be concluded that this analyses is dependent on our ability to access and exploit the production
databases which is useless if the data models are not updated continuously with new data dimensions.
Figure 2.16 Scope of design rules, DFM rules and DFM models [DFM Dictionary, 2004]
69
collaborative engineering and data flow between design and manufacturing [Morinaga et al., 2006].
Proposed architecture suggests distributed relational databases where interface layer is focused on providing
the alternative naming conventions and data sharing across design and manufacturing flow.
Most semiconductor manufacturers expect 193nm immersion lithography to remain the dominant
patterning technology at 32nm and 22nm technology nodes [Venkataraman, 2007]. The primary goal of
DFM is to enlarge process yield window [Raina, 2006] while the primary goal of MFD (APC/AEC) is to
keep the manufacturing process in that yield window. The RET to overcome diffraction effects including
OPC, PSM and immersion lithography are being stretched, resulting in restrictions on layout and an
explosive growth in the number of design rules being employed to enable manufacturable designs. The DFM
techniques that analyze design content, flag areas of the design that limit yield, and make changes to improve
yield are being developed and employed [Venkataraman, 2007]; however it is extremely difficult, if not
impossible, to predict all the problems that are likely to occur and what impact they will have on a product.
From the above discussion, we extend our argument that we are objectively focused on fining
challenges and limitations that resulted in DFM ineffectiveness followed by generic solutions to resolve
these issues to put DFM back on track. In this thesis we are not working on improving the simulation time
with extended design and DFM rules; however we are working on the extraction of these rules and
improving the DFM models. Please refer to Figure 2.40, the feedback loop presented here is meant for the
new technology which is frozen upon its transfer from an alliance and we are not authorized to change it. To
be more competitive, we have to extend these feedback loops for the technology derivative or improvement
initiatives (alignment) and product design and development (adoption). It requires efficient data access,
dynamic exploitation and effective root cause analysis.
70
Factors that lead the realization of DFM concept are (i) the CAD/CAE tools to ensure proper
linkage between the design and manufacturing, (ii) relational databases and computer networking
capabilities for efficient and effective data exploitation, (iii) process health feedback (SPC) with possible
preventive maintenance predictions and (iv) the yield/performance trade-off modeling and simulation tools.
71
significant shifts of Moore, more Moore and more than Moore has benefited 5 technology shrinks from
180nm to 45nm within last decade resulting from improvements in (i) Lithography (248 nm, 193 nm, 193 nm
immersion) (ii) Materials (Al, Cu, low-k, high-k dielectrics, metal gate, resist chemistry) (iii) strain
techniques (stress liners, embedded SiGe, shallow trench isolation) and (iv) Sensitivity to random
effects/defects (dopant fluctuation, line edge roughness). In addition to this [Trybula 1995] electronic design
automation has empowered industry with capabilities to bridge not only the increasing gap between design
and manufacturing but also the cost reduction with computer aided design tools (ECAD/TCAD). The DFM
efforts are beneficial if they exceed implementation costs like software license cost, process characterization
and model calibration resources, loss of design efficiency and time to market opportunity and loss of layout
density [Liebmann, 2008]. Variability has always been present in IC’s [Aitken et al., 2008], but the most
advanced data analysis tool used to manage variability has been (and still is) “the spreadsheet”. There is no
single DFM methodology integrated within EDA flow that can distribute residual variance; however authors
do agree that DFM has evolved from rule based to model based (physical to electrical DFM) and is defined
as broad set of practices that helps produce compelling products at high yield levels on a competitive
schedule.
It is evident from the above discussion that DFM is critical for the success of the SI but its
effectiveness depends on our ability to exploit the production data source by continuously adding new data
dimensions. In the SI business models engineers still use excel for a variety of analysis. It is an important
question that even in the presence of advances and sophisticated GUI tools for data extraction and analysis,
why R&D engineers are using excel as an intermediate tool. The most logical answer that we received from
the engineers in the production line is that newly emerging variations need multi-source analysis, hence they
use excel to align multi-source data which is captured at different levels (wafer/site/die). This methodology is
not working, because they are not able to find root causes against abnormal drifts or variations.
72
emerging with shrinking technologies. The overwhelming amount of manufacturing data is turning into a
potential challenge where knowledge extraction is highly temporal and success lies in timely extraction. A
potential solution requires focus not only on the data and its representation but also on the analysis
methodologies to achieve benefits. Industry has to shift from data towards information and/or knowledge
capitalization with descriptive, predictive and prescriptive models as proposed in this article. We believe that
the IDM has an ideal structure to shift from data driven DFM towards information/knowledge driven DFM
efforts in comparison with fabless to provide designers sufficient margins.
From the above discussions we can summarize that DFM is focused on manufacturability and yield
along with 4 improvement axis (Figure 2.17) as (i) technology (process, equipment and material
innovations), (ii) product design (design driven yield loss) and (iii) manufacturing (defect driven yield loss)
axes. The defect driven yield losses are controlled through APC/AEC (MFD) efforts characterized as
descriptive models whereas design driven yield losses require data mining (DFM) efforts as being
predictive/prescriptive models. The MFD efforts map to the horizontal axis (technology scale) whereas DFM
efforts are focused on design and manufacturing (vertical axes) and partially technology axis (process and
equipment engineering). The process, equipment and material axis constitute the technology scale whereas
design and manufacturability axes correspond to the product and variability respectively. The MFD efforts
when integrated with DFM, in a fully automated system, provide significant benefits not only in terms of
yield improvement and robust process control but also in pattern learning while moving towards process
improvements, techno shrink and future nodes. The technology scale directly impacts the design driven
defects and design in turn challenge our computational capabilities for potential manufacturability and yield.
To have an industry wide uniform understanding of the DFM concept, we can categorize it as (i) data driven,
(ii) information driven and (iii) knowledge driven efforts that correspond to descriptive, predictive and
prescriptive models.
We can simplify the discussion by defining our objective to move from the data driven DFM efforts
(as today) towards information and knowledge driven DFM initiatives. It requires dynamic exploitation of
the production data sources to find systematic patterns that can be transformed into rules and/or models. The
DFM effectiveness depends on the quality of the input data followed by information integration to generate
knowledge. This argument forms the basis for our 2nd and 3rd research question that (i) what are the true
DFM challenges, (ii) what are the limiting factors in the dynamic exploitation of the production data that has
led DFM ineffectiveness and (ii) what are the generic solutions to address those limitations? The answers to
these questions can be found in the chapters 4 and 5.
Yield
Figure 2.17 - 4-dimensional innovation framework for unified agile DFM system
73
2.3 INFORMATION INTEGRATION CHALLENGES TOWARDS MORE EFFECTIVE DFM METHODS
From the above discussion we have established a unified understanding of the DFM concept based on the
shift from data driven DFM efforts towards information and knowledge driven DFM initiatives. The unified
DFM concept is based on the [Zeigler et al., 2000] knowledge models as (i) descriptive, (ii) predictive and
(iii) prescriptive where transition among these models require a shift from data towards information and
knowledge. The data analysis framework where engineers transform data into information and knowledge is
best explained by the Data-Method-Stat (DMS) triangle (Figure 2.18). Data is generated from the methods
(manufacturing processes) and is stored at three levels, (i) Operational Data Sources (ODS), (ii) Data
Warehouse (DWH) and (iii) Data Marts (DM). The statistics includes data analysis methodologies that
generate information from data and further transform it into knowledge with machine learning algorithms.
The results from this analysis are used to control, align or fix the drifting process so that manufacturability
and yield can be ensured. The DMS triangle follows a cyclic improvement pattern to transform data into
information and knowledge.
74
2.3.1 Data/Information Integration Issues
The database technology to store data/information was coined around 1960 and since then engineers have
always been complaining about data volume and dimensions to perform accurate statistical analysis. Recent
revolutions in IT technologies have enabled huge data volumes and dimensions along with decreasing
storage and computational costs. It has raised the need for data integration and first integration approach was
introduced in 1980 as Multibase [Landers and Rosenberg, 1982 and Hurson and Bright, 1991]. The data
dimension here refers to a specific type of data and in technical terms a new table in the database.
The data collected across the production line is of heterogeneous nature; hence it requires integration
prior to multi-source analysis. The data integration is defined as unifying data that share common semantics
but originates from different sources. It is expected to address 4 types of heterogeneities: (i) structural
heterogeneity, involving different data models; (ii) syntactical heterogeneity, involving different languages
and data representations; (iii) systemic heterogeneity, involving hardware and operating systems and (iv)
semantics heterogeneity, involving different concepts and their interpretations [Gruber, 1993 and Busse et
al., 1999]. The new database technologies have solved the syntactic and systematic heterogeneity but
semantic and structural heterogeneities are still to be addressed. The semantic heterogeneity is mainly
focused in research with reference to data/information integration and it deals with three types of concepts:
(i) the semantically equivalent concepts, (ii) the semantically unrelated concepts, and (iii) the semantically
related concepts [Cui and O’Brien, 2000]. The Ontology and Metadata are two approaches used till date to
remove such type of heterogeneities for an efficient data and information integration.
75
information systems [Corcho and Gomez-Perez, 2000]. The federated information systems supported with
federated database provides a tightly coupled federation layer between information system and federated
databases, tightly coupled through schema matching (metadata) and single federated unified schema.
Ontology has to deal with the all type of information like structured (databases), semi-structured (XML) and
non-structured (web pages). Increasing web based non-structured and semi-structured information have
pushed all the concentration in ontology to focus on removing semantic heterogeneity for web resulting in
significant developments like Web Ontology Language (OWL), Description Logic (DL), SPARKLE
(ontology based language), Resource Description Framework (RDF), Document Type Definition (DTD)
etc…
76
ontology (global ontology for shared vocabulary for semantics e.g. SIMS [Arens et al., 1996]), (ii) multiple
ontology (multiple ontologies no common shared vocabulary e.g. OBSERVER [Mena et al., 1996]) and (iii)
hybrid ontology (multiple ontologies with shared vocabulary e.g. COIN [Goh, 1997]). It is also used as a
global query schema, e.g. SIMS [Arens et al., 1996] user query is expressed in terms of ontology which is
reformulated in respective sub-queries followed by results combination. The ontology acts as a global
schema for query; hence it requires an automated tool otherwise it would be difficult for the user to formulate
a query from this schema without knowing each and every detail of individual schema [Wache et al., 2001].
There are many languages for the representation of the ontology as: description logic (DL), terminological
(T-Box: vocabulary of the application domain), non-terminological (A-Box: assertions about vocabulary)
and Classical Frame based languages.
[Buccella and Cechich, 2003] proposed a three step approach for data integration where first of all
shared vocabulary is developed, followed by local ontology development and then associated mapping. This
approach is only focused on reducing semantic heterogeneity. The case study from an automobile industry
[Maier et al., 2003] demonstrates the benefits of using ontology for efficient and reliable data integration in
Product Life Cycle Management (PLM). It is a 6-step methodology: (i) schema import (table is taken as
concept), (ii) creating relations (logical relations between tables), (iii) create mappings (concept to concept,
attribute to attribute and attribute to concept), (iv) business logic (deductive logic), (v) rule based modeling,
(vi) inferencing and schema export. It is recommended as a beneficial methodology for heterogeneous data
integration.
[Buccella et al., 2005] compared 7 systems (SIMS, OBSERVER, DOME, KRAFT, Carnot,
InfoSleuth and COIN) for ontology based distributed and heterogeneous data integration including structured
and unstructured data sources (web). Framework used for evaluation purpose is architecture, semantic
heterogeneity (semantically equivalent concepts, semantically unrelated concepts and semantically related
concepts) and query resolution [Cui and Brien, 2000]. Many ontology based data-integration surveys can be
found, that focus on ontology development and mapping whereas other focus on languages used to represent
ontologies [Wache et al., 2001 and Corcho and Gomez-Perez, 2000].
The query resolution is an important but neglected factor within ontology based data integration
tools/systems, however optimization approaches could be paramount in overall ontology based system
efficiency for integration. The SI have three-level storage architecture hence ODS is expected to store the
measurement/observation time as transaction time whereas aggregated information is recorded through ETL
routines in DWH and DM’s. Risk of database servers being overloaded due to heavy queries from users
might result in temporal data mismatch resulting in an opportunity loss. It is necessary to find a solution for
an efficient data extraction and integration considering temporal nature of data in such a way that scheduled
ETL (extraction, transformation and loading) routines are least impacted. This 3-layer storage architecture
provides an opportunity to handle temporal issue but still require semantic and structural heterogeneity to be
addressed.
77
[Aparício et al., 2005] has demonstrated that ontologies help us in addressing semantic heterogeneity
better than building global schema, hence database integration if supported with building ontology (single or
multiple) followed by query answering using global schema proves its benefits. In this approach the global
schema is presented with database as SQL view to enhance the performance of system. [Fiedler et al., 2005]
present a very interesting debate and argues that the solution heterogeneous databases integration is difficult
but it can be achieved with assumptions; however they suggest that collaboration instead of integration is
better approach in such scenarios. The idea is very simple, that global schema provides possible opportunities
to merge the databases resulting in a global view for the global users. The coordination concept presented here
is similar to the correspondence table between databases and provide means to communicate, coordinate and
cooperate through SQL views. Such ideas can be tested with few small databases but serious problems shall
arise in case of huge temporal databases with semantic heterogeneity.
So far all above approaches talk about developing ontology and then using it to find similarities
between schema for query model but no approach talk about possible real time live databases where schema
or concepts could be added on as and when required basis; hence it results in the need for dynamic
resynchronization. In our case, the scenario is quite complex, overall domain is semiconductor manufacturing,
but diversity and heterogeneity in comparison to the evolution during last two decades, presents a challenge
for information integration. In order to keep-up pace with the technological evolutions it has become
important to push the concept of data/information integration and knowledge acquisition for competitive
advantage.
78
2.3.6 RDB Schema to Ontology Mapping Approaches
Due to increasing volume of data across web, relational database to ontology mapping has emerged as a
significant domain for knowledge sharing and exchange of data [Konstantinou et al., 2006 and Liddle et al.,
2003 and Calvanese et al., 2001]. [Zuling and Hongbing, 2005] proposed a methodology to map relational
databases (using MySQL) to ontology by mapping its schema and instances to T-BOX and A-BOX (using
“Protégé” software) and then user request is parsed into RDQL (RDF query language) which is transformed to
formulate SQL query for execution. There are certain questions which are still unanswered like in a situation
if the instances are in the size of terabytes: are we obliged to bring database instances within the developed
ontology domain where SPARKLE or any other web semantic query language can be used to answer user
requests? Perhaps it is not a good solution as far as query performance and storage space is concerned.
[Calvanese et al., 2001] proposed the notion of ontology of integration where we intend to develop a
global ontology from local ontologies for information extraction. [Zuling and Hongbing, 2005] proposed an
interesting idea to implement ontology database with user defined rule based querying and reasoning.
[Astrova et al., 2007] propose to store the web ontologies into relational databases in order to benefit from
computational efficiencies. [Curé, 2005] proposed DBOM (database to ontology mapping) framework for
semantic web data integration. [Nyulas et al., 2007] developed a plug-in to be used with “protégé” to import
database schema with first hand mapping to OWL and provides a strong feature where you can and cannot
import data along with the schema, in case if data is not imported it can be accessed through OWL-database
bridge.
2.3.7 Ontology Driven Data Extraction Tools
The ontology development and management environment (freeware and commercial) exploded as the interest
in web information integration increased. It has resulted in the emergence of technologies like RDF, XML
schema, DTD, OWL etc. Our focus is on data alignment, integration and extraction of temporal
manufacturing data; hence we shall focus on the integrated environments developed for data/information
extraction and integration based on 3-layer storage (ODS, DWH and DM). Two commercial and three
freeware ontology tools are selected out of 100+ available based on storage schema (database) as under for
possible integration and utilization of schema matching and ontology management [Youn and McLeod,
2006].
a) Link-Factory Workbench (commercial tool): It is a 3-tier architecture for large medical
ontologies, user interface is client-side application that is used to connect to the business tier
(implemented as ontology server) to manage or query data-tier i.e. relational database. Input output
formats supported are XML, RDF, OWL and multi user capabilities with information extraction
whereas ontology stored in the database is just used for linking purpose. It does not support graphical
output view.
b) K-Infinity (commercial tool): Knowledge builder utility with two major components, graph and
concept editors, facilitates the knowledge engineers to build objects, relations, network of knowledge,
concepts, individual attributes and relations.
c) Protégé 2000 (freeware): It is developed by Stanford University, USA and it is a freeware having
a lot of functionalities. It is a graphical tool for domain ontology development and management
supported with plug-ins to graphically view the tables and diagrams. Further it can be used to learn
ontology from relational database schema with and without importing instances. Query requests
require an understanding of RDF and SPARKLE. The I/O format is RDF, XML Schema and Java
[Arens et al., 1996, Mena et al., 1996 and Ramakrishnan and Gehrke, 2003].
d) WebODE (freeware): It is one of the powerful freeware developed to serve three purposes (i)
ontology development, (ii) ontology integration (iii) ontology based application development, all
three purposes are met through graphical user interface provided with its 3-tier ontology editor
79
WebODE. It is developed in JAVA; hence support CORBA, XML and RMI. It provides GUI, web
and multi user support including prominent web formats.
e) ICOM (freeware): It is a powerful tool in terms of capabilities of conceptual Entity Relation (ER)
modeling with an objective to ensure the visualization of all constraints as one single conceptual
model. It could also be used for the translation of ER conceptual models developed in ontologies for
various purposes. ICOM tool reasons with (multiple) diagrams by encoding them in a single
description logic knowledge base, and shows the result of any deductions such as inferred links, new
constraints, and inconsistent entities or relationships. Theoretical results guarantee the correctness and
the completeness of the reasoning process.
One thing is apparent that ontology has evolved with a focus on sharing and exchanging web
knowledge. There is no approach proposed so far that takes into account the model evolution and its dynamic
synchronizations. In SI, the databases are growing at an immense pace in data volume and dimension; hence
we are obliged to continuously restructure the data models so that new dimensions can be timely updated. Our
success lies in our ability to dynamically exploit all data volumes and dimensions to find root causes against
the abnormal drifts emerging from spatial variations. So we need an ontology based system that can pre-
assess the failures of potential changes in the data models. These assessments must be made at the application
and user levels. Any potential changes must follow the compliance loop (updation) where end user
application is affected by this change and end users must be informed for all structural changes in data
dimensions. This argument complement 2nd research question that how we can enable dynamic restructuring
of data models to continuously support the addition and deletion of data dimensions? The answer to this
question can be found in chapter-6.
2.4 SUMMARY AND CONCLUSIONS
In this chapter the literature has been analyzed across three dimensions (i) semiconductor industry background
and its evolution, (ii) role of DFM methods and current challenges and (iii) information integration
approaches for dynamic data exploitation. It is observed that the SI is characterized by the cyclic demand
patterns with positive CAGR (+8.72%) that guarantees a cumulative demand; hence we are obliged to
respond to the growing market with new, fast and high value but low cost products. It has led a shift in
business objective (T2M, T2V to ramp-up rate) and emergence of new technical challenges (design and
manufacturing interface complexities). The SI can only respond market growth by introducing new
technologies every 2 to 3 years or quickly maturing the technology derivative and improvement efforts with
local DFM efforts. The DFM methods had been adopted by the SI around 1980 as a yield enhancement
strategy which worked well till 250nm technology node, but beyond this node DFM has resulted in high cost
and ineffective R&D efforts, creating a challenge in accomplishing quick ramp-up rates with the existing SI
business models.
In order to address the increased technology lead times and R&D costs, the SI model (IDM) has
structurally transformed into fablite and fabless business models. But still, IDM fablite model is reported to
be the best models in terms of revenue generation. We argue that the IDM fablite model provides a coherent
platform for the knowledge capitalization from production line and our ability to improve the challenges
faced by the R&D engineers shall result in the continuous improvement efforts for technology alignment or
adoption. It shall reduce technology lead times, associated costs along with an opportunity for early
penetration in the market with higher profits.
The SI is a 300+ B$ industry with an equal opportunity to capture maximum market for its stake
holders. In order to get maximum share, it is necessary to revisit and rank the SI objectives so that a strategy
can be formulated truly in line with the top ranked objective. The existing business models (IDM, Fablite and
Fabless) must be analyzed to evaluate their potential to support the newly formulated business strategy. In
case if none of the existing model is capable to support it, a new model or an extension to the existing
80
business models must be proposed to ensure its compliance with the new strategy in line with the top
ranked business objectives. This argument formulates our first research question, which is addressed in
chapter-3.
The SI emerged as one of the most complex, competitive and technologically fastest growing
manufacturing domain. From the literature review on the evolution of the DFM methods, it is evident that
success lies in the ability to put DFM back to track. Few significant factors identified being major causes for
DFM ineffectiveness (i) inability to exploit the production data for multi-source root cause analysis, (ii)
improper analysis due to difficulties in data alignment because of the varying metrology coordinates and (iii)
heterogeneous understanding of the DFM concept. In this chapter, a unified understanding of the DFM
concept as the data driven R&D effort is proposed with an objective to shift these data centered efforts
towards information and knowledge centered R&D initiatives. It requires dynamic exploitation of the
production data sources to find systematic patterns that can be transformed into rules and/or models. The
DFM effectiveness depends on the quality of input data followed by information integration and analysis to
generate knowledge. This argument forms the basis for 2nd and 3rd research questions as presented in section
1.4. They can be further divided into smaller questions as (i) what are the true DFM challenges, (ii) what are
the limiting factors in exploitation of the production data sources that has led the DFM ineffectiveness and
(ii) what are the generic solutions to address those limitations? The answers to these questions can be found
in the chapters 4 and 5.
The data analysis framework where engineers transform data collected across the production line
into information and knowledge is best explained by the data-method-stat (DMS) triangle. The data is
generated from methods (processes) and is stored at three levels, (i) operational data sources (ODS), (ii) data
warehouse (DWH) and (iii) data marts (DM). The Statistics includes analysis methodologies that generate
information from the data and further transform it into knowledge with advanced statistical and/or machine
learning algorithms. The results from these analyses are used to control, align or fix the drifting processes so
that manufacturability and yield can be ensured. The DMS triangle follows an improvement cycle to
transform the data into information and knowledge. The online transaction (OLTP) and analytical processing
(OLAP) are key concepts which are built around the methodologies to store and process the data to generate
information and knowledge. The enterprise wide information is stored at three levels ODS, DWH and DM.
The OLTP system is based on the relational databases where end-user queries are optimized for transactional
processing (insert, delete, and update). In Comparison to the OLTP, OLAP systems are based on the DWH
and DM concepts and are focused on the query performance and quick transformation of data into
information and knowledge for decision making. The data is stored in a multidimensional array in the OLAP
systems which results in fast slicing, dicing and drill up/down operations. The primary objective of the
OLAP system is on fast data aggregation. In the SI, R&D engineers are primarily focused on multi-source
analysis not for the purpose of aggregation but for mapping and alignment so that effective root cause
analyses can be performed to correct the drifts processes.
The above presented OLAP and OLTP systems do not match with the objectives of R&D engineers;
however both systems offer interesting benefits which can be used to improve the R&D productivity. The
OLAP system and multidimensional modeling can be used to quickly access multi-source data from multiple
databases accumulate in huge volumes for the purpose of mapping and alignment instead of aggregation. The
slicing, dicing and drill up/down are of very high interest which can help engineers to find root causes across
multiple dimensions of the data sources. The OLTP systems offer transactional efficiency; hence we can use
this quality to manage data model evolutions. The difficult part of the SI is that ODS and DWH data sources
are of proprietary nature; hence they cannot be changed easily. It has resulted in serious issues about the
inclusion of new data dimensions and huge data volumes. The consequence of this fact is that R&D
engineers are not able to exploit the production data sources which indirectly contribute to the DFM
ineffectiveness.
81
The data collected across the production line is of heterogeneous nature; hence it requires integration
prior to multi-source analysis. The data integration is defined as unifying data that share common semantics
but originates from the different sources. The SI being high-tech with potentially growing volumes of
temporal data and urgent need for data alignment, extraction and integration require a solution that includes
remedy for the semantic and structural heterogeneities with model evolution management. The ontology and
Metadata based data, information and database integration approaches are used most commonly for the
sharing and exchange of web data. These approaches neither support the huge data volumes nor the dynamic
restructuring of data models.
In SI, the databases are growing at an immense pace in data volume and dimensions; hence success
lies in the ability to dynamically exploit huge data volumes and dimensions for an effective root cause
analysis against the abnormal drifts emerging from the spatial variations. So we need an ontology based
system that can pre-assess the failures of potential changes in the data models. These assessments must be
made at the application and user levels. Any potential changes must follow the compliance loop (updation)
where end user application is affected by this change and end users must be informed for all structural
changes across all data dimensions. This argument helps us in complementing 2 nd research question that how
we can enable dynamic restructuring of data models to continuously support the addition and deletion of new
data dimensions? The answer to this question can be found in chapter-6.
82
Chapter 3: An Extended IDM (e-IDM) Business Model 8
In this chapter, SI business models are analyzed for their support to a recent shift in the business objectives
towards quick ramp-up-rate. The Strategic Creative ANalysis (SCAN) and Strength, Weakness, Opportunity
and Threat (SWOT) analyses techniques are used in addition to the brain storming sessions to (i) identify key
improvement areas in today’s most successful IDM-fablite business model and (ii) formulate business
strategy to achieve maximum economic benefits. The technology development/improvement process is
further analyzed to find the limitations that have led to the DFM ineffectiveness and extended ramp-up rates.
The challenges found here along with the key improvement areas form the basis for the proposition of an
extended IDM business (e-IDM) model with integrated DFM methodology.
Contents
3.1 Introduction ......................................................................................................................... 85
3.1.1 Strategic Planning Analysis .................................................................................. 85
3.1.2 SCAN Analysis ...................................................................................................... 86
3.2 SCAN Analysis: Part-1 (Top Ranked Business Objectives in SI) ....................................... 87
3.3 Key Improvement Areas in IDM-fablite Business Model.................................................... 88
3.4 Technology Derivative/Improvement Process Analysis. ..................................................... 90
3.5 Key Challenges in Technology Derivative/Improvement Process ....................................... 91
3.5.1 Data Extraction Issues .......................................................................................... 91
3.5.2 Variance Analysis Challenges ............................................................................... 92
3.5.3 Silicon Based Correlation Limitations .................................................................. 92
3.6 SWOT Analysis on IDM-fablite Business Model ................................................................ 92
3.7 Proposed Extended IDM-fablite (e-IDM) Business Model .................................................. 94
3.8 Research Schematic and Advancement (e-IDM Model) ...................................................... 95
3.9 Summary and Conclusions .................................................................................................. 97
8 SHAHZAD M.K., HUBAC S., SIADAT A., TOLLENAERE M., An Extended Business Model to Ensure Time-to-Quality in Semiconductor
Manufacturing Industry, International Conference on Enterprise Information Systems, 2011, Portugal
83
84
3.1 INTRODUCTION
The traditional SI business model (IDM) has structurally transformed into IDM-fablite, fabless and foundry
business models in order to share the exponentially increasing technology R&D costs and lead times. It is
very important to understand that to cope with the cumulative demand growth (+8.78%), new IC
technologies must be introduced every 2-3 years. The design and manufacturing interface complexities (new
manufacturability and yield loss mechanisms) due to miniaturization are rising and being compensated with
the design, process, material and equipment innovations. It requires more human resource and
experimentations that give rise to technology development costs to ensure its timely introduction in the
market. So it is evident from the above facts that structural transformation was just to share the increasing
R&D costs. In the current situation an IDM-fablite is classified as the best model for associated economic
opportunities.
The recent shift in the SI business objectives from T2M and T2V towards the ramp-up rate needs to
be reassessed in the context of potential opportunities associated with cumulative demand growth (+8.78
CAGR). So in simple words, one needs to reconsider that the best IDM-fablite model is capable to sustain
this growth in demand for maximum economic benefits. This assessment has been divided into three parts as
(i) identify business strategy in compliance with increasing demands and economic opportunities, (ii) assess
existing business model for possible compliance with the quick ramp-up objective and (iii) find key
improvement areas and/or propose a business model.
a) SWOT analysis: It is most widely used approach in which the Strengths, Weaknesses,
Opportunities and Threats associated with the business activity are identified. The first step is to
define business objective of the activity and find internal and external factors important to achieve
the objectives. The strengths and weaknesses are usually internal whereas opportunities and threats
are always external. It is a generic tool and has a wide application for the formulation of a business
strategy to achieve the objectives [Mindtools, 2007].
b) PEST analysis: It is a tool used to understand the Political, Economic, Socio-cultural and
Technological environment of the organization and is commonly used to analyze the market growth,
decline or such factors to align the future business directions. The four PEST factors can be treated
like opportunities and a threat as in SWOT analysis but its emphasis is more on the socio-economic
factors. Its application is highly dependent on the type of business and or activity being analyzed
[Businessballs, 2006].
c) Porter’s five forces: This concept is based on the five forces and is used to find out the power
center in the current business situation. These five forces are (i) suppliers power, (ii) buyer power,
(iii) competitive rivalry, (iv) threat of substitution and (v) threat of new entry. By understanding
power center, this concept can also be used to find the areas of strength, to improve weaknesses and
to avoid mistakes [Porter, 1996 and Porter, 1998].
85
d) Four corners analysis: It is a predictive tool designed by Michael E. Porter that helps in
determining a competitor’s course of action and is based on the basic principle of knowing what
motivates the competitor. This dimension helps in finding a more accurate and realistic assessment
of the competitors possible moves [Gilad, 2011]. The four corners are (i) motivation (drivers), (ii)
motivation (management), (iii) actions (strategy) and (iv) actions (capabilities).
e) Value chain analysis: It is based on the concept that all the activities in the organization create
and add a value. This methodology is used to identify those activities which do not add value; hence
they can either be removed or replaced with more efficient activities. Each value adding activity is
considered a source of competitive advantage [Porter, 1996].
f) Early warning system: The purpose of strategic early warning system is to detect or predict
strategically important events as early as possible. They are often used to identify the first strategic
move from competitor or to assess the likelihood of a given scenario becoming reality. The seven
components of this system are market definition, open systems, filtering, predictive intelligence,
communicating intelligence, contingency planning and the cyclic process [Comai and Tena, 2007].
g) War gaming: This technique is used to find competitive vulnerabilities and wrong internal
assumptions about competitors’ strategies. They are used for critical strategic decisions but depend
on the correct assessment of the competitors moves [Treat et al., 1996].
h) SCAN: It is used for the formulation of business strategies by first identifying and listing the
business objectives. These business objectives are ranked and then the top ranked business objective
is selected for further investigation using SWOT analysis that result in potential set of strategies. The
most appropriate strategy is selected and is always based on the strengths, weaknesses, opportunities
and threats [Winer, 1993].
If above listed commonly used tools are analyzed for the formulation of a business strategy, it can be
determined that early warning system, war gaming, four corner analysis and peters’ five forces system are
specifically designed for marketing environment in which one is obliged to assess the moves of his
competitors to take an appropriate action. Although they help in the formulation of strategic moves, they do
not directly comply with our generic needs. The value chain analysis is an excellent tool used to identify
inefficient activities so that they can be improved afterwards. But in this case, it is quite difficult to quantify
the value against each activity; hence, value chain does not seem appropriate either. The PEST analysis
emphasize on socio-economic elements for the business strategy which are not relevant in our case. The
SCAN and SWOT analysis being generic for the business strategy formulation are applicable in our case.
The SCAN analysis has an advantage over SWOT as it starts by ranking the objectives followed by SWOT
analysis to formulate a business strategy in line with the top ranked objective.
87
Get leadership position for ST in SI
How ?
How else? Why ?
How ? How ?
Why ?
Achieve Customer
Introduce New Technologies Im prove Sales Satisfaction
every 2 to 3 Years Revenues
How ?
How ? How ?
Why ? How else?
Why ? Why ?
Get Maxim um Market
Share Reduce Im prove Reduce Tim e to m arket Im prove quality
Develop leading Reduce R&D product cost yield and Tim e to volum es and reliability
Technology Cost on New How ?
Alliances and Technology How ?
Partnerships Development Why ?
Quick Ram p-Up Why ?
Rates
Reduce Technology
How ? Adoption Lead Tim es and
Costs
Why ? How ?
Reduce Technology
Alignm ent Lead Tim es and Why ?
Costs (derivative and Im prove MFD
im provement initiatives) (APC/AEC)
How ? effectiveness
How ?
Why ?
Im prove R&D
Efforts Efficiencies
In this analysis, the listed objectives are presented in rectangular boxes and they can be read either
from bottom to top by asking question “why is this objective being pursued?” or from top to bottom by
asking question “how can this objective be achieved?”. Answer to these questions can be found in terms of
another objective on the end of the link. The starting question that was posed was why R&D efficiencies
should be improved, and the answer is that they should be improved so that the DFM methods can be
improved. The next question is: why DFM methods should be improved? The answer is that, it should be
improved because the quick silicon based device and interconnect models validation are desirable. This
process continues until the top ranked objective << to get a leadership position and quick ramp-up rate >>
are reached. Now, looking at the example from top to bottom flow, a question is posed that how a leadership
position for STMicroelectronics can be achieved, and the answer is that by introducing new technologies
every 2-3 years, maximizing sales revenues and total customer satisfaction. The “how else” part supplements
the answer by providing evidence for the additional objectives that can contribute towards the target
objective.
88
MFD Efforts/Process Control Technology Alliance
Effective Root
Technology
Cause
Design Platform
Design Platform
Analysis ? Internal Tech Platform(s) Transfer ?
Design Platform FE/BE Technology
Process/
Product/
Design
Manufacturing
database
issues ?
FE/BE Technology
FE/BE Technology
(ITP > CTP)
89
a) Do methods exist (Fig.3.2) to improve device and interconnect models (technology alignment)
and product development process (technology adoption)?
b) Can manufacturing databases and EIS support continuous improvement in the ITP (Fig.3.2)?
3.4 TECHNOLOGY DERIVATIVE/IMPROVEMENT PROCESS ANALYSIS
In the context of above defined questions in section 3.3, technology derivative or improvement process is
analyzed. The answers are further used during SWOT analysis for the formulation of a business strategy. The
Integrated DEFinition (IDEF0) model for the said analysis is presented in Figure 3.3. In this process the
method, business functions and sub-functions have been represented with different actors involved. The
interactions between actors and functions are presented using arrows whereas interaction between functions
and sub-functions are shown by <<uses>> and <<extends>> arrows. The <<uses>> type arrow defines the
relationship that the function always uses the sub-function prior to its interaction with the actors defined in
the IDEF0 model whereas the <<extends>> type refers to potentially extending the functionality of the
source function/sub-function to the target function/sub-function.
The Process Integration (PI), technology R&D, device engineering, Central CAD Design Solution
(CCDS), DFM, Design, EDA and interconnect modeling teams are directly involved in the process. It can be
seen that the need for FE technology development/improvement could arise as per company policy to move
towards technology shrinks (55nm, 50nm, 40nm) or customer feedback for technology improvements. The
technology shrink a.k.a. technology derivative is a process where new technology is derived from the
existing technology and is capable of manufacturing smaller features (40nm is derived from 45nm, 55nm and
50nm are derived from 65nm). These derived technologies are based on design, process and/or material
innovations and result in the gain of area and costs. On the other hand, the technology improvement refers to
the efforts fully dependent on process innovations that enable to manufacture same design in smaller area.
These improvements enable to put features more closely but feature sizes are not reduced.
Manage Product
Device Modeling
Device Modeling Portfolio Strategy ST Product Groups
EDA Team
«uses»
«extends»
«extends»
«uses»
«uses» DfM Kit
DfM Kit Updated
Updated
Revised
Revised DRMDevice &
for New
T2D / APLES
Interconnect
Technology / Imrpovement Design Team
Models
«extends» «extends»
Interconnect
Interconnect
Modeling
Modeling «uses»
PI Team
«uses» Yield Modeling DFM Team
«uses» PT
PTVariation
Variation
Analysis
Analysis
«uses»
«uses»
«uses»
90
The device and interconnect modeling are the key processes in the technology derivative/
improvement initiatives where SPICE models (device and interconnect models) are developed through
CAD simulations and validated based on the measured electrical (PT) results. This process starts with the
definition of target values on critical and key parameters. The physical parametric stack is constructed based
on the source technology design rule manual (DRM) and is simulated until the electrical and parasitic
extraction satisfies the SPICE model corners. This stack is revised to validate until the electrical and parasitic
parameters comply with SPICE models. The geometric stack is identified through CAD simulations which
are validated through test products.
The focus here is on interconnect modeling where key sub-processes are listed as (i) data extraction,
(ii) drift analysis and (iii) silicon based correlation. The test products undergo the manufacturing process
flow and the data is collected across the production line. The first step in silicon based validation of the
interconnect models is the extraction of the data from multiple data sources for analysis purposes. These
measured parameters are benchmarked against the simulated results and significant deviations are
highlighted (single-source analysis). Any significant deviations are further investigated for the root cause
analysis (multi-source analysis) to find the source cause followed by model corrections. In this process,
success lies in the ability to quickly resolve the conflicts that arise from the stack validated vs. simulated
results that shall reduce the respins and technology lead times. These results are used to revise the device and
interconnect models in the DRM.
The DFM teams are focused on identifying the manufacturability and yield loss mechanisms where
the drifts/variations leading to parametric or functional yield loss are further investigated for their
classification as systematic or random mechanisms. These drifts and variations are transformed into rules
and/or models for inclusion in the DFM kits and subsequent use during CAD simulations. The design and
DFM kits are used by the designers in addition to the design libraries to quickly design, simulate and assess
its manufacturability and yield as a first time correct design. The similar process is required in the device
modeling; hence, they result in a chain reaction of improvements ultimately leading to a new technology
derivative/improvement effort (alignment or adoption) with reduced lead times and costs.
9 The complete UML model, GUI, Algorithms and experimental results can be found in Appendix-E
91
a) Multiple manufacturing data sources (relational databases) are dedicated to operational
excellence and do not support DFM/MFD efforts; hence, engineers spend a lot of time in
extraction, cleaning and alignment before analysis and in most cases, it results in zero value
addition.
b) Manufacturing data resources have serious ontology issues (same parameter with different
semantics in different databases), as a consequence it becomes difficult to align and correlate
multi-source data resulting in a missed opportunity.
c) Unstructured evolution of local databases has resulted missing links which are the key to
perform a multivariate or predictive modeling.
It is evident from these observations that multi-source data extraction has serious issues which
restrict engineers to find root causes using single source analysis resulting in extended lead times and in most
cases no improvements.
3.5.2 Variance Analysis Challenges
The variance analysis is the first step towards root cause analysis and provides initial signals about potential
drifts or variations. It is based on the electrical parametric data collected across the production line. The key
observations made during this project are listed as under:
a) R&D engineers may apply ± σ filters during single-source data extraction to remove outliers that
could possibly highlight a significant drift. The resulting data when checked against its
compliance with the normality law, is often found in compliance.
b) Excel is the widely used tool in SI besides multiple advanced statistical tools but engineers
prefer excel because they could easily handle and manipulate data in excel that might result in
misleading conclusions
c) The data collected across the production line is huge and the tools being used for the data pre-
processing are too slow to handle it because of their relational database limitations.
It is evident from these observations that when different analysis tools are used, including excel, they
are like black boxes in terms of the algorithm used for the computation of statistics. It was found that the
same data when used for certain statistics computation with different tools, often result in slight variation that
might be misleading. If the data is filtered then it is not possible to improve or go for a technology derivative.
3.5.3 Silicon Based Correlation Limitations
This sub-process is not yet implemented because R&D engineers are not able to align the extracted data at
site/die and test structure position levels. It means that it is not possible to perform multi-source analysis to
find root causes against drifts/variations. It can simply be concluded that the knowledge capitalization is at
its minimum in this sub-process. Now there exists sufficient information about the strengths and weaknesses
of an IDM-fablite business model to move towards SWOT analysis so that an appropriate business strategy
can be agreed prior to propose modification in the business model to ensure its compliance with a recent shift
in the business objective i.e. ramp-up rate.
92
a) Strength/Opportunity Option: This option suggests joint ventures with top ranked IDMs,
foundries and fabless companies to best exploit the strengths e.g. intellectual capital, state of the art
equipment, data and methods against potential opportunities (high revenues and market share).
INTERNAL
TRO TRO
Strength Weaknesses
Technology • Intellectual Capital in R&D • Ontology Issues
T2M
T2Q
Alignment • State of the art equipment • Missing links between
+ Adoption T2V Databases
• Quality standards & procedures
TRO TRO • Huge manufacturing Data • Excel (widely used) + Data
• Product/test chip Filtering
• Interconnect & Device
SWOT Analysis for TRO characterization, FDC, APC,
modelling based on previous
Device and Interconnect modeling
“Achieve leadership position in SI” methodologies DRM
• 11+ EDA tools & 6+ databases • Min knowledge capitalization
b) Strength/Threat Option: This option suggest focus on the design, process, equipment and
material innovation to mitigate threats like limiting physics laws, technology platform development
and backward compatibility and dynamic customer requirements.
c) Weakness/Opportunity Option: The ontology issues, missing database links, usage of excel for
data analysis, minimum knowledge capitalization (correlation between geometric and electrical
measurements) are characterized as the core weaknesses; hence, in order to exploit the opportunities
there must be a focus on the knowledge capitalization and improved coordination between R&D
functions.
93
3.7 PROPOSED EXTENDED IDM (E-IDM) BUSINESS MODEL
It can be concluded that multi-source data extraction, alignment and mapping is critical to improve the
effectiveness of DFM and MFD efforts during new technology transfer (alignment), technology derivative
and improvement efforts (alignment) and product design and development (adoption). Based on these facts
and business strategies (c & d) it can also be concluded that existing IDM-fablite mode is not coherent with
the recent shift in the business objective (ramp-up rate). So an extended IDM-fablite business model has
been proposed which provides true coherence with recent shift in business objectives (ramp-up rate) and top
ranked SI objective (Figure 3.5).
Design Platform
Design Platform
Root Cause Analysis Internal Tech Platform(s) Transfer.
(DFM + MFD efforts) Design Platform FE/BE Technology
Process/
Product/
Design
No manufacturing
database issues.
FE/BE Technology
FE/BE Technology
(ITP > CTP)
94
example of inward manufacturing window is recently witnessed when the revenues of STMicroelectronics
dropped by 6% whereas the worldwide sales revenues have increased. The reason was that Nokia fired 5000
employs and it had a significant impact on the sales orders for STMicroelectronics.
95
Figure 3.6 - The research schematic and advancement with e-IDM business model
96
3.9 SUMMARY AND CONCLUSIONS
This chapter has analyzed the most successful IDM-fablite business model from the SI to assess its support
towards a recent shift in the business objectives “to achieve leadership position and quick ramp-up rate”. The
first step was the identification of business (T2M, T2V, ramp-up-rate) and technical (area, power, timing,
leakage) objectives using brainstorming sessions held with engineers and managers, followed by their
subsequent ranking using SCAN analysis. The top ranked business objective “to achieve leadership position
and quick ramp-up rate” and key improvement areas as the (i) fast technology transfer, (ii) manufacturing
databases and (iii) effective root cause analysis (R&D) from IDM-fablite model, are further investigated
using SWOT analysis. The objective is to align business strategy to assess its potential compliance with
IDM-fablite business model against identified TROs. The technology development/improvement process
was further investigated to assess potential challenges faced during its compliance with the TROs and (i) data
extraction, (ii) alignment and (iii) pre-processing due to ontology issues and (iv) missing database links are
found as key failure modes.
The conclusions highlighted that the existing IDM-fablite model do not fully support the TRO
objective. It is because of the fact that our R&D engineers are not able to exploit multi-source data collected
across production line for the root causes analysis which has led ineffectiveness in the DFM methods. The
DFM methods play a critical role in technology development, alignment and/or adoption. It is important to
improve technology alignment and adoption lead times and associated costs; hence an extended IDM-fablite
(e-IDM) business model is proposed with the integrated DFM efforts. The proposed model provides
compliance with TROs and enhances inward manufacturing window while maintaining its backward
compatibility with the technology developed in alliance.
In next chapter we shall find out true DFM challenges within the proposed e-IDM model so that
associated failure modes and root causes against DFM integration are removed. The objective is to ensure the
success of proposed e-IDM model so that recent shift in business objectives quick-ramp-up rate is achieved
for potential economic benefits and leadership position.
97
98
Chapter 4: I-FMEA Methodology for True DFM Challenges10
In the previous chapter, an extended IDM-fablite (e-IDM) model is presented with DFM integration to continuously
improve the technology derivative/improvement alignment and adoption. The propose i-FMEA methodology helps
in identifying and removing true DFM challenges, which are critical for the success of e-IDM model. The i-FMEA is
different than traditional FMEA as it searches failure modes and root causes across business functions. It is applied
on two groups of failure modes as (i) ineffective root cause analysis (infield and scribe line test structure positions)
and (ii) data extraction, mapping and alignment. It has been seen that most of the cyclic root causes (repeating
causes) are traced back to IT and EDA business functions. The identified root causes form the basis for generic
scientific contributions as (i) MAM, (ii) SPM, (iii) ROMMII and (iv) YASS. The experience learned during the
identification of true DFM challenges has led the proposition of 4-step i-FMEA methodology which is capable of
finding cyclic failure modes and root causes, which require generic solutions rather than operational fixes.
Contents
4.1 Introduction.........................................................................................................................101
4.2 Historical Evolution of FMEA Methodology ....................................................................101
4.2.1 FMEA Process and Evolution ..............................................................................101
4.2.2 Basic FMEA Vocabulary ......................................................................................102
4.2.3 Benefits and Limitations of Traditional Approach ...............................................102
4.3 Proposed Interdisciplinary (i-FMEA) Methodology ...........................................................103
4.3.1 Comparison of i-FMEA with Traditional FMEA Approach .................................103
4.3.2 i-FMEA Methodology and Thesis Schematic ........................................................104
4.4 i-FMEA Methodology Results ...........................................................................................114
4.4.1 Step-2: Initial Failure Modes and Root Causes ...................................................114
4.4.2 Operational Fixes through Joint Projects ............................................................111
4.4.3 Step-3: Cyclic Failure Modes and Root Causes ...................................................113
4.4.4 Generic R&D Solutions ........................................................................................115
4.5 Research Schematic and Advancement (i-FMEA Methodology) ......................................116
4.6 Summary and Conclusions .................................................................................................118
10 Shahzad M.K., Hubac S., Siadat A. and Tollenaere M., An interdisciplinary FMEA methodology to find true DFM
challenges, 12th European APCM Conference, Grenoble France 2012
99
100
4.1 INTRODUCTION
The previous chapter has proposed an extended IDM-fablite (e-IDM) business model to address the recent
shift in business objectives “ramp-up rate” in compliance with top ranked business objective “to achieve the
leadership position”. The success of the proposed model depends on the ability to remove DFM
ineffectiveness across three key improvement areas: (i) fast technology transfer, (ii) ineffective root cause
analysis and (iii) extraction, mapping and alignment to improve technology alignment and adoption efforts.
In this chapter, the objective is to identify true failure modes and root causes for the DFM ineffectiveness for
generic solutions rather than operational fixes. The proposed i-FMEA methodology is based on well known
approach used for the concept, design and process known as Failure Mode Effect Analysis (FMEA). This
approach is limited by the expert’s knowledge and scope because failure modes and root causes are searched
at system/subsystem/ component levels. The root causes associated with the failure modes are removed with
operational fixes, which are not permanent solution. The proposed 4-step interdisciplinary FMEA (i-FMEA)
methodology is superior to the traditional FMEA, because it searches for cyclic failure modes and root
causes across all the business functions. It is objectively focused on the continuous scanning of business
environment for any potential change followed by the identification of key challenges and failure modes with
cyclic causes. The results are very promising as the key failure modes and root causes identified were never
thought to be the source of the problem.
101
Find Failure Find Potential Controls for
Modes & Effects Root Causes Detection
2
Initial Failure
Model Analysis
1
Scope / RPN
3
Revue Priority
5
Operational Fixes
4
102
being designed or redesigned, (ii) when an existing product, process or service is being applied in different
way, (iii) when developing control plans for new or modified process, (iv) when improving the existing
products, processes or services, (v) when analyzing failures of an existing process and (vi) periodically
throughout life of process, product or service.
Track
Environment
for Changes
SCAN
Analysis Key
Challenges
103
b) The success of traditional FMEA depends on our ability to precisely define the scope and functional
analysis before applying the basic steps. It requires expert’s knowledge and could easily lead to the
waste of time and resources; however, i-FMEA methodology continuously analyze the business
environment for any potential change followed by SCAN analysis to find out critical processes that
might get effected. This methodology is a better approach as it emphasizes to align all business
functions to quickly respond to any potential change in the business environment.
c) The traditional FMEA approach is subjectively focused on applying operational fixes to improve the
design, process or system imperfection but i-FMEA methodology is focused on identifying failure
modes with cyclic causes followed by generic R&D fixes so that any chances for associated causes
to become causes for other failure modes can be eliminated.
The above discussed potential differences clearly defines the advantages of our proposed i-FMEA
methodology over traditional FMEA approach, which is primarily/subjectively focused on aligning the
business processes to quickly respond to business environmental changes for maximum market share and
higher profits.
104
Figure 4.3 – 4-Step i-FMEA methodology and research thesis schematic
105
4.4.1.1 Technology Derivative/Improvement Initiative
This process comprises of four sub-processes: (i) device/interconnect modeling, (ii) data extraction and pre-
processing, (iii) variance analysis and (iv) PT-Inline correlation. The traditional FMEA results are presented
in tables 4.1 and 4.2. Let us discuss the FMEA results of device/interconnect modeling (Table 4.1) where
sub-processes are benchmarked as (i) to generate typical stack, (ii) to generate stack for corner models, (iii)
QA and impact analysis, (iv) design impact analysis, (v) variance analysis and (vi) model validation. The
cutoff value for RPN based prioritization is taken as 200. It is known that (i) variance analysis and (ii) model
validation are the critical sub-processes; hence, the discussion will be restricted to these two critical sub-
processes. The simulated target and corner models against typical and corner stack definitions are validated
where significant deviations are further investigated to find root causes so that the process, equipment or
SPICE models can be adjusted. The variance analysis and model validation steps are then the most important
tasks that have direct impact on the technology lead times and associate costs.
The most important and critical root causes identified are incorrect tool, incorrect statistics, incorrect
measurements, wrong data pre-processing and site level data alignment for the electrical and physical
geometric measurements. The fact that the extracted data is not understandable by the R&D engineers makes
it almost impossible to analyze the variances and/or model validations. If the preventions and proposed
actions are closely analyzed, it can be easily concluded that the biggest challenge is data extraction in right
format.
The critical causes highlighted in device/interconnect modeling are as (i) data extraction, (ii)
alignment and (iii) pre-processing, which are further investigated as presented in the table 4.2 to ascertain the
proposed corrections. It can be concluded that these sub-processes are very critical for device/interconnect
model variance analysis in technology derivative/improvement initiatives based on the fact that computed
RPN values are above the cutoff value (200). These critical elements are further discussed as under:
a) Data Extraction Sub-Process
The SI production line data is stored in 4-layer data storage architecture: (i) flat files, (ii) operational data
sources or stores (ODS), (iii) data warehouses (DWH) and (iv) data marts (DM). The higher storage layers
(DWH and DM) represent aggregation and integration of the data from lower layers (flat files and ODS). The
industry normally uses multiple data extraction tools which are classified as single-source or multi-source
data extraction and analysis utilities. More often it is found that either data is not available or users are not
able to properly extract it using these utilities. In both cases neither variance analysis nor model validation
could be performed, which result in high severity. The most apparent root causes identified against these
limitations are: (i) the databases have different data retention periods and (ii) users are not able to use
multiple tools due to training issues. The users use multiple tools for data extraction from different data
sources that result in data incoherencies and pose severe data alignment issues because ODS have longer data
retentions than DWH and DM. It is often seen that, multi-source data extracted by end users has varying
formats and is not understandable due to different vocabulary used in different databases. The databases are
often not supported with up-to-date data dictionaries to present semantics of data and dimensions. All of
these limitations result in data alignment; hence, the single-source variance analysis is likely possible but
multi-source analysis due to data alignment issues is almost impossible. The proposed solutions could be to
either change DWH/DM databases or develop automated data extraction tools.
106
Table 4.1 – Device / Interconnect modeling FMEA result
107
a) Data Alignment Sub-Process
The multi-source data alignment has emerged as a core limiting factors in finding root causes where missing
common shared identifiers, identifiers with different names and insufficient identifiers are the key failure
modes resulting in the data misalignment. The missing data dictionaries, inconsistent ETL routines, ontology
issues (same concept with different semantics in different data sources) and database up gradation are the key
root causes. Unfortunately there is no prevention possible until an overall database up-gradation is
performed, which is not a practical solution; hence, the proposed solutions include: (i) database audits to find
potential links between multiple data source, (ii) ETL routines audit to ensure accurate and complete data
population in data sources, (iii) upgrade data sources to remove ontology issues and (iv) add more data
dimensions along with an automated data extraction and alignment process.
b) Data Pre-Processing Sub-Process
Besides the fact that data is perfectly extracted, the incorrect data pre-processing might destroy the value
within data. The root cause analysis is performed with an objective to quickly get the value from data as it
degrades with time. The above proposed data extraction and alignment solutions do contribute in timely
extraction of the value. The key failure modes in pre-processing are outliers skipping, inaccurate filtering and
inappropriate data transformations, which refer to the inability to select and apply correct outlier detection
and removal process. It requires an automated process or good end user training so that accurate value can be
quickly extracted from the data.
108
Table 4.2 – Data extraction, alignment and pre-processing FMEA Results
109
Table 4.3 – Fast technology transfer FMEA results
110
The most critical root cause is the misalignment of PT and SPICE model parameters that result in
inaccurate PT-Specs. The PT and SPICE model parameters are the same electrical parameters but they have
different names because SPICE models are developed in new technology alliance and then are transferred to
partners’ alliance manufacturing facilities for alignment. It is the responsibility of the receiving plant to
maintain the PT-SPICE parametric relationship against each revision or maturity level of the model. The LW
scaling refers to different geometries and potential shrink due to process imperfection. The L/W
(length/width) of geometries requires expertise in the Design of Experiment (DOE) which might be an issue
without proper training and competence. The proposed solution is an automated software tool that removes
the human intervention and generates PT aligned specs as accurately as possible. It results in quick alignment
and validation of the SPICE (device and interconnect) models and technology lead times and costs are
significantly improved.
From the above discussion, it can be concluded that the core limitations are associated with the
ability to extract, align and pre-process the production line huge data volumes and dimensions. It is an
important step prior to variance analysis and model validation steps during technology alignment adoption
efforts. The proposed solutions as the operational fixes are the automated data extraction, alignment and pre-
processing utilities that provide clean aligned data for the root cause analysis.
Our proposed tool facilitates engineers in computing and aligning the SPICE parameters with the simulated
and measured electrical parameters for different technologies. It is a network driven utility being shared by multiple
engineers and it supports them to change normalization formulas on geometric specification variations to analyze the
potential impact in the simulated results. This tool has highlighted two significant failure modes (i) unstructured
naming conventions and (ii) varying file format output.
111
elements and computes, inter die, intra die and scribe line inter metal layer variances. The R&D engineers use excel
to analyze the data and it is found that they apply ± σ filter while data extraction. This step screens out the significant
deviations which are potential suspects; hence the resulting data closely follow the normal distribution which is not
the case. A multi-dimensional data model has been proposed to perform requisite computations along with a
comparison of results obtained with and without filters. This tool has resulted in the identification of missing values,
wrong filters and missing coordinates to compute the geometric specifications of interconnect metal lines as the
causes to our inability to model the newly emerging spatial variations or finding an answer to the yield drifts or
excursions. The proposed tool has significantly reduced the processing time and resulted in error free computations.
113
Table 4.2 – Device / Interconnect Modeling Traditional FMEA Results
Table 4.4 – FMEA results on cyclic failure modes and root causes
114
We also noted that with the new revolution in IT technology, new metrology equipment are capable
of providing new additional data dimensions which are not updated or included in the existing data models
based on the fact that those databases are of proprietary nature and managed by the external companies. It
has resulted in inconsistent data models with a lot of missing dimensions that might help engineers in
effective root cause analysis as well as means to effectively bridge the missing links between databases. The
size of the database is the biggest issue not because of the cost or storage capacities but subsequent data
exploitation. It is the reason that data retention period in ODS is longer than the DWH and DM. The ODS
hold single source data; hence our R&D engineers spend huge amount of time in data extraction using
single-source data sources followed by manual mapping and alignment. One more convincing reason is that
today the production and R&D engineers’ use same data sources where they have different data needs. The
production and process engineers do not need data more than the product life cycle which is 8-12 weeks
whereas R&D engineers’ need data for at least 1 year. We can conclude that today’s engineering and
production databases, data extraction tools and analysis utilities are tuned to support production and process
engineers whereas R&D engineers also try to benefit from the same source but without success. It requires a
separate R&D data source which is designed and controlled by our internal IT section and we can structure
its evolution on as and when need basis without impacting the performance of existing R&D data extraction
tools. This situation shall persist until and unless a generic R&D solution is searched and applied.
The cyclic root causes as discussed above are true challenges that result in ineffectiveness of the
DFM methods. Generic solutions have been proposed as (i) ROMMII (referential ontology Meta Model for
information integration) and (ii) R&D data model to address inconsistent data model, unstructured evolution
and data retention period issues. These solutions are presented and discussed in detail in the chapter-6.
The proposed model have resulted in a dream come true for the engineers because till now they have been
trying to find causes against newly emerging spatial variations using single-source wafer, site or die level analyses
that contributed towards the inefficient DFM methods. These inefficiencies are being compensated in the technology
alliances that result in exponential increase in the technology costs and extended technology adoption and alignment
lead times. The proposed MAM model enables the local DFM efforts for continuous technology derivative
alignment and adoption improvements. It removes the die/site level mismatches as well as generates die to site
qualification for multi-source die/site level effective root cause analyses.
b) SPM (spatial positioning) Model (SC2)
The results obtained from the MAM model enables us to base our root cause analysis at die or site levels. It is
important to note that the parameters are measured using scribe line or infield test structures. These test structures are
not true products but are taken as the representatives of the actual products and are spread across the surface and
scribe lines. The newly emerging drifts are due to spatial variations; hence root cause analyses performed based on
the die/site level might not accurately capture these spatial variations. The SPM model is proposed to perform
115
correlation based on the shortest distance between test structures used for the measurements, to ensure that spatial
variations are well taken into consideration. It computes the shortest distance and identifies the parameters along
with coordinates for the correlation purposes. It is important to note that with the increasing number of parameters
and test structures the computational cost increases exponentially. An optimized algorithm is presented that improves
the efficiency by 50%. This model is important for the technology alignment and adoption efforts specially when
improving the yield excursions.
c) ROMMII (referential ontology Meta model for information integration) Architecture
and R&D Data Model (SC3 and SC4)
The proposed architecture enables us to remove the limitation associated with the unstructured evolutions of data
models. It allows us to add new data dimensions when needed to ensure that engineers are provided with up-to-date
dimension for an accurate root cause analysis. This platform learns the Meta model for the R&D data model and
performs an accurate versioning against all potential evolutions. Any data query spanning over multiple periods is
divided into sub-queries to avoid errors and resulting data is merged prior to its distribution to end user.
The R&D data model (SC4) is proposed to avoid the inconsistencies in the retention periods for different
databases. The growth in the data sources often result in archiving the old data, however the archived data can be
uploaded to the respective databases if needed. To make it more convenient for the engineers, the database must hold
12 months data. Today, the production databases are used for R&D purposes; hence 12 months period data retention
is not likely possible due to volume and performance issues. It must also be noted that most of the existing
production databases are of proprietary nature, hence any structural change is to be requested that further delays our
efforts as well increases the costs. This R&D data model is to be maintained in house by IT or R&D people to allow
its evolution and restructuring on as per need basis under quality control loop.
The algorithms with this ROMMII platform perform pre-failure analysis upon any potential structural
change in the data model. It computes statistics at application and user levels and upon any potential change
performs a risk analysis on the likely impacted users and applications. The potential users are also intimated about
the newly added data dimension to ensure that all users are aware of any new changes.
4.5 RESEARCH SCHEMATIC AND ADVANCEMENTS (I-FMEA METHODOLOGY)
The research schematic is presented in Figure 4.4 to show the advancement. The proposed i-FMEA
methodology holds most of the advancements except the YASS (yield aware sampling strategy). The block
A0 correspond to step-1 of the proposed methodology and blocks A1 to A4 correspond to the step-2. These
blocks have already been discussed in the chapter-3 where based on the analysis results an extended IDM (e-
IDM) fablite model is proposed. The blocks A5 and A6 present the traditional FMEA results (step-3) along
with 5 industrial contributions a.k.a. operational fixes from IC1 to IC5. These operational fixes are
highlighted as screen shots of the developed and deployed software tools. The blocks A7 and A8 correspond
to the step-4 which is focused on identifying the cyclic failure modes and root causes followed by generic
solutions a.k.a. R&D fixes as SC1 to SC4. The SC5 (i-FMEA) is the scientific contribution summarized and
presented in this chapter to find cyclic failure modes and root causes.
116
Figure 4.4 - The research schematic and advancement with i-FMEA methodology
117
4.6 SUMMARY AND CONCLUSIONS
In this chapter we used our proposed i-FMEA approach to find true DFM challenges. In this approach, the first and
second steps are attributed to chapter-2 where the top ranked business objectives are identified as the “leadership
position and quick ramp-up rate” along with data extraction, alignment and pre-processing as initial failure modes.
The initial root causes identified are ontology issues, missing links between databases, missing values and varying
measurement reference frames that restrict engineers to the single-source root cause analysis. We have developed 5
different tools (i) BEOL-variance analysis to analyze parametric drifts in back-end-of-line interconnect modeling
process, (ii) KLA-Ace recipe for the mapping and alignment of coordinates to enable multi-source site level analysis,
(iii) EPP (equipment, product, process) life cycle extraction tool to extract multi-source contextual data to support
drift/excursion analysis and (iv) ACM (alarm control and management) tool to extract, analyze and manage alarm
data for the tool capacity optimization. These tools address the above initially identified root causes. The objective is
to either confirm the existing or identify new failure modes and it was surprising to find new failure modes as: (i)
unstructured data model evolutions, (ii) missing data dimensions and (iii) wrong correlation due to test structure
positions.
These failure modes were grouped as (i) ineffective root-cause analysis (infield and scribe line test structure
positions) and (ii) data extraction, mapping and alignment, which need generic solutions. A detailed FMEA analysis
on these two grouped failure modes was performed and resulting root causes were used to propose generic solutions.
The new root causes linked with the ineffective root cause analysis failure mode are found as missing and varying
coordinates due to varying metrology reference frames, rotation of the wafer prior to measurement steps and
infield/scribe line test structure positions. It is very important to note that the root causes identified could fall in
different business function and in our case these are IT and EDA.
The FMEA is also performed on the technology transfer that resulted in issues between SPICE model and
PT parameters relationship. The names of the parameters in these SPICE models are generic whereas the names used
by the test engineers are customized; hence the mismatch creates a significant problem. The initial root causes that
were found are incorrect and error prone results because of manual data alignment and pre-processing; hence, we
developed SMA (spice model alignment) tool that resulted in automation and removed the associated inefficiencies.
The further use of this software tool identified unstructured naming convention for the metrology parameters and the
varying format for the CAD simulation results as key root causes.
The generic scientific contributions have been proposed based on the identified root causes. The generic
solutions are presented in chapters 5 and 6 as (i) MAM and SPM models and (ii) ROMMII and R&D DWH model.
The proposed i-FMEA approach has evolved based on the experience through these phases where it is noted that true
failure modes and root causes can be traced to other business functions. These steps are formulated into 4-step i-
FMEA methodology as (i) identify initial failure modes using FMEA, (ii) identify cyclic failure modes through joint
projects, (iii) identify root causes against failure modes and (iv) propose generic R&D solutions. In the next chapter 5
and 6 we present (i) MAM and SPM models and (ii) ROMMII platform and R&D DWH model respectively as
generic solutions.
118
Chapter 5: Measurement Coordinates Mapping, Alignment and Positioning
11,12,13,14
All the measurements during production processes are performed on the test structures, which are placed in the
scribe lines and/or in the fields. The wafer position is marked by its notch position and it is rotated prior to
measurements depending on test structures orientation; hence, measurement coordinates potentially vary with each
metrology step. It is critically important to map and align the multi-source measurement data prior to the root cause
analysis. At present, engineers spend huge amount of time in data alignment because of the missing x,y coordinates
for measurements in the database, issues with metrology reference frames and missing identifier between site and die
levels data. A generic MAM (mapping and alignment) model for site/site and die/die levels coordinates mapping and
alignment followed by die/site qualification is proposed. In order to capture newly emerging spatial variations, it is
important to perform correlation analysis based on the position of the test structure used to measure the source and
target parameters. The SPM (spatial positioning) model is also presented that perform mapping based on the shortest
distance between test structures on the wafer surface.
Contents
5.1 Introduction to Device/Interconnect Modeling ..................................................................... 121
5.2 Challenges in Multi-Source Data Analysis ........................................................................... 124
5.3 Site/Die Level Mismatch Problem ........................................................................................ 126
5.4 Proposed Die/Site Level Mapping, Alignment and Qualification (MAM) Model ................. 128
5.4.1 Site/Site or Die/Die Level Mapping and Alignment ................................................. 128
5.4.2 Die/Site Level Qualification .................................................................................... 131
5.5 Test Structure Position Based Mapping and Alignment (SPM) Model .................................. 134
5.5.1 SPM (spatial positioning) Problem [Source/Target 1*1] Context .......................... 135
5.5.2 Step-Circle based Basic Algorithm [Source/Target (1*1)] for Mapping ................. 135
5.5.3 Example for Basic Step Circle Algorithm [Source/Target (1*1)] ............................ 136
5.5.4 SPM (spatial positioning) Problem [Source/Target 1*n] Context .......................... 138
5.5.5 Optimized Step-Circle Based Algorithm (1*n) Problem .......................................... 139
5.5.6 Example for Optimized Step-Circle Algorithm [Source/Target (1*n)] .................... 139
5.6 Data Model for Position Based Site/Site Mapping................................................................... 140
5.7 Research Schematic and Advancements (MAM and SPM Models) ........................................ 142
5.8 Summary and Conclusions...................................................................................................... 144
11 Shahzad M.K., Tollenaere M., Hubac S., Siadat A., Extension des Méthodes DFM pour l’industrialisation de produits microélectroniques, 9e
Congrès Internationale de Génie Industriel Montréal, 2011, Canada
12 Shahzad M.K., Hubac S., Siadat A. and Tollenaere M., MAM (mapping and alignment model) for inspection data in semiconductor industry, 12th
European APCM Conference, Grenoble France
13 2012Shahzad M.K., Hubac S., Siadat A. and Tollenaere M., SPM (spatial positioning model) model to improve DFM methods, 12th European
APCM Conference, Grenoble France 2012
14 Shahzad M.K., Siadat A. Tollenaere M. and Hubac S. (2012), Towards more effective DFM methods in Semiconductor Industry (SI), International
Journal of Production Research, (submitted) TPRS-2012-IJPR-0638
119
120
5.1 INTRODUCTION TO DEVICE/INTERCONNECT MODELING
It is seen that the success of SI lies in our ability to quickly align and adopt new technology transfer
followed by continuous technology derivative/improvement efforts. The device and interconnect modeling
have (section 3.4, Figure 3.3) been demonstrated as the key processes. In chapter-3 and 4, a detailed
discussion is done on the data extraction, alignment and pre-processing limitations but here interconnect
modeling process will be analyzed to highlight the data mapping and alignment problem. The basic
interconnect modeling process is presented in the Figure 5.1.
Mat0
Ma30
Mat5 Mat10 Mat20
121
Figure 5.2 – The BEOL - Interconnect modeling process (tentative model)
The production model (Figure 5.4) is the last phase in interconnect modeling process where test
structures (PCM: process control monitors, PCS: process control structure) are finalized that shall be used to
capture the process and metrology information during normal production time line. A comparison between
simulated and measured electrical and parasitic parameters is carried out with a formal report. This report is
based on the results and analysis received from the PI team on 3 lots.
122
for both typical and corner models are time consuming tasks that directly increase the technology lead times.
The improved process is presented in Figure 5.5.
123
5.2 CHALLENGES IN MULTI-SOURCE DATA ANALYSIS
The SI manufacturing operations are grouped as Front-End-Of-Line (FEOL) and Back-End-Of-Line (BEOL)
operations to manufacture the transistors and interconnections between them respectively. The production
line data is classified as the contextual and inspection data. The contextual data includes maintenance,
process recipe, Fault Detection and Classification (FDC), Work in Progress (WIP) and Out of Control (OOC)
data. The inspection data comprises of PT, Inline, Electrical Wafer Sort (EWS) and defectivity data. The
FEOL and BEOL operations can be divided into 1 to n operations where defectivity and inline data is
collected across all operations whereas the PT data is collected at the end of metal-1 and metal-7 when
contacts are manufactured for subsequent use of electrical tests. The EWS data is collected at the end of all
processes to sort dies (chips) into good and bad dies. The most commonly used type of data during root cause
analysis is the inspection data. The overall process is presented in the Figure 5.6.
124
Figure 5.7 – The BEOL – process for interconnections
The possible type of multi-source root cause analysis could be easily classified based on the levels
on which they are collected across the production lines. The PT and Inline data is collected at site level
whereas defectivity and EWS data are collected at die levels (Figure 5.6). The potential root cause analysis
could be performed at the site/site, die/die and die/site levels. The further investigation on existing
production data sources highlight that all PT measurements are supported with the x,y coordinates and
indices whereas inline data is supported only with the site numbers. It is also seen in the chapters 1 and 4 that
wafers are rotated prior to the measurements based on the position of the test structure on the wafer. It is
because of the fact that metrology equipment cannot move right or left to align itself with the test structure
input pins or pads. It is evident that PT and Inline data collected across the wafer cannot be mapped because
the site numbers present in both measurements correspond to different coordinates. In the absence of x,y
coordinates for inline data it can be concluded that site level mapping is not highly difficult, time consuming
and error prone. The die level x,y coordinates do exist in the source data files, however neither these values
125
nor the position of the wafer are uploaded in the database; hence it is highly difficult for the R&D engineers
to perform even die/die level correlation analysis. The die/site level analysis is also not possible because
there do not exist common identifiers that can be used for die/site level mapping.
126
The reference frames used for the metrology can be shifted from one to another by a clockwise
rotation to accurately map source and target site/die level measurements. The site/site level mismatches are
presented in the Figure 5.10 where PT and inline measurements with notches at bottom and right respectively
follow the 1st reference frame. Hence, notch position of the inline measurement is rotated by 270°
anticlockwise to shift its notch position downwards for site/site mapping.
127
different notch positions and missing die/site level qualification data. First of all, it is required to rotate the
die level wafer data by 270° clockwise and 270° anticlockwise to bring it to the 1st reference frame with
notch position at the bottom. The x and y components of sites (inline data) associated with each die in EWS
data must be found (Figure 5.12) for the die/site qualification.
5.4 PROPOSED DIE/SITE LEVEL MAPPING, ALIGNMENT AND QUALIFICATION (MAM) MODEL
From the problem context discussed above in section-3, the problem is separated in two parts as (i) site/site,
die/die mapping and alignment and (ii) die/site qualification. Let us first start with the site/site, die/die
mapping and alignment issue.
P’ (x’, y’)
P (x, y)
r
r
r sin(φ)
θ
φ
r cos(φ)
If this point P(x,y,φ) is further rotated with reference to the same origin by θ° to P′(x’,y′, φ+θ) then it can be
represented as under:
(3)
(4)
128
The new coordinates of the point P′(x’,y′,φ+θ) with the substitution can be rewritten in the matrix notation
as under:
(5)
(6)
(7)
In site/site and die/die mapping and alignment problem each site/die within the source and target
wafers is taken as a point located in different reference frames with different notch positions. So the above
presented concept is used to always rotate the coordinates of our sites and dies to the 1st reference frame
with notch at bottom. In the proposed model is taken as the clockwise angle between two reference frames
and as an anticlockwise angle that rotates the sit/die level coordinates to move the notch position at bottom.
To better understand the steps involved in this proposed formulation let us present an example as
shown in the Figure 5.14 where a wafer is located in the 2nd reference frame and the 4th quadrant so all die
or site level coordinates are known with reference to its origin. It is important to find these coordinates with
reference to the origin located in the 1st reference frame. The original coordinates of the wafer are rotated by
= - 270º to get new coordinates in the 1st reference frame. The resulting wafer is found in 4th quadrant and
1st reference frame but to move the notch position at bottom the wafer is further rotated around its center by
θ = 90º followed by its translation to center.
Figure 5.14 – Reference frame and notch position rotation and translation
The problem is generalized such that the point is located in the 2nd reference frame and the
coordinates of this point are known with reference to its origin .We have to find the coordinates of
the point with reference to the new origin located in the target reference frame where this
point shall be further rotated by the angle to get the notch position at bottom. Finally it shall be translated
to the center of the reference frame. This formulation is presented in Figure 5.15.
Figure 5.15 – Generic formulation for reference frame and notch position rotation
129
The points , and can be represented as vectors. The said point
is known with reference to the point and our objective is to compute the same point with
reference to . The both origins and have x and y coordinates as (0,0) and are
superimposed on one another hence the angle between them is taken as 0. The new coordinates of the point
with reference to can be found by taking a vector sum as under:
(8)
In this formulation the is a vector between points and , is a vector between points
and and is a vector between points and . The matrix representation of the vector sum based on
the Cartesian coordinates (transformed polar coordinates) is as under:
(9)
(10)
In this formulation and shall rotate the reference frame and the notch position respectively
followed by its translation to centre given by and a.k.a. alignment factor.
Now the algorithm developed to normalize the given wafer coordinates is presented below. Let us start with
the description of the variables (Table 5.1).
Step-1: Initialize
Step-2: Compute ( ), , ,
130
transformed to the 1st reference frame followed by its rotation of the notch position at bottom and translation
towards the center. We shall transform P(-4,2) ’ P’(x’,y’) [ = -180, = 90].
Figure 5.16 – Reference frame and notch position rotation with translation example
After transformation of all the coordinates, the results are presented in the Figure 5.17.
131
find the exact wafer center followed by X,Y (min/max) distances from this computed wafer center in the
central site/die. It is the key information to accurately count the number of sites or dies along X+, X-, Y+ and
Y- axes. The step-1 to step-3 in the algorithm corresponds to these computations. The counted sites and dies
are further used to find the X and Y site components for each die as per step-4 and step-5 in the algorithm.
Figure 5.18 – Die/Site qualification, mask and wafer center and site counts
Step-1: Compute the center of the wafer in site source
If then
else
Step-2: Initialize 0
Step-3: Compute
While ,
While ,
While ,
132
While ,
Update
Update
The Edge Bevel Removal (EBR) plays a very critical role during production and is defined as the
outer most area on the wafer surface which is not used during production, so this area must be excluded prior
to any computations. It can be seen in the Figure 5.19 below as the distance between the outer solid and
dotted lines. The above algorithm takes an assumption that source and target wafers are already mapped and
aligned to the 1st reference frame with notch position at bottom based on our model as presented in section
4.1. To validate the proposed die/site qualification algorithm, we start with an example (Figure 5. 18) where
site (x, y) and shift (x, y) have the coordinates (10, 15) and (3.5, 6.5) respectively. The wafer under
consideration has a 300 mm diameter and 3mm (EBR) edge base removal. The shift (x,y) is used to compute
the center of the central die/site from the actual wafer center; hence the wafer center is simply defined by
shift(-x,-y) as (-3.5,6.5). This wafer center is used to compute the minimum and maximum distances along x
and y axes as Min(x)=1.5, Max(x)=8.5, Min(y)=14 and Max(y)=1. The number of sites along x and y axes
are computed with reference to the wafer center as per step-3 of the algorithm.
10
Corners 5
9
8
4
7
Online Metrology 6
3
Circuit 5
4
A 2
3
2
HS-VS Lines 1
1
0
0
-1
-2
B -1
-3
PCM Structure
-4
-2
Top Left Top Right -5
PCS Structure
-6 -3
Center -7
Center Left -8 -4
-9
Notch Position
The Figure 5.22(a) shows a simple site in the wafer with both infield and scribe lines test structures.
The 5.22(b) shows the source and target parameters that must be mapped based on the site level mapping and
alignment. The position of the test structures with reference to the center point is known; hence the shortest
distance between source and target parameters can be easily computed as presented in 5.22(c). In an
anticipation of alternative mapping, a step-circle is drawn and distance between source parameter of source
site ‘E’ is computed with all the sites in step circle as shown in 5.22(d) and 5.22(e). Based on the shortest
distance formulae the source parameter of source site ‘E’ is mapped with the target parameter of site ‘F’. It is
more realistic in an anticipation of tracking the spatial variations emerging especially in the development of
new technologies.
135
The step-circle based basic algorithm for site to site mapping is presented as under:
Step-1:
Step-2:
Step-3:
4
TP = (-3.5, 6.5)
3
2 A
0
15
-1
C = (0,0)
B
-2
-3
-4
-5
SP = (-4.5, -7.0)
-6 -5 -4 -3 -2 -1 0 1 2 3 4 5 6 10
(11)
(12)
= 13.536 (13)
The graphical description of this computation is presented in the Figure 5.25(a). The computation of
source and target distance between source site and target step-circle sites is presented in Figure 5.25(b),
5.25(c), 5.25(d) and 5.25(e). All computed distances are presented in the Figure 5.25(f). The Step-3 of the
proposed algorithm finds the shortest distance resulting in mapping of the source site (-2, +2) and target site
(-2, +1).
(a) (b)
(c) (d)
(e) (f)
(Site Level)
(a)
(Die Level)
(b)
138
5.5.5 Optimized Step-Circle Based Algorithms (i*n) Problem
The variable declaration in the table 5.3 for the basic algorithms hold valid in this section, however
additional variables are presented as under in Table 5.4.
Notations Description Notations Description
Distance from the source parameter to the
top left (TL), bottom left (BL), top right Direction of traversing
(TR) and bottom right (BR) corners
Table 5.4 – Description of step-circle (O) algorithms
The proposed algorithm is presented as under:
Step-1:
Step-2:
Step-3:
Step-4:
140
The proposed data model fully supports the spatial variation modeling on the wafer surface for PT, Inline,
EWS and defectivity measurements data.
141
5.7 RESEARCH SCHEMATIC AND ADVANCEMENTS (MAM AND SPM MODELS)
The research schematic and advancement is presented in Figure 5.30. The MAM (SC1) and SPM (SC2)
models are key generic contributions made in this thesis which lead us towards a shift from data driven
ineffective DFM efforts towards information and knowledge driven efficient DFM efforts. The proposed
generic scientific contributions address the cyclic failure modes and root causes identified in the step-3 of i-
FMEA methodology presented in chapter-4. Our proposed i-FMEA methodology has significant advantages
as it helps in identifying the
Cyclic failure modes and root cause which are repeated until and unless they are fixed with generic
solutions. The proposed MAM and SPM models ate the partial contributions in solving the issues associated
with data extraction, mapping and alignment during technology alignment and adoption efforts.
142
Figure 5.30 - The research schematic and advancement with MAM and SPM models
143
5.8 SUMMARY AND CONCLUSIONS
In this chapter, we have analyzed the traditional interconnect modeling process being one of the critical
processes in technology alignment and adoption initiatives. The key objective is to analyze the effect of
cyclic failure modes and causes as: (i) data extraction, (ii) pre-processing and (iii) multi-source root cause
analysis, so that generic R&D solutions can be proposed. The BPR analysis on the interconnect modeling
process has resulted that the cyclic failure modes and root causes have a significant influence on the model
correction and interconnect stack definitions steps that lead to an exponential increase in technology lead
times and associated costs. We have proposed a revised interconnect modeling process by adding a
knowledge base, which provides an initial feedback for the first time correct stack definition followed by
quick model correction based on effective root cause analysis.
The effective root cause analysis is the key to improve existing ineffectiveness in the DFM methods
and is defined as the ability to analyze data to find answers against drifts and/or variations. These drifts and
variations are further classified as systematic or random for their transformation into rules and/or models. At
present, engineers are facing huge difficulty in the data mapping and alignment to perform multi-source
analyses based on fact that the measurements result in varying coordinate system due to test structure
orientations. The wafer is rotated prior to the metrology steps and its position is monitored using notch or flat
positions. The varying coordinates issue is further complicated with different metrology reference frames.
These issues have a significant impact on stack definition, model correction and validation steps.
We have proposed mapping and alignment (MAM) model and spatial positioning (SPM) model. The
MAM model is capable of performing site/site, die/die and die/site levels data mapping and alignment. It is
highly important in correlation analysis because inspection data are captured at different levels (PT and
Inline are captured at Site and defectivity and EWS data are captured at die levels). The proposed MAM
model enables multi-source root cause analysis resulting in quick knowledge capitalization. We know that
during production the wafers undergo multiple inspection steps and all the tests are performed on test
structures placed inside the field and/or scribe lines. To accurately capture the spatial variations we have
proposed SPM model, which enables the mapping of source and target parameters based on distance between
tests structures used for measurements. The analyses results are more accurate and help us in capturing the
spatial variations emerging due to miniaturization. The MAM model is suggested for mapping and alignment
during technology adoption efforts whereas SPM model is a best fit during technology alignment initiatives.
We concluded the chapter with a data model, which is filled using source metrology, test program
specifications and layout files, provided by the Reticule Assembly Teams (RAT). The data model provides
an efficient way of internal processing prior to data mapping and alignment. The solution for missing data
dimensions and continuous data models restructuring for the successful evolutions is provided in next
chapter, which results in a generic solutions, focused on removing weaknesses and improving DFM
effectiveness. It enables an effective multi-source root cause analyses for the R&D and/or
product/process/equipment engineers.
144
Chapter 6: ROMMII and R&D Data Model for Information Integration15
In this chapter, the cyclic failure modes and root causes associated with multi-source data extraction function are
addressed. The key causes are identified as (i) unstructured data model evolution, (ii) ontology issues and (ii) data
retention periods. The ROMMII platform is presented to address unstructured data model and ontology issues where
the R&D model provides an efficient solution to data retention issues easing production data sources. The ROMMII
platform enables us to exploit huge data volumes and dimensions by removing (i) model inconsistencies, (ii) pre-
failure assessments to avoid extraction and analysis utilities failures, and (iii) information diffusion to perspective end
users about the inclusion of new data dimensions. The R&D data model resolves the issue of different data retention
periods in existing production sources and proprietary constraints associated with multiple production data sources.
Contents
15 Shahzad M.K., Hubac S., Siadat A., Tollenaere M., ROMMII (referential ontology meta model for information integration) Architecture for
Dynamic Restructuring of DWH data models, 12th European APCM Conference, Grenoble France, 2012
145
146
6.1 INTRODUCTION
Looking at the historical evolution of data needs by R&D engineers as presented in Figure 6.1, it can be seen
that in past engineers have been complaining about few data volumes and dimensions for their inability to
capitalize knowledge. The memory and computational costs were high in comparison to the performance and
efficiency. However recent revolutions in the IT technologies have resulted in continuous reduction in the
memory and computational costs, and increase in the computation power. It is because of this fact that today
we are able to store large data volumes and can exploit them at high efficiency BUT R&D engineers are still
complaining and this time it is not about the volume and dimensions but their inability to exploit these huge
volumes and dimensions. As a consequence, the productivity of our R&D engineers is low which impacts the
competitiveness of SI resulting in extended technology alignment and adoption lead times, so it is necessary
to work on solutions to facilitate engineers in their efforts to quickly transform the data into information and
then knowledge for technology lead times and costs reductions.
147
with sequential, indexed and random access methods. The third generation programming languages (e.g.
COBOL, BASIC) were used to write programs to read, write, delete and update the data in these file. These
file systems are still used today for the system configuration files but they are not being used for commercial
purpose because of the limitations as (i) weak security, (ii) no sharing, (iii) likely duplication and (iv) high
maintenance costs for data consistency and access controls [Impagliazzo, 2012].
6.2.2 Non-Relational Database Era (1968-1980)
In this era the emphasis was to overcome the limitations associated with the initial file systems and IBM
again took lead and proposed the first non-relational databases management system called IMS (information
management system). The IMS is a hierarchical database with a tree like storage architecture having parent-
child as one-to-many relationship. It was developed in a joint project with Rockwell and Caterpillar (1966-
1968) to manage large bill-of-material (BOM) for the Apollo space shuttle. The network database model has
an advantage of efficient searching over flat file databases. It also results in less data redundancy, security
and integrity; however the biggest disadvantage is the difficulty in implementing the many-to-many
relationships [Blackman, 1998].
The second major contributions in this era are the network database models the (i) CODASYL
DBTG and (ii) IDS model. The CODASYL (conference on data systems languages) also credited for the first
general purpose business programming language (COBOL) in 1968. The database task group (DBTG)
proposed the specifications for the network database model in 1969 along with interfaces in the COBOL
language for data definition (DDL) and data manipulation languages (DML). The network database model
(IDS at Honeywell) uses directed acyclic graph with nodes as records and edges as relationships [Bachman,
1973]. It has the similar advantages like IMS with many-to-many relationship; however, it lacks structural
independence [Blackman, 1998].
6.2.3 Relational Database Era (1970 till present)
The relational database is the first major breakthrough of the century that has changed the way in which data
could be stored and used for the decision support systems. There is no doubt that the rich database efforts in
terms of hierarchical and network based databases resulted the relational databases. The idea of relational
database model was initially proposed by Ted Codd at IBM in 1970 [Codd, 1970]. This model provides
advantages over all above discussed database models as (i) no redundancy, (ii) security and integrity, (iii) all
types of possible relations and (iv) the economies of scale whereas, the only associated advantage is the size
and cost of the DBMS. It has also led the emergence of object oriented database management systems
(OODBMS) around 1985; however, it has resulted in little success because the conversion cost is too high
for the billions of bytes of data [Blackman, 1998]. The examples of the RDBMS in this era are (i) Ingress
which ended into Informix and (ii) System R by IBM which resulted in DB2.
6.2.4 Dimensional Database Era (1990 till present)
The revolution in the database technology for efficient data storage and subsequent retrievals was supported
with the decreasing memory and computational costs. It has resulted in the availability of huge volume of
data. The drawback of size associated with the relational databases has resulted in the inefficient data
exploitation with rise of interest in the methods for quicker response queries. The dimensional database era
started (1990) with the emergence of enterprise resource planning (ERP) and management resource planning
(MRP) concepts. The biggest advantage is that data is stored in multi-dimensional cubes which results in
quicker aggregation and faster response to the end user queries [Blackman, 1998].
The above discussion on the evolution from flat-files towards dimensional databases indicates that
the driving force behind all these efforts is the ability to efficiently store and exploit data resources with the
security and integrity. In today’s competitive environment where knowledge is the key for success, it is
needed to transform data into information and knowledge so that varying processes and business strategies
can be reformulated for competitive gains. We hypothesize that the principle objective is data/information
148
integration for the knowledge capitalization. Now we shall focus in detail on the dimensional database
systems in order to assess its appropriateness for our proposed e-IDM business model (section 3.7, chapter-
3).
6.3 EXISTING DATA/INFORMATION INTEGRATION SYSTEMS
The existing information systems used for the data/information extraction/integration efforts are classified as
online transaction processing (OLTP) and online analytical process (OLAP) systems. The OLTP systems are
supported with the relational databases and are focused on efficient insert, update and delete queries whereas
OLAP systems require DWH/DM architecture with emphasis on data retrieval. It is also important to note
that all business intelligence (BI) tools require the ability to quickly aggregate, dice and slice data for
decision making purposes using OLAP cubes which are built from the DWH/DM databases. A brief
comparison is presented below for reference:
The consistent and validated data is bulk loaded for The database is optimized for the validation of data
large, complex and unpredictable queries and transactional processes
The query response is optimized with multiple It uses few indices as the emphasis is on the
indices with few joins accurate transactions with many joins
It follows periodic updates through extraction,
transformation and loading (ETL) routines with few It requires frequent modifications
frequent modifications
It operates with huge data volumes and large data It operates on small data volumes and few
aggregations aggregations
149
performance and efficiency. The DWH and DM can be implemented using relational database management
systems (e.g. Oracle, SQL Server) or purely dimensional database management systems (e.g. TeraData). The
relational database systems provide additional utilities to build OLAP cubes whereas dimensional database
systems provide an edge on relational databases in terms of query performance and ability to treat huge data
volumes.
6.4 DWH-DM: INFORMATION INTEGRATION AND BUSINESS INTELLIGENCE PLATFORM
The growth in the availability of huge data volumes and technologies to store and process them has resulted
significant challenges in the integration. The DWH and DM databases as discussed above provide us an
opportunity to efficiently store and exploit the huge data volumes. The latest BI tools use the DWH and/or
DM architecture to quickly respond to the varying information needs of R&D engineers. The DWH and DM
are defined as under:
DWH: The DWH is a subject oriented, integrated, non-volatile and time-variant collection of data
to support management decisions [Inmon, 2005]. It is a relational database that uses ETL routines to
populate the DWH with historical data which is collected and validated from multiple data sources.
It takes off the processing load from the transactional databases (OLTP) for in-depth but
computationally expensive analysis.
DM: The DM can be simply defined as a smaller version of a DWH with the data from one source
e.g. sales. The data stored in the data marts is highly or partially summarized.
Let us start with the basic definitions and concepts prior to discuss the new database architecture for
DWH and DM.
6.4.1 Basic Definitions and Concepts
Let us start with basic concepts and definitions to better understand the problem and further discuss the topic:
o Facts: It is a type of table in the dimensional database which includes two types of columns (i) facts
(quantitative measures) and (ii) foreign keys to dimension tables. The facts are numerical
measurements of a certain business process and they are aggregated in the OLAP cubes e.g.
aggregated sales volume against country, region or city.
o Dimensions: The dimensions are the tables that store records related to the particular dimension.
These are the attributes by which the facts in the fact table are grouped. The dimensions that change
over time are referred as slowly changing dimensions and can be treated by (i) overwriting the old
value, (ii) adding a new column and (iv) adding new row and version. The dimensions are classified
as confirm dimensions if it is linked to multiple fact tables whereas the single dimensions, with few
attributes having yes/no values, are referred as junk dimensions. They are grouped into single
dimension to reduce number of referred dimensions in respective fact tables.
o Attribute: It refers to the fields within fact and dimensions tables.
o Hierarchy: The dimensions are further decomposed into sub-dimensions and it is referred to as
dimensional hierarchy e.g. time dimension (year, quarter, and month) and geographic dimension
(country, sales region, state, city and store).
o Drill up/down: The drill up/down are specific operations applied on the OLAP cube where
the data is presented at higher or lower hierarchies based on the dimensions. It requires either
ROLAP or HOLAP cubes (see section 6.3).
o OLAP Cube: It is defined as a multidimensional representation of data from the fact and dimension
tables. The internal storage of OLAP cube is different but efficient as show below.
150
Produc
t
Metrology
Equipment
Process
o Slicing and Dicing: The slicing operation is defined as the process of retrieving data from
OLAP cube by filtering it on a given dimension whereas dicing operation filters the original
cube data across all dimensions as show below.
Product Slicing
Dicing
Metrology
Equipment
Process
Figure 6.3 – Slicing and dicing operations on OLAP cube
6.4.2 The DWH Architectures, Models and Schemas
In order to better understand the advantages associated with the DWH and DM databases, it is important to
discuss its potentially possible architectures, models and schemas as under:
DWH Architectures/Frameworks: The most commonly used DWH architectures are as (i)
basic DWH architecture (Figure 6.4a), (ii) DWH with staging (Figure 6.4b) and (iii) DWH and DM
with staging (Figure 6.4c). In the basic DWH structure, the data from different data sources is
directly moved to the data warehouse through ETL routines and end users are provided with direct
access over the DWH through OLAP based data extraction and analysis utilities. The disadvantage
associated with this architecture is the inconsistencies in the ETL routines; hence, data validation
could raise serious issues during subsequent analysis by end users. The staging is a temporary
storage which is used for the validation of data prior to its permanent storage in the DWH; hence, it
overcomes the drawbacks associated with the basic DWH architecture (Figure 6.4 b & c). The
typical DWH architecture being used in most of the BI tools, where data from the DWH is further
aggregated to the DM and end users are provided with access to these data marts for efficient query
responses (Figure 6.4c).
151
c) DWH and DM with Staging Architecture
Figure 6.4 – Principle DWH architectures and frameworks
DWH Models: The DWH follows three modelling levels as the (i) conceptual, (ii) logical and (iii)
physical models (Figure 6.5). The conceptual model provides the highest level of abstraction where
only names of entities and their relationships are considered. The logical level in comparison to
conceptual level describes the system in much more detail; however, the physical implementation is
still an abstraction. It includes identification of attributes and primary and foreign keys followed by
normalization process. The physical model takes into account the target database management
system (Oracle, SQL Server, MySQL, Access etc.) constraints and data types and domains are
detailed for each attribute. The physical model can be demoralised at this level based on the user
requirements.
152
DWH Schemas: The data warehouse design and development use three commonly used schemas
as (i) Star, (ii) Snowflake and (iii) Constellation Schemas. The Star schema (Figure 6.6a) consists of
fact and dimension tables; however, there are no hierarchies for dimensions or the fact tables. The
fact tables do hold quantitative or additive facts because primarily the DWH is focused on
aggregation on different dimensions and periods. These are huge tables with millions of records
where dimensions are used to build OLAP cubes which are further sliced, dices or drilled up/down to
search relevant information. The queries used within the OLAP cube are SQL queries with multiple
joins. It is very important to note down that as the number of joins increases, queries performance
degrades but DWH architecture has a multi-dimensional storage with multiple indices for fast query
results. The major difference between Star and Snowflake schemas is that the snowflake schema
allows hierarchical dimensions; however, all dimensions are linked to the central fact table (Figure
6.6b). The advantage is that multi-dimensional hierarchy results in data aggregation at multiple
levels where information is sliced and/or diced with the precise granularity. The multiple
fact/dimension tables’ joins affect the query performance but such limitations are compensated with
additional indexing. The third most commonly used DWH schema is the Constellation schema
(Figure 6.6c) which is also called the hybrid of Star and Snowflake schemas. In this schema, the
facts tables are divided in multiple fact tables and hierarchical dimensions are shared by these fact
tables.
Region: Sub-Dimension
-Region_ID (PK) Region Sales: Fact
Sales: Fact Sales: Fact Product: DImension
-Description Time: Dimension -Product_ID (FK)
-Store_ID (FK) Time: Dimension -Store_ID (FK) -Product_ID (PK) -Time_ID (FK)
-Time_ID (FK) -TIme_ID (PK)
-TIme_ID (PK) -Product_ID (FK) -Name -Region
-Product_ID (FK) -Year
-Year -Time_ID (FK) -Description -Qty_Sold
-Qty_Sold -Month
-Month -Qty_Sold -Make -Sales_Value
-Day
-Sales_Value -Day -Sales_Value -Color
Model: Sub-Dimension
-Weight
-Model_ID (PK)
-Description State Sales: Fact
Product: DImension -Marker_ID (PK) Product: DImension
-Product_ID (PK) -Time_ID (FK)
-Product_ID (PK) Time: Dimension -Product_ID (FK)
-Name
-Name -TIme_ID (PK) -State
-Description
-Description -Year -Qty_Sold
Markerl: Sub-Dimension -Model_ID
-Make -Month -Sales_Value
-Color
-Color -Marker_ID (PK) -Weight -Day
-Weight -Description
Figure 6.7 – Data warehouse Meta model for business intelligence [Darmont et al., 2007]
6.4.3 Inmon and Kimbell DWH Philosophies
William Inmon and Ralph Kimball are well known in the domain of information management for decision
support and have played a pivotal role in changing the information management concept. Mr. William Inmon
is known as the father of data warehousing and is credited for proposing the concept and term ‘Data
Warehouse’ in 1991. Mr. Ralph Kimball is known as the father of business intelligence and is credited for
the concept ‘Data Marts’, star schema and snowflake schema. These legends have contributed most to the
domain of data warehousing and business intelligence tools. In order to better understand the contributions of
these great people let us define the term business intelligence (BI) as BI = Inmon’s Corporate DWH +
Kimball’s Data Marts + Data Mining. It can be said that both subjectively focus on developing means
(DWH+DM) to efficiently exploit the huge data volumes to ultimately support the decision making process.
Inmon’s advocates that it is necessary to start building large enterprise wide data warehouse
followed by multiple data marts to meet with the customized or specific needs of different business functions
within the organization. In comparison to this philosophy, Kimball proposes that we must start by building
data marts to serve data analysis requirements of respective organizational business functions followed by its
virtual integration for enterprise wide data warehouse. These approaches by Inmon and Kimball can be stated
as the ‘top down’ and ‘bottom up’ philosophies respectively; however, both are based on the dimensional
modeling.
154
The Kimball approach is more realistic and follows the natural sequence of enterprise information
evolution where departments started modeling and managing their information needs as per business
requirements and later the need to interlink these information sources resulted in the enterprise wide
information systems. The Inmon’s approach is difficult to follow because of the fact that it is difficult to
potentially forecast the information needs in advance; hence, even if historical data/information is stored we
are not sure, if it shall ever be used? In comparison, Kimball’s method is more practical as it give rise to the
evolution of the real information needs and once modeled as DWH and DM serve the basis for business
intelligence. The only threat to this approach is that constantly changing business requirements and needs
requires frequent modification in the data models which results in high level of complexity because of cross
version queries. The Inmon’s model in longer run requires fewer changes and can come up as a robust and
reliable source for analyses purposes.
155
databases are not built for the fast retrieval of summarized data from huge data collection; hence, even upon
the implementation of summary tables, retrieval efficiency cannot be compared with DWH projects. The
DWH projects are focused on fast data aggregation and analyses across multiple dimensions to fulfil the
varying information needs.
156
The proposed R&D data warehouse is presented as under in the Figure 6.8 a & b. The basic concepts
of facts and dimensions tables and DWH architecture are used in the context of multi-source data extraction,
mapping and alignment from huge data volumes. It is very important to note that at present, there are a
number of excellent data analysis tools based on advanced data mining and artificial intelligence techniques
but they cannot give results until they have multi-source input data in the correct format. The concept of
multiple fact tables linked together like hierarchical dimensions has been proposed with an objective to
accurately map and align different facts for further analyses.
The presented schema is the logical representation of DWH model without cardinalities and
associations because as per our best knowledge till now there do not exist a standard modelling language for
the DWH/DM schemas. The above presented DWH model is partial representation of actual schema because
of confidentiality issues with contextual facts and dimensions, however, the above model is sufficient to
present and demonstrate the concept. In this representation, [S.D], [D], [C.F] and [M.F] notations are used
for sub-dimension, dimension, contextual facts and measure facts respectively. The product dimension has
product_type, mask_set, process and mask sub-dimensions. The process dimension summarizes the process
information for the given product and can be used for further exploitation. The mask_set and mask sub-
dimensions provide critical information about wafers’ structural description with quantitative measures. The
process_plan dimension is composed of brick, operation, EDC_plan (engineering data collection) and step
sub-dimensions. It provides complete description of the total steps including process and measurement
during production operations for each product. The measurement steps are linked with the EDC plan which
further links it with the test_program dimension to list out parameters to be measured in this respect. The
lot_wafer dimension provides the list of effected lots and names of the wafers included in the lot. The
chamber-tracking and run are the additional contextual facts linked to the lot_wafer_steps central contextual
fact. The lot_wafer_steps is the central contextual fact which provides the actual execution of each
production process step in detail for slicing/dicing operations during root cause analysis and knowledge
capitalization efforts.
In this model we have included 4 measure fact tables as (i) PT_Measure, (ii) EWS_Measure, (iii)
Inline_measure and (iv) Defectivity_Measure facts. Each measure fact is linked with the fact in the
lot_wafer_steps fact table. The defectivity_measure fact is linked with the defectivity_program dimension
and class_lookup and class programs sub-dimensions. The facts collected in this case are at the die levels and
it can be seen that there are no traces of site level information at this level. The PT, EWS and Inline measure
facts have the test_program dimension and test_bin_def, test_bin_relation, test_program_specs,
test_structure_map and parameter sub-dimensions. The test_bin and test_program_specs provide list of
parameters that can be tested against the given test_program along with specification limits. The
test_structure-map sub-dimension is critical because it provides us access to the test structures and their
location which are used during the measurement of the parameters. The additional contextual facts and
dimensions which are helpful in establishing effective R&D efforts e.g. fault detection and classification
(FDC) information is not presented in the above model being highly confidential. This and similar
information can be added any time in the similar structure using our proposed Meta model for knowledge
capitalization (Figure 6.8).
157
Figure 6.8(a) – Proposed R&D data warehouse for knowledge capitalization
While designing the R&D DWH model, there were two options as (i) the measured fact data could
be treated prior to its storage for coordinates mapping and alignment using our proposed MAM and SPM
models or (ii) store data in original shape with supplement or missing information like notch position,
measure reference frame and wafer structural information and treat the data upon its extraction. The second
option was chosen because it is believed that the data should be kept in its original format and end users must
be given an option to transform it on as required basis. The advantage is that data is in its original shape and
it can be used for different transformations as per choice of the end users. The proposed R&D DWH model is
implemented using relational databases (MS Access); however, the rise of data volume for R&D and
engineering tasks dictates us to select a commercial RDMS like Oracle, SQL Server or TeraData. The end
users can use the above model to start the analysis from individual measured facts (PT, Inline, EWS, and
Defectivity) and subsequently move towards its correlation with other measured facts and contextual facts. It
is not possible to accurately define the potential size of the database to justify the use of commercial database
management systems, but for the R&D and engineering initiatives, the estimated size for one year R&D data
is expected to rise above 1.5 TB (tera bytes).
158
Figure 6.8(b) – Proposed R&D data warehouse for knowledge capitalization
The above presented data model for R&D DWH is a simple logical presentation with notations as
[S.D], [D], [C.F] and [M.F] and without the declaration of primary and foreign keys. It is done on purpose to
highlight that to the best of our knowledge there do not exist a standard DWH modelling language like UML.
The DWH data model is not normalized and to certain extent duplication of the data is allowed to improve
data retrieval efficiency. If a DWH is implemented in RDBMS then indexing and views can be used to
improve the data retrieval efficiency; however, the best way is to use OLAP tools provided by the respective
vendors to slice/dice and drill up/down the OLAP cubes for knowledge capitalization. It is also important to
note that no primary or foreign keys are added or highlighted because duplication is allowed. The fact tables
include keys from dimensional tables (foreign keys). They are used as composite key in the fact tables;
however, if required, a simple identifier can be added within the fact tables. The same rule is also applicable
for the dimension tables.
A Meta model could be extracted (Figure 6.9) for the knowledge capitalization from the above
presented R&D DWH model to improve R&D effectiveness. In this schema, there are central fact tables that
hold the contextual data whereas child fact tables associated with the respective central fact table are the
quantitative measured facts. The central and child fact tables are connected in a star formation whereas the
dimensions are shared across these fact tables and they are placed in hierarchal association resulting in the
formation of flower schema. This schema has commonalities with star and constellation schemas as (i) star
schema and (ii) shared hierarchal dimensions. The major differences in existing and proposed schemas are
that in our proposed schema (i) contextual data is placed in central fact tables and (ii) quantitative measures
or facts are stored in child tables in the star schema. The objective of the proposed flower schema is to enable
159
huge data volume exploitation for the purpose of mapping and alignment across multiple dimensions for
effective root cause analyses (knowledge capitalization) but existing schemas are focused on efficient and
fast aggregation of quantitative measures.
1..*
-Parent/Child 0..*
1..*
1..*
0..* 1..*
0..*
0..*
-Parent/Child 0..*
Attribute
Value
Contextual Fact Table Measure Fact Table
1..*
0..* 1 0..*
Context Measures
M.F M.F
M.F
D M.F
M.F C.F
C.F M.F
S.D D S.D
C.F D
M.F
C.F M.F
C.F
D
M.F M.F
The above SQL multiple join query shall result all PT and EWS parameters IDs and values.
The above example returns the same PT and EWS data at die and site levels but along with the test
structure positions. These test structure positions are further used in the SPM models to perform accurate
mapping based on site and die levels generated through MAM model. The key is that this data model
provides and generates all type of measured and contextual data needs. The data extracted is further
processed using MAM and SPM models for the coordinate normalizations. The above presented SQL queries
are not the only set of queries, stored procedures can also be built as per user needs which can be further
supported with user define SQL functions for slicing, dicing and drill up/down operations. The size of one
year’s data storage is estimated to be 1.5 TB which means that multiple join queries are likely to crash;
hence, OLAP supported RDBS (Oracle, SQL Server) or purely dimensional DBMS as TeraData are needed
for efficient and fast query results. To best serve the end users, it is suggested that based on this R&D DWH
161
model, customized data extraction utilities must be developed and provided to R&D engineers for the
productivity and improvement.
6.6 PROBLEM CONTEXT AND CURRENT CHALLENGES
Let us discuss current challenges in the data extraction and analyses framework (Figure 6.11) where there are
multiple engineering and R&D teams working for the technology derivative/improvement
alignment/adoption and process control efforts and initiatives. The R&D efforts by these teams require data
extraction and analysis for which there exist a large number of customized applications CA1, CA2 … CAn.
The end users use these applications to access the data which is stored in 3-level storage architecture (i)
ODS, (ii) RDBS and (iii) DWH and DM. The data extraction requests are classified as single-source or
multi-source data analysis initiatives. At present, the single-source data extraction and subsequent analysis is
likely possible and is being carried out but multi-source data extraction is not possible because of the missing
database links, missing values, incompatible data formats and users’ inability to use multi-source data
extraction utilities. The IT revolutions have resulted in huge data volumes and availability of new facts
which must be constantly updated in existing data resources to improve the R&D effectiveness. This is not
possible because of the fact that the operational data sources are of proprietary nature and its structure cannot
be changed. It results in a situation where multi-source analysis is never possible and often results in the
waste of time and resources when engineers try to align and map the data using excel sheets.
The proposed R&D DWH solution to exploit huge data volumes for mapping and alignment is
sufficient to provide means for effective root cause analyses; however, it does not provide solution for
unstructured evolution of data models that have been established as critical factor for effective R&D efforts.
6.7 PROPOSED ROMMII FRAMEWORK
The proposed R&D DWH does not provide solution for the relational data model evolution because they are
proprietary databases and changing the database structure shall result in crashing end-user applications. It is
also not possible to redesign and developing the existing relational databases and applications; hence, we
have proposed ROMMII (Relational/Referential Ontology Meta Model for Information Integration)
framework to bridge the gap. This framework provides a high level Meta model to implement the ontology
for information integration without changing the structure of existing data sources. It is implemented as a
relational database instead of using ontology based languages. This is because of the fact that SQL query
performance for the validation and optimization of the end-user requests is much higher if manipulated as a
162
relational model. The relational database is also easy to manage and control against the rapid and frequent
modifications.
The proposed framework is called Meta Model because it precisely defines specifications of the
relational model which is used to track the modifications and missing links between databases. It is referred
as ontology because of the fact that it offers a shared collection of missing knowledge across the databases
for subsequent use during query pre-processing and its reformulation for the query optimization. The word
“information integration” represents the objective of the proposed framework besides different semantics and
syntaxes through alias, feed forward and feed backward transformation functions. The objective is to
optimally use the existing relational data resources supplemented with additional relational or DWH
databases to provide end users with updated and correct information for knowledge capitalization.
163
The end users execute the SQL queries in search of data/information but these queries must be
validated against the evolved model for consistencies. The ‘synchronize data model’ function ensures that
upon any modification the end user applications are properly updated by application administrators. The
query optimization function extend its services to ‘validate query’ function where SQL query is reformulated
for (i) retrieval efficiency, (ii) exclusion of deleted fields/tables and (iii) inclusion of new fields if they are
not synchronized with end user applications. The ‘compute user statistics’ function computes and updates the
user statistics based upon the data queries executed by end users through respective applications. It is
important to note that based on ROMMII platform, customized end user data extraction applications are
suggested.
e) User_Statistics
(provides pre-
assessment on the
application failures, it
ensure compatibility.
164
6.7.3 Activity and Sequence Diagrams against Use Cases
In this section, the activity diagrams against uses cases are presented as (i) learn Meta model, (ii)
synchronize Meta model, (iii) query validation and optimization, (iv) log file parsing and users statistics
computations and (v) pre-failure assessments. The modify data model structure (model evolution) use cases
are based on the pre-failure assessment and learn Meta model use cases; hence, they are included in the
sequence diagrams but not in the activity diagrams.
Figure 6.14 – Learning Meta model for new database, table and/or fields
165
The learn Meta model activity starts with a choice for the database/data administrators and
redirects itself based on the choices made as the database, table or field. A list if attributes and paths for
each selected database is prepared. We loop through each database and verify its duplication from
existing Metabase. The databases with duplication are skipped and we proceed to the next database in
list. If the database is not in the Metabase, then the database and its version info is generated and added
to the Metabase. A list of tables in t heselected database is prepared and looped followed by the addition
of its version info into Metabase. At this stage, a list of all fields from the Metabase is generated
excluding the table being processed for the identification of potential missing links. We, then start by
first adding information of each field in the table being processed, into Metabase along with alias, local
and global similarity indices. If the similarity index computed on two fields is > 50%, then they are
suspected to have potential link. The data/database administrators make decisions about the inclusion of
this as a potential link, if accepted, the potential link info along with feed forward and/or feed backward
transformation functions are developed and added by the database administrators.We continue until all
fields, tables or databases are looped through and added into Metabase for query validation and
optimizations efforts. The role of each participant in the ROMMII platform is explained in their relevant
activity diagram.
166
case. The presented sequence diagram is compatible with the activity diagram and is self
explanatory. It is important to note that potentially identified links between databases and/or tables
are pointed by the system whereas their inclusion in the Metabase is always at the discretion of the
administrators. All potential links are further supported with the identification of feed forward and/or
feed backward transformation rules, which are programmed as internal database functions to
improve query performance.
i) Synchronize Meta Model
The “synchronize Meta model” is an important use case that extends its services to modify the data
models. It uses “pre-failure assessment” and “computes user statistics” use cases. The basic objective
of this use case is to manage and resolve all the inconsistencies that exist between the end user
applications used for data extraction and the Metabase. It happens when application administrators
fail to comply with the notifications from ROMMII for the application modifications. This use case
can be initiated by the application administrators to synchronize model modifications and ensure that
no failures are associated with respective user applications. The activity and sequence diagrams are
presented in the Figure 6.16 and 6.17 respectively.
In the Figure 6.16(a), database/data administrators are provided with an option to modify
table or field. The table option further requires the selection of modification type as add, delete or
modify the table. The modify table option is similar to the modify field option because in both cases
users can add, delete or modify the fields. The modify field or table options are presented in Figure
6.16(b). The table deletion is the most simple process where the table status is set to deleted in the
Metabase along with start and end valid times. The table is not deleted physically from the Metabase
which help us in SQL query optimization. This step requires similar deletion of corresponding alias,
domains, samples, potential links, relations and transformation rules. The add table option is simple
as it requires just new entries. It starts with the generation of table properties and its version info
which are then added to the Metbase. The “all-field-table” is generated from the Metabase which is
further used to find out potential links with the table fields being added into the Metabase. The table
fields are added to the Metabase with alias, potential links, relations and transformation rules. These
rules help us to remove data type and/or semantic inconsistencies.
The modify table option requires to check if the table exists in Metabase otherwise the next
table is selected. The modification in the table is classified as addition, deletion or modification of
field type. The addition and modification of field type requires deletion of their current versions
along with addition of new information and version. The new field is then checked against the
existing fields for potential links and relations. All these changes are recorded in the Metabase which
generates new database version. The actions performed are stored against each new database version
which can be easily explored and reversed to generate the previous database versions.
167
Figure 6.16(a) – Modification and synchronization activity diagram
168
Figure 6.16(b) – Modification and synchronization activity diagram
169
Figure 6.17 – Sequence diagram to modify and synchronize Meta model
ii) Query Validation and Optimization
This is one of the most important and key use case in the ROMMII platform where all the end users
queries for data extraction are first validated followed by its optimization. It is important to note that user
express their data needs through complex GUI (graphical user interfaces) where upon the user selections, a
complex SQL query is generated at the back end. This query is required to be validated to ensure that it
shall return valid results. If inconsistencies at this level are found then optimization algorithm reformulates
SQL query to avoid the application crashes. It provides an opportunity to move towards application agility;
however all the queries might not be optimized e.g. if the inner join or where clause criteria involves a
table or field which have been deleted or modified. The activity and sequence diagrams for this use case
170
are presented in the Figures 6.18 and 6.19 respectively. In the proposed ROMMII platform, the focus is on
data manipulation language (DML) which includes select, insert, update and delete SQL queries with
‘inner joins’ and ‘where’ clauses; however, it can be equally upgraded to include data definition (DDL)
and control (DCL) languages.
The primary objective of this use case is to validate and optimize query prior to execution. It can be
seen in Figure 6.17 that we start by parsing the SQL query in three lists as (i) where_fields, (ii)
inner_join_fields and (iii) select_fields lists. These lists hold the respective fields in the format of
‘database.table.field’. The tables from where_fields and inner_join_fields are checked in the latest database
and respective table versions. It is important to note that if ‘where’ and ‘inner join’ clauses have wrong or
deleted fields the execution of the SQL query shall result in a failure. If any of these tables are missing then
we do not proceed and inform the user about non executable SQL query; however, if all the tables are
traced back in the Metabase then list of all the fields from the select_fields list is prepared. These fields are
checked against the latest field versions in respective database; however if a field is missing it is traced
back in the older versions. The fields not found in the older databases versions are removed from the SQL
queries whereas traced fields are kept for “select” type SQL queries. The traced fields are removed for
insert/update and delete queries. The SQL query is updated with the fields, which were newly added in the
current version besides the fact that they were not selected in user defined queries. Finally the numbers of
fields in the validated and optimized SQL query are computed. If the field count is < 1 then user is
informed and query execution is aborted but for field count >=1, end users are provided with the requested
data. The optimization in this process is defined as steps (i) remove missing fields, (ii) keep fields not
found in current version but available in the older versions and (iii) add new fields in current version from
the previous version. It is referred as optimization because it ensures that the SQL query is executed even if
it includes previous version fields. It provides end users with additional fields which are recently added in
new versions but not selected by the end user for reference.
The sequence diagram for the query validation and optimization is presented in Figure 6.18. It is
initiated by the end-user where he selects the data as per his need and requirement through a GUI. Based
on this selection, the SQL query is automatically generated with “where” and “inner join” clauses, which
requires the validation and optimization before execution. This sequence diagram includes only two
objects (i) tables and (ii) fields, and is quite efficient in its execution; however, database and database
version objects can also be included if multiple database query is to be executed. In our case, it is assumed
that the users are working with single flexible DWH database. The execution flow is almost similar as
explained above.
The list of excluded fields from the SQL query are also reported back to the end user along with
the inclusion of new fields in the current version to ensure that end users are familiar with new type of data
for its potential use during analysis efforts. This activity diagram complements the “pre-failure assessment”
use case as it provides means to compute potential failure in terms of RPN (risk priority number) using
‘query risk levels’ object from the Metabase. The RPN is not used to prioritize the potential failures but for
the relative comparison of gains achieved through ROMMII platform.
171
Figure 6.18 – Query validation and optimization activity diagram
172
Figure 6.19 – Query validation and optimization sequence diagram
173
iii) Log file Parsing and Users Statistics Computations
The “log file” parsing and “user statistics computation” use cases generate knowledge that shall be
used during pre-failure assessments. The log files are generated by database server on 24 hours basis
using automated routine and accordingly user statistics are updated. These log files can be adjusted
by database/data administrators on weekly basis e.g. if it is changed to weekly basis the log files are
still generated on 24 hours being sequenced as 1 to 7 for the given week. The activity diagram for
these use cases is presented in the Figures 6.20; however, sequence diagram does not seem to
contribute or elaborate the working of proposed ROMMII platform so we have restricted ourselves
to the activity diagram only.
Figure 6.20 – Log File parsing and user statistics computation activity diagram
174
In the above presented activity diagram (Figure 6.21), we start by importing log files. The
distinct users and applications lists are computed by parsing log files one by one along with
database, table and field level frequencies. From this step, we follow two parallel flows as (i)
application level statistics and (ii) user level statistics. In application level statistics, we compute
database, table and field level frequencies followed by the computation of current application
probability (CAP). This step follows an average application probability (AAP) computation where
simple average of the two computed statistics is taken. These steps are repeated for each application
identified from the current log file being processed. The CAP and AAP are computed for user level
statistics against all distinct users identified from the log file. These computations are repeated until
all the log files are processed.
175
Figure 6.21 – Pre-failure assessment activity diagram
176
CA3
DM3 DM4
CA2
ROM2I2: DM2
It facilitates multi CA1 CA n
source data DM1 DM n
extraction and
dynamic changes in
the data model a) SPM (spatial
positioning model)
ROM2I2 MAM R&D Data Warehouse SPM
b) MAM (mapping and
alignment model)
Data Wrappers c) R&D WareHouse
(1year data)
Product Life Cycle,
Equipment Life
Cycle, OOC,Mask,
ODS RDB DWH DM
Metrology, EWS,
Defectivity , PT …
6.9 RESEARCH SCHEMATIC AND ADVANCEMENTS (ROMMII FRAMEWORK AND R&D DWH
DATA MODEL)
The research schematic and advancement are presented in the Figure 6.23. The ROMMII (SC3) platform and
R&D data model (SC4) models are the key generic contributions made in this thesis which enable us to
exploit huge data volumes and dimensions, and continuously evolve the existing data models for the
inclusion of new dimensions. The significance of these contributions is that it performs pre-failure
assessments on the existing single or multi-source data extraction and analyses utilities against potential
structural changes in the data models. It ensures that all the appropriate measures are taken prior affecting the
model evolution, resulting in the agility of existing utilities. In addition to this, it also ensures that all
structural changes are well communicated to potentially interested users based on the user and application
historical statistics. The ability to exploit huge data volumes and dimensions along with our proposed MAM
and SPM models shall ensure timely value extraction and knowledge capitalization, which is a key to
improve technology alignment and adoption lead times and costs.
177
Figure 6.23 - The research schematic and advancement with ROMMII platform and R&D DWH model
178
6.10 SUMMARY AND CONCLUSIONS
The R&D engineers have always been complaining about the availability of insufficient data and dimensions
to perform statistical analysis due to high storage and computational cost. The recent IT revolutions have
resulted in decreasing storage and computational costs at higher performance. It has ensured the availability
of huge data volumes and dimensions for R&D engineers, but they are still complaining that besides the
availability, now they are not able to exploit these huge data volumes due to inconsistent data models and
their unstructured evolutions. It is because of the fact that root cause analysis has shifted from the single-
source to multi-source analyses and if multiple data sources have ontology issues and missing common
identifiers then it is not possible to extract, map and align the multi-source data for effective root cause
analysis. The production databases have converged to data archiving due to increasing data volumes; hence,
they have varying data retention periods. It results in a big issue for the R&D engineer, because they need at
least one year’s data for the R&D analyses. The reason is that existing production data resources are equally
being used for the R&D as well as production management and engineering purposes. Moreover they are of
proprietary nature; hence, we do not have the right to alter or modify the data models. Any such effort
directly results in the failure of associated data extraction and analysis tools.
The proposed ROMMII platform shifts unstructured data model evolutions towards structured
evolutions and provides a pre-failure assessment upon any intended potential change in the data models. The
impact of potential change is computed at the end users and application administrator levels and they are
respectively intimated for potential changes in the applications to avoid failures and for information
purposes. The application administrators comply with the potential changes and evolutions in the model and
change the application to support extraction and analysis efforts to avoid potential failures. It also ensures
that new data dimensions are always available for the engineers and moreover they are intimated upon any
inclusion or exclusions of data dimensions. To avoid the data retention and proprietary issues, we have
proposed R&D DWH data model, which motivates customized application development for engineering
teams.
The proposed R&D DWH data model is a multi-dimensional data warehouse whereas the ROMMII
model is implemented as a relational database. The ROMMII platform is modeled and presented using the
UML and process flow diagrams. Its industrialization is highly specific to the type of database management
system being used for the R&D data model and existing databases. It leads to a new research area as well
that, how the agility provided by our proposed ROMMII and R&D data model can be extended, if R&D data
model comprises of different databases for the efficiency and performance.
The potential industrialization of proposed generic solutions as (i) MAM, (ii) SPM and (iii)
ROMMII model, is likely to result in increasing demand for metrology/inspection of engineering and R&D
lots. It has direct impact on the existing metrology/inspection tools capacities, which are dedicated for the
production lots. It is unlikely that we shall purchase new metrology equipments; hence, we need to spare the
tool capacities if we want smooth integration of the proposed generic solutions. In next chapter, we present a
yield aware sampling strategy, which spares the inspection capacities to be used for engineering and R&D
purposes.
179
180
Chapter 7: Yield Aware Sampling Strategy (YASS) for Tool Capacity
Optimization16
In the previous chapters, generic R&D solutions are proposed as the (i) MAM and SPM Models and (ii) ROMII
Platform and R&D DWH model. The potential industrialization of these proposed models in an IDM-fablite
business model shall result in additional requests for R&D lots metrology and inspection. It is highly likely that the
measurement capacities run out because production lots have the priority to use these capacities to ensure product
quality. In order to ensure that proposed solutions are industrialized, IDM-fablite business model must be provided
with the methodology to generate additional measurement capacitates. The existing metrology/inspection strategies
are being optimized using static, dynamic and smart sampling strategies based on the risk and associated delays. In
this chapter we have proposed a yield aware sampling strategy, which is objectively focused on finding and
inspecting the bad lots while moving the good lots to the next production steps. The equipment used in the SI are
highly sophisticated and hold three different key levels of information as (i) alarms, (ii) states and (iii) meters data.
This data provides key information about potential health of the equipment while it was processing production lot(s);
hence, this data could easily be used to learn the predictive models for subsequent predictions. These reliable
predictions in comparison with the static, dynamic and smart sampling strategies shall result in additional capacities
to be used for R&D purposes.
Contents
7.1 Introduction ......................................................................................................................................... 183
7.2 Metrology/Inspection and Production Tools Capacities Issue ............................................................ 183
7.2.1 Why we need 100% inspection? .......................................................................................... 183
7.2.2 Why additional capacities? ................................................................................................. 184
7.2.3 What is wrong with the Sampling Strategies? ..................................................................... 184
7.3 Proposed 3-Step Yield Aware Sampling Strategy (YASS) ................................................................. 185
7.3.1 Heuristic Algorithm for [PAM, PSM] models (Step-1) ....................................................... 186
7.3.2 Example for [PAM, PSM] prediction? ................................................................................ 187
7.3.3 Clustering and Priority queue allocation (Step-2 and Step-3) ............................................ 188
7.4 Data model to Support [PAM, PSM] Models...................................................................................... 190
7.5 Research Schematic and Advancements (YASS Strategy) ................................................................. 191
7.6 Summary and Conclusions .................................................................................................................. 193
16 Shahzad M.K., Chaillou Thomas, Hubac S., Siadat A., Tollenaere M., A yield aware sampling strategy for inspection tools capacity optimization,
International Conference on Artificial Intelligence, Las Vegas, USA 2012
181
182
7.1 INTRODUCTION
The proposed solutions as (i) MAM model, (ii) SPM Model, (iii) ROMMII Platform and (iv) R&D data model
exactly match with the needs of the R&D engineers. It provides an opportunity for our engineers to perform accurate
root cause analysis taking into account the newly emerging spatial variations. To best capitalize this opportunity,
additional dies/sites must be inspected in addition to the existing 9-17 sites metrology/inspection strategy. This
additional metrology requires more inspection tools or the available capacities that primarily serve to control
production lots in an e-IDM business model. As a result we have two options: (i) either purchase new
metrology/inspection tools or (ii) optimize the inspection capacities. It must be noted that the inspection/metrology
tools in an IDM primarily serve the production to control the product quality with 100% inspection at each step
hence, inspection tools quickly run out of capacities resulting in the production cycle delays. These equipments are
costly and add up fixed cost, which is not welcomed. To best utilize the available inspection capacities, we use
sampling strategies (static, dynamic and smart), which are based on the risk, delays and capacities. Industrialization
of such blind strategies is still a big question that might result in skipping bad lots to move to the next production
steps, resulting in waste of resources and customer dissatisfaction.
The industrialization and success of proposed generic R&D solutions depend on the additional inspection
capacities; hence, in order to generate these inspection capacities we move one step ahead and propose a yield aware
sampling strategy that predicts all production lots as good, bad or suspected lots. These predictions are made based
on the likely yield loss with predictive state (PSM) and alarm (PAM) models. The steps involved in this strategy are:
(i) classify potentially suspected lots, (ii) cluster and/or populate suspected lots in the priority queues and (iii) apply
Last in First Out (LIFO) to optimize capacities. It provides sufficient spare metrology/inspection capacities that can
be used for extended R&D purposes.
7.2 METROLOGY/INSPECTION AND PRODUCTION TOOLS CAPACITIES ISSUES
The SI has revolutionized our daily lives with electronic chips that can be found in almost all the equipments
around us and follows the slogan smaller, faster and cheaper driven by Moore’s law [Moore, 1998]. It
postulates that the number of transistors shall double in every 18 to 24 months at reduced cost and power.
Since then the SI has kept its pace as per Moore’s law by continuously investing in R&D for the new
technologies. New equipments are being manufactured to support and keep up with the emerging demands
and the pace defined by the Moore’s law. The equipments are highly expensive; hence, decisions to purchase
new production equipments are based on business strategy and estimated ROI (return on investment). The
metrology/inspection tools carry fixed costs and often phase out or need changes to cope up with new
technologies; hence, capacity optimization strategies are used to balance inspection load instead of
purchasing new metrology/inspection tools.
7.2.1 Why do we need 100% inspection?
The IC chip manufacturing has become a complex but expensive production process resulting in process
control challenges to find lots with yield issues before they consume the expensive production resources. An
electronic chip undergoes approximately 200 operations, 1100+ steps and 8 weeks of processing prior to
packaging and assembly. To ensure the product quality, metrology /inspection steps are added within the
manufacturing flow, almost after every manufacturing step. The design of an economical control for a
production process has always been an interesting research area that was initiated by [Duncan, 1956] but it
lacks robust results and requires a balance between controls and their costs. Hsu proposed a control plan with
the concept of skipping by decreasing control frequency as compared to a 100% inspection plan that
indirectly added risk of skipping bad lots to the next steps [Hsu, 1977]. Reynolds et al, (1988) presented, that
sampling size and frequency as two levels of controls are the better solution to quickly identify the issues
while minimizing the cost of errors [Reynolds et al., 1988]. Sanos introduced the concept of SPC in the
semiconductor industry that served as the basis for an updated control plan [Spanos, 1991]. Many approaches
have been proposed for an adaptive control: (i) sampling strategy based on the number of wafers passed on
metrology tools [Raaij and Verhallen, 2005], (ii) updating control plan based on process excursions [Mouli
183
and Scott, 2007] and (iii) updating control plan based on risk encountered during productions. Coledani and
Tolio have designed a buffer for control machines taking into account the quality and cycle time expectations
and it is the latest contribution in research regarding control plans [Colledani, 2008] and [Colledani and
Tolio, 2009].
7.2.2 Why additional capacities?
It is known that inspection tools have limited capacities; hence, optimal capacity utilization in a high product
mix is a key for success. There are some critical steps as well where the delay due to inspection capacity
limitation might have a strong impact on the next production step, so these priority products must be
inspected before other products in queue. Limited metrology/inspection tools capacities has a strong impact
on the production cycle times; hence, an efficient sampling and control strategy is required to optimize the
capacities and exploit economic benefits. In addition to this, there is also a strong need for the additional
metrology/inspection capacities based on the shift from MFD to local DFM efforts in our proposed extended
IDM (e-IDM) model. The objective is to capture newly emerging spatial variations and model the systematic
variations and drifts into rules and/or models for subsequent use during CAD simulations to assess yield and
manufacturability.
7.2.3 What is wrong with the sampling strategies?
The existing strategies are classified as static and dynamic where static sampling [Lee, 2002] selects the
same numbers of lots but dynamic sampling [Raaij and Verhallen, 2005] selects the number of lots for
inspection, based on the overall production. Smart sampling is a new approach that samples lots by taking
into account the risk associated with production tools, inspection tools capacities and delays to dynamically
minimize the wafers at risk [Dauzere-Peres et al., 2011and Sahnoun et al., 2011]. It is a better approach than
the static and dynamic strategies. In this strategy, if a lot in the waiting queue is controlled and it passes the
inspection step then all lots in the waiting queue, processed before this lot are removed with a confidence
that they are good lots. However none of them provide an evidence for a likely yield loss against sampled
lots resulting in skipping the suspected lots to move to the next process steps. We need a yield aware strategy
to classify good, bad and suspected lots to reduce the inspection load significantly followed by an
optimization strategy that exploits the production resources against limiting inspection tools capacities.
A generic production process is presented in Figure 7.1 where lots are processed, controlled and/or
skipped at the production and inspection tools to avoid bad lots moving to the next steps. The sampling and
scheduling strategies in this production process are focused on priority queues and lots are sampled based on
the risk levels associated with the product or process. These strategies are an effort to balance the difference
between production and inspection tools capacities. These strategies can be viewed as a blind strategy with a
high risk of skipping bad lots to the next steps. Our proposed YASS strategy empowers the control with PSM
and PAM models to filter good and bad lots followed by capacity optimization. We argue that the focus
should be shifted to find and control bad products rather than inspecting 100% product using blind inspection
strategies that do not differentiate between bad/good products.
R2R Feedback / Rework
184
It is evident that the inspection capacity allocation problem is linked with the process control plans in
the semiconductor industry. If mistakes are committed like skipping the bad lots to the next production steps
then the consequences are evident in terms of customer dissatisfaction and costs. It is due to the fact that
existing approaches do not provide any evidence of the likely yield loss. The additional capacities can be
spared from the production control if a reliable inspection strategy is adopted that avoids the inspection of
good lots. The existing sampling strategies do not guarantee a likely yield loss; hence, to ensure smooth
integration of the scientific contributions (SC1, SC2 and SC3) in the e-IDM model a 3-step yield aware
sampling strategy is presented that provides us with the additional inspection capacities. In this strategy, the
focus is on the identification of bad or suspected lots while moving good lots to the next production steps. It
shall not only increase the optimal utilization of the inspection capacities, but also provide us additional
capacities for the R&D purposes. Our proposed yield aware 3-step strategy uses the predictions made by the
PSM and PAM models. Based on the predictive output combinations, lots are either added to the priority
inspection queues or moved to the next production steps followed by the capacity optimization.
Insp. Eq-n
Lot9 Lot6
Equipment-i
Alarm-1
Lot8 Lot7
PSM
Scrap
SPC
Insp. Eq-2
[PSM, PAM]
Alarm-n
Lot4 Lot1
Queue P2
State-1
Operation-n Equipment-n Module-n
ASM
Insp. Eq-1
Lot3 Lot2
State-n
Queue P3
Next Step [Prod Capacity >> Insp. Capacity]
Figure 7.2 - Methodology with predictive state (PSM) and alarm (PAM) models
We have used only the states and alarms data; however, the module level meter data shall be
included in the future to compute the weighted probabilities to refine the state and alarm level predictions.
The first step in the inspection strategy is to classify the good, bad and/or suspected lots. In this step, we start
185
with the exploitation of the historical equipment states, alarms and SPC (statistical process control) data from
the process, maintenance and alarms databases to build predictive state [PSM] and alarm [PAM] models.
These models [PSM, PAM] are then used to classify the new production lots as good and/or bad lots and
generate four possible outputs: (i) [good, good], (ii) [good, bad], (iii) [bad, good] and (iv) [bad, bad].
The good production lots [good, good] are moved to the next production steps without metrology
and bad lots [bad, bad] undergo the 100% inspection and their results are used to update the prediction
[PAM] and [PSM] models. The suspected lots [good, bad] or [bad, good] are clustered (2nd step) based on
the equipment, product and recipe. It is followed by a priority queue allocation algorithm (3rd step) that
enters the suspected lot clusters into priority queues for further optimization based on LIFO (last in first out)
principle. It states that if a suspected lot defies the prediction upon inspection then all the lots in the same
cluster shall be subjected to 100% inspection otherwise, the cluster members are skipped. The predictive
state and alarm models [PSM, PAM] are updated with the feedback against all coherent and incoherent
predictions as good and/or bad examples. It provides us an intelligent way to reliably sample only bad or
potentially suspected lots followed by priority queuing and optimization for the economic benefits. The step-
1, where predictive [PSM, PAM] models are learned, is implemented with a heuristic algorithm as presented
in section 7.4. These models classify new production lots as good, bad or suspected lots. These PAM and
PSM models correspond to the step-1 of the proposed methodology. The success of this approach depends on
the accuracy of learning and classification algorithm for PSM and PAM models. We propose two heuristics,
first for the PSM and PAM models and second to cluster the suspected lots and applying the queue
optimization. A tuning parameter is also provided at the discretion of the user to control the PSM and PAM
prediction confidence levels.
Confusion (j)
Matrixes of [G, B, C] alarms or Global Support of alarms and
states data for all equipment E1.n states for equipment Ei and wafer
and Wafers W1.n Wj
Local Support for each set of Target defined by the user for
alarms and states in the Matrix for each equipment Ei to be used for
equipment Ei and wafer Wj model predictions
Matrix of alarms or states
Matrix for alarms and states,
sequence for equipment Ei and
historical data for equipment Ei
wafer Wj
186
Proposed algorithm for the prediction of [PSM, PAM] model is as under. We start with the
predictive model learning by computing good, bad and confusion matrixes of alarms and states sequence sets
for all equipments as shown in step-1. It shall be used during the computation of the local support. In Step-2,
the local support for each set of alarms and states sequence for a given wafer Wj is computed by counting the
similar set of alarms or states sequences in [G,B,C] matrixes. This count is divided by the respective count of
sequences in [G,B,C] matrixes to get the local support values for each sequence of alarms and states in wafer
Wj . Further global support for the wafer Wj is computed by multiplying the local supports computed for
each set of alarms and states sequence against [G,B,C] matrixes with the sum of computed local supports.
The results are summed at [G,B,C] level for the prediction by comparing it against the T[Ei] in step-4. If the
global support is smaller than the T[Ei] then the benefit of the doubt is given to the prediction by adding the
global support for [C]. Said algorithm for [PSM, PAM] models is presented as under:
Step-1: Compute for state and alarms data for each equipment Ei
Step-2: Compute local support for each set of alarms/states sequences for a given wafer Wj and equipment Ei
187
Alarm Matrix for Equipment Ei [Bad Yield]
T_ID Module1 Module2 Module3 Module4
1 107(0x6B) 384(0xA04)
2 384(0xF04) 84(0xC09)
3 388(0xA04) 988(0xA04)
4 374(0xA04)
5 14(0xC09) 384(0xC04) 24(0xC09)
6
7 384(0xF04) 312(0xD04)
8 84(0xC09)
9 384(0xF04) 172(0xD04) 38(0xF14)
10 34(0xF22)
Table 7.6 - Local and global support for wafer Wj [54%, Good]
7.3.3 Clustering and priority queue allocation [Step-2 and Step-3]
Based on the predictions from [PSM, PAM] models (section 3.1), we follow the 2nd and 3rd step in the
proposed yield aware inspection strategy where suspected lots are clustered and added to the priority queues
followed by LIFO optimization. It is presented with a simple flow chart (Fig.5). All production lots with
188
prediction combination [PSM’Good and PAM’Good] are simply skipped whereas other lots are populated in
the priority queues P1, P2 and P3 where P1>P2>P3. Lots with the combination [PSM’Bad] and PAM’Bad]
are first clustered based on the similarity of product, technology and recipe followed by the population of last
lot from each cluster in P1. If an inspected lot from the P1 validates the model prediction then its respective
cluster members are simply scrapped; otherwise, each member of the cluster is inspected and the predictive
[PSM, PAM] models are updated. In case of differences in the model predictions, lots are declared as
suspected and are populated in the priority queues P2 [PSM’Good and PAM’Bad] and P3 [PSM’Bad and
PAM’Good]. The lots from P2 are sequentially inspected; if it defies the models then all lots processed
before the inspected lot in P2 are given the benefit of the doubt and are skipped. If a lot inspected from the
P3 defies the model then all the respective cluster members are inspected and predictive [PSM, PAM]
models are updated for coherences and incoherencies.
It is evident from the above discussion that the [PAM] predictions have higher priority than the
[PSM] predictions based on the two facts, (i) child modules influence the statesof their parent modules,
hence prediction model developed at the module level might have a dual impact and (ii) alarms count and
duration result in the change of state of the modules. The states data is an aggregation of the alarms data at
module level; hence, alarms data provide more low level detailed information with no influence on the
alarms of parent modules. Based on these facts in this proposed methodology we have given higher priority
to PAM prediction while performing information fusion of modules alarms and states data. The [PSM, PAM]
prediction weights shall be defined in the future by including the meter data. The meter data is very critical
because the values of the meters initiate the preventive maintenance actions on modules. We believe that it
shall play a pivotal role in defining the prediction accuracies of the PAM model. The alarms and resulting
prediction shall get more weight when meter data falls within the distribution of previous preventive
maintenance actions.
Compute PSM, PAM Models Predictions
[PSM, PAM]
[Bad, Good]
[Good, Bad]
Check
Update PSM, Yes More
Scrap Lot Lots
PAM Models
No
No Check
Yes
More
Clusters
189
7.4 DATA MODEL TO SUPPORT [PAM, PSM] MODELS
The biggest challenge in building and deploying these PAM and PSM predictive models is the multi-source
data extraction, alignment and preprocessing from SPC, maintenance, process and alarms data sources. To
facilitate this, a data model (Fig.6) is proposed with the ASCM (Alarm and State Control Management) tool
that allows engineers a quick extraction and alignment of the data. In this data model, the equipment
(equipment class) is composed of modules (module class) and every module has a state (state_history class)
and alarms (module_alarms class) history. The usage meter data is also available (usagemeter class) but in
this paper this data has not been used. The parent-child associations between modules are captured by the
parent_child_relation class. A product (product class) is manufactured using multiple lots (lot class) but
follows a single process plan (process_plan class). The process plan has multiple operations
(process_operation) and each operation can have multiple steps (process_steps). The process step undergoes
different step runs (step_run class) as the production or metrology runs. The production step runs are
associated with the equipment that has the capability (equipment_capability class) to perform the process
steps. This data model is translated into a relational database, which is implemented using SQL server to
support the ASCM tool (IC5, Appendix F).
1
1
1 1 State_Model
Process_Plan 1 1 1..*
1 -State_Model_ID
-Process_Plan
1 Measure -Name
-Tech_Version
-Soft_Version -Measure_ID -Description 1 UsageMeter
1 -Description -Measure_Run_ID
-Usage_Meter_ID
-Technology -Process_Run_ID
-Module_ID
1 -Parameter 1 -Meter_Name
-Eq_ID 1..* -Units
-Lot_ID
-Meter_Value
-Wafer_ID
Model_to_States -Last_Reading
-Last_Reading_Date
Product 1 -State_Model_ID
-Last_PM_Value
-Name
-Product_ID 1 -Next_PM_Value
-State_Name
-Process_Plan
-Soft_Version Yield
-Product_Type
-Reticle_Set_ID 1 UsageMeter_History
-Recipe_Set_ID
0..*
1
1
0..* State_History
1 Out_of_Control -EQ_Module_ID
1..* -Name
-State_Name 0..*
1 -Start_Date
Lot
-End_Date
-Lot_ID -Previous_State_Name
-Product_ID 1 1
-Lot_Type
0..* Scrap
EQ_State_History Module_State_History
-Lot_ID
-Yield
190
7.5 RESEARCH SCHEMATIC AND ADVANCEMENTS (YASS STRATEGY)
The research schematic and advancement is presented in Figure 7.5. The YASS strategy is the key
contribution made in this chapter with an objective to generate additional measurement capacities. The
YASS is a 4-step methodology and it is supported by PSM and PAM predictive models. These models
predict each and every production lot as good, bad or suspected and based on proposed clustering and queue
optimization algorithms, we skip, inspect, scrap or partially control production lots. The skipped lots are the
good ones and this strategy is based on the simple principle that we should find and inspect the bad lots
whereas good lots must be pushed up to the next processing steps to reduce cycle times. In addition to cycle
time gains we get additional inspection capacities that shall be used to support the industrialization of our
proposed generic R&D models. The success of this strategy depends on the accuracy of PSM and PAM
models. These models are learned in the due course of production. We have full inspection load in the start
of the production because PSM and PAM models have minimum accuracy but as more and more production
lots are processed, we get accurate PSM and PAM models, that result in additional inspection capacities. The
granularity of the PSM and PAM models are still to be determined. At present we have hypothesized that
PSM and PAM models shall be learned at production step level for each equipment; however, in future the
similarity between production plans can also be used to cluster these PSM and PAM models.
191
Figure 7.5: The research schematic and advancement with YASS sampling strategy
192
7.6 SUMMARY AND CONCLUSIONS
The YASS strategy has significant advantages over existing static, dynamic and smart sampling strategies as it
identifies the production lots with likely yield loss; hence, good lots are skipped to next production steps resulting in
additional tools capacities. This strategy also provides a confidence and is directly linked with the alarms and states
data coming from the equipment. The alarms and states data describes a precise equipment condition and evolution;
hence, it could be easily linked with its impact on the yield. The advantage is not limited to additional capacities but
it also provides an optimized utilization of the inspection capacities. The T[Ei] (tuning) parameter provides a
control to the end users if they want to tighten the sampling or decrease the sampling rates in good periods.
It has one disadvantage of granularity of the prediction models, which we believe has to be at the product
and equipment levels. In the start these models are blank; however, they can be used for rough predictions based on
the equipment and process similarities. The product plays a significant role; hence, these models need to be adopted
for each new product. So in all cases to adopt these models we need to start with 100% inspection till the maturity of
the models. The long term benefit for the capacities optimization can be utilized by the R&D engineers. To extend
the proposed approach in future, we believe that equipment must be made capable with built-in modules to
automatically learn and manage these models at the product level and as soon as a product finishes a process step, it
is tagged for inspection or a skip.
193
194
Discussions and Conclusions
The semiconductor industry is the most fragile, fastest growing and most competitive manufacturing domain.
It is characterized by the cyclic demand patterns with a positive CAGR (+8.72) that guarantees cumulative
demand growth. It motivates us to continuously invest in R&D efforts so that we are ready to respond to the
growing market with new, fast and high value but low cost products. The SI, since its birth, has followed the
industrial slogan << smaller, faster and cheaper >> driven by Moore’s law that postulates doubling transistor
density every 18-24 months. This miniaturization trend has led the emergence of technical (design and
manufacturing interface) challenges that often result in serious manufacturability and yield level issues. It
has also resulted in the shift of business objectives from time-to-market and time-to-volume towards ramp-
up-rate, which can be achieved by introducing (i) new technologies every 2 to 3 years, (ii) technology
derivative(s) and/or (iii) current technology improvements.
The new technology is developed in an alliance to share R&D costs and model emerging design and
manufacturing complexities, supported with innovation in design, material, process and equipment. The cost
and lead times involved in a new technology are 5-7 B$ and 2-3 years respectively, as modeling design and
manufacturing issues are highly complex and has turned into a high cost R&D activity. The competitiveness,
however requires a quick technology alignment and/or adoption whether it is the new technology, technology
derivatives or improvements. The manufacturing data collected across the production line is the key towards
technology derivatives and improvement efforts success, besides new technologies. It requires dynamic and
effective production data exploitation so that the new technology derivatives and improvements can be
quickly aligned and adapted at reduced costs.
The design for manufacturing (DFM) is a well-known approach in the manufacturing domain where
manufacturing variations are transformed into rules and/or models for subsequent use during CAD
simulations to assess potential manufacturability and yield prior to prototyping or production. It helps us in
finding the first time correct design, which effectively reduces the lead times and avoids design respin costs.
It was adopted by SI in 1980 as a yield enhancement strategy that worked very well till 250nm technology,
but beyond this it has become ineffective. The ineffectiveness means that it is difficult to classify
manufacturing drifts and variations as random or systematic patterns followed by their transformation into
rules and/or models. There is nothing wrong with the DFM methods, rather, the issues lies with the quality of
input data, which is fed to these methods. So, if we can improve the quality of the data, we can put DFM
back on track to help us in reducing the lead times and costs associated with the technology derivatives and
improvement initiatives for the competitive advantage.
The above discussion highlights the need for literature review across three domains as (i) the SI
business model evolutions and challenges, (ii) role of DFM methods in the success of SI and (iii) information
integration needs to enable dynamic exploitation of the manufacturing data collected across the production
line. We found transformation of traditional IDM business model into fabless and fablite models to capture
maximum market share in the SI market demand with a positive CAGR. This transformation is truly in line
with shifts in business objectives to address the growing market needs but recent shift towards the ramp-up-
rate needs its realignment. The miniaturization driven by Moore’s law has resulted in the emergence of new
design/manufacturing interface complexities. The spatial variation on the wafer surface is a recent
phenomenon, which is critical to the ineffectiveness of the DFM methods. The data collected across the
production line is categorized as measurement data (PT, Inline, EWS and Defectivity, etc.) and contextual
data (WIP, product/equipment life cycle, SPC, etc.). The existing DFM methods are based on the single
source analyses, which are insufficient to capture the spatial variations; hence, the need to shift from single
to multi-source data analyses to improve DFM effectiveness. It is to be further strengthened by moving from
the site level analyses towards die and test structure position based analyses as well. The wafer orientation
due to test structure positions and measurement reference frames often results in varying site/die level
coordinates, which requires a generic solution because engineers spend huge amount of time in mapping and
aligning site/die level data. It often results in bad alignment, which adds to the DFM ineffectiveness. The
195
DFM methods cannot be improved in the present emerging spatial variations until the shift from single to
multi-source data analysis at site, die and test structure position based analyses. The methodology for data
mapping and alignment cannot work alone until it is supported with the data. The recent revolutions in IT
have enabled low cost storage and exploitation of huge data volumes. It provided the opportunity to
continuously restructure the data models so that new data dimensions can be added to be further used during
analyses.
This thesis is objectively focused in putting DFM back on track. There has been a transformation of
SI business models and shift in business objectives based on market dynamics; hence, the need to first
analyze the existing business models against a recent shift towards ramp-up-rate. The objective is to propose
an extension in the existing business models to support the new objective as well as a methodology to
synchronize the business models with the future shifts in business objectives. The SCAN analysis is selected
as the methodology to identify the top rank objectives followed by its alignment with SI business model
based on extension or modification. The SCAN analysis highlighted the leadership position and quick ramp-
up-rate as the top ranked objectives. The SWOT analysis resulted in the adoption of strategies to mitigate
weaknesses and exploit the opportunities. The best known IDM-fablite business model is bench marked to
assess its potential success against the shift in business objectives and we identified (i) fast technology
transfer, (ii) manufacturing databases and (iii) effective root cause analysis (R&D) as the key improvement
areas for its compliance with recent shift in business objectives along with strategy to mitigate weaknesses.
The analysis was further extended to technology development process, which highlighted the (i) data
extraction, (ii) alignment and (iii) pre-processing due to ontology issues and (iv) missing database links as
the key weaknesses for mitigation. These weaknesses are classified as the main reasons for DFM
ineffectiveness based on the fact that due to these challenges R&D engineers are not able to exploit
manufacturing data sources besides huge data volumes and dimensions. The proposed e-IDM fablite business
model is an extension of existing IDM-fablite model where these weaknesses are removed and engineers can
potentially exploit the production resources to perform multi-source site/die/position based root cause
analyses.
Compliance with the proposed e-IDM business model necessitated finding root causes against
respective identified weaknesses (failure modes) grouped as (i) ineffective root-cause analysis (infield and
scribe line test structure positions) and (ii) data extraction, mapping and alignment. We have proposed and
used i-FMEA approach to identify the cyclic root causes against these failure modes. The proposed 4-step
methodology helps identify the cyclic root causes that result in failure and need generic R&D solutions
instead of operational fixes as proposed by traditional FMEA approach. The root causes clearly demonstrate
the inability of R&D engineers to exploit huge volumes of data and allow the inclusion of new data
dimensions in existing data models. The proposed i-FMEA approach helps in tracing back the potential root
causes in other business functions. The key cyclic root causes identified are the (i) data format issues, (ii)
unstructured evolution of data models and (iii) missing database links and dimensions that need generic
R&D fixes. The SCAN and i-FMEA methodologies are suggested to be used upon the detection of any
change in the business environment to quickly align the business model and find weaknesses along with
respective root causes for, which generic R&D solutions are critical.
The effective root cause analysis is the key to improve the existing ineffectiveness in the DFM
methods and can be defined as the ability to analyze the data to find answers against drifts and/or variations.
These drifts and variations are further classified as systematic or random for their transformation into rules
and/or models. At present, engineers are facing huge difficulties in data mapping and alignment to perform
multi-source analyses. It is based on the fact that measurements result in varying coordinate systems due to
test structure orientation on the wafer. The wafer is rotated prior to the metrology steps and its position is
monitored using notch or flat positions. The varying coordinates issue is further complicated with the
different metrology reference frames. These issues have a significant impact on stack definition, model
correction and validation steps during the interconnect modeling process.
196
The proposed generic R&D solutions include MAM and SPM models, which take into account the
information from the wafer, its mask and test structure positions, and normalize the site (PT and Inline) and
die (EWS and Defectivity) level measurements. The MAM model provides fixes to sites and dies level
measurements so that they can be used for the correlation purposes. The SPM model provides ability to
correlate parameters based on the position of the respective test structures whereas proposed die/site
qualification model enables R&D engineers to perform site/die level root cause analyses. The above generic
models remove all inconsistencies with the measurement coordinates so that engineers could perform multi-
source root causes analysis. Recent IT technologies have enabled huge data volumes storage and
exploitation. The engineers need multi-source data coming from multiple data sources for effective root
cause analysis; hence, the biggest challenge faced at this level is to allow restructuring of data models.
The SPM model is highly efficient; however, there is still room for improvement to obviate
computing the positions w.r.t. each wafer for all the test structures. The structural position map can be
computed and saved as a search pattern, which can be used for mapping and alignment of measurement data
prior to data analyses. We can apply a similar search pattern based idea for the MAM model based on the
fact that wafer position and measurement reference frame have discrete values, which can be used as a
reference for mapping and alignment in an effort to reduce the computational costs and efforts.
The R&D engineers have always been complaining about the availability of insufficient data
volumes and dimensions to perform statistical analysis due to storage and computational cost issues. The
recent IT revolutions have resulted in decreasing storage and computational costs at higher performance. It
has enabled the availability of huge data volumes and dimensions to the R&D engineers, but they are still
complaining that besides the availability, they are not able to exploit huge data volumes and dimensions due
to unstructured data model evolutions and inconsistent data models. It is because of the fact that root cause
analysis has shifted from single-source analysis to multi-source analysis. If multiple data sources have
ontology issues and missing common identifiers then it is not possible to extract multi-source data for the
root cause analysis. The production databases have converged to data archiving due to increasing data
volumes; hence, they have different retention periods. It causes a big issue for the R&D engineer because
they need at least one year’s data available all the time for the R&D purposes. The reason is that existing
production data resources are equally being used for the R&D and moreover they are of proprietary nature;
hence, the inability or right to alter or modify the model. Any such effort directly results in the failure of
associated data extraction and analysis tools, which is not less than a disaster.
The proposed ROMMII platform shifts unstructured model evolutions towards structured evolutions
and provides a pre-failure assessment upon any intended potential change in the data model. The impact of
potential change is computed at end users and application administrator levels and they are respectively
intimated for potential changes in the applications to avoid failures and for information purposes. The
application administrators comply with potential change, and evolution in the model is ensured to support the
extraction and analysis utilities to avoid failures. It also ensures that all new data dimensions are available at
all times for the R&D engineers and moreover, they are intimated of any inclusion or removal of data
dimension from the R&D data sources. To avoid the data retention period and proprietary issues, we have
proposed an R&D data model, which shall result in the customized application development for each R&D
group. It shall be managed and controlled by ROMMII framework where we are able to manage the data
dimensions.
The R&D data model is a multi-dimensional data warehouse whereas the ROMMII model is
implemented as a relational database. The ROMMII platform is modeled and presented using the UML and
process flow diagram. A simple database with three tables is used to present the potential results; however,
its industrialization is highly specific to the type of database management system being used for the R&D
data model; hence, its industrialization requires the brainstorming sessions on the choice of DMS first. It
leads to a new research area as well, that how the agility provided by proposed ROMMII platform can be
extended if R&D data model comprises of different databases.
197
In this thesis we have demonstrated that the existing IDM-fablite business model does not support
the recent shift in the business objective i.e. ramp-up-rate; hence, we have proposed the e-IDM fablite
business model to fully comply with the new business objectives. The root causes against weaknesses found
during the business model analysis are identified using our proposed i-FMEA methodology. The generic
R&D solutions as the MAM, SPM, ROMMII and R&D DWH model are proposed instead of operational
fixes. The industrialization of these solutions shall result in extended requests from the R&D engineers for
additional measurements at site and/or die levels. It shall challenge the production and inspection capacities
of an IDM. The proposed solutions can only be industrialized if they do not add additional fixed cost; hence,
we have further analyzed the inspection capacities and existing strategies to assess if we could spare
inspection capacities for R&D purposes. The existing static, dynamic and smart sampling strategies for the
inspection are blind strategies and are based on the risk, delay and capacities. The proposed 3-step YASS
strategy uses PAM and PSM prediction algorithms based on alarms and states data. The alarms and states
patterns likely predict the potential bad or suspected lots and reduce inspection load. The additional
capacities generated can be used for R&D purposes to support the smooth industrialization of proposed
scientific contributions.
It is highly important to note that in the present organizational structure the role of IT is limited to
database management where control and administration of these resources are outsourced. The proprietary
nature of database resources further complicates the issue with regards to data models restructuring to add
new dimensions. The proposed ROMMII framework suggests the need for the role of data administrator to
be created in parallel to IT. The administrator must be responsible for all data sources in the organization
along with data dictionaries and any potential change must pass through this role to ensure consistency of
concepts and new data dimension along with naming standards and description. The proposed solutions
suggest the development of customized data extraction and analysis utilities for different teams in order to
improve the productivity of R&D engineers. These customized applications must be updated upon the
request of the end user for additional data analysis capabilities. These utilities must be able to use the data
analysis algorithms developed and tested by the local R&D teams. The proposed solutions in this thesis are
critical for the IDM-fablite business models to maintain their leadership position by capturing maximum
market share, which can be achieved by putting DFM back on track. It requires enhancing the productivity of
R&D engineers by enabling them to use and exploit the huge production data volumes for effective root
cause analyses.
198
Appendix A: List of Publications
We have published scientific contributions in the well known international conferences and
journals. The count for these publications includes Journal (1), Conference (7) and others (2). In this
section we present the list of the articles published/accepted/submitted along with the abstract for
the readers’ interest.
A.1 Journal Publications:
1. Shahzad M.K., Siadat A., Tollenaere M. and Hubac S. (2012), Towards more effective DFM
methods in Semiconductor Industry (SI), International Journal of Production Research, (submitted)
TPRS-2012-IJPR-0638
Abstract: The DFM methods are used during the technology alignment and adoption processes in
the semiconductor industry (SI) for manufacturability and yield assessments. These methods have
worked well till 250nm technology for the transformation of systematic variations into rules and/or
models based on single-source analysis but afterwards they have become ineffective R&D efforts.
The site/die level mismatches due to metrology reference frames and test structure orientations are
the main causes for this ineffectiveness. It restricts us from modeling newly emerging spatial
variation resulting from miniaturization that often results in increasing the yield losses and lead
times. The purpose of this paper is to improve the DFM effectiveness; hence we present a generic
coordinates mapping, alignment and qualification model to remove the site/die level mismatches. It
provides us an accurate computation of the physical dimensions followed by correlation against the
electrical and inspection data at site/die levels. This ability shall effectively transform the newly
emerging spatial variations into DFM rules and/or models, based on the multi-source root cause
analysis. The presented model is further integrated in BEOL (back-end-of-line) interconnect
modeling and an extended BEOL (e-BEOL) process is proposed for the yield and lead time
improvements based on effective DFM methods.
Keywords: design for manufacturing (DFM); coordinates mapping and alignment; BEOL
interconnect modeling and root cause analysis
Abstract: Semiconductor manufacturing industry (SMI) has shifted from an IDM (integrated device
manufacturer) to a fabless structure where technology is developed in an alliance to share high R&D
costs and address time to market and time to volume challenges. In this fabless structure, electronic
design automation has emerged as a key stake holder to model increasing design and manufacturing
interface complexities and its integration within design flow, but collaboration within alliances have
resulted information sharing and technology transfer as the key challenges. We argue that IDM
model is superior to a fabless structure due to its inherent ability for faster/superior knowledge
capitalization. We benchmarked and analyzed a world reputed IDM with use-case and SWOT
(strength, weakness, opportunity, threat) analyses to identify the limiting factors that led this
transformation and found data and statistics as the core issues. We have proposed an extended IDM
business model where engineering information systems (EIS) are tuned for design for
manufacturability (DFM) compliance to achieve time to quality (time to volume, time to market) and
yield ramp up rate at low cost but effective R&D efforts.
199
Keywords: SMI business model; time-to-market (T2M); time-to-volume (T2V); design for
manufacturing (DFM); yield ramp-up rate
2. Shahzad M.K., Tollenaere M., Hubac S., Siadat A., Extension des Méthodes DFM pour
l’industrialisation de produits microélectroniques, 9e Congrès Internationale de Génie Industriel
Montréal, 2011, Canada
Abstract : Semiconductor manufacturing industry (SMI) is characterized by the fastest change in
smallest period of time; hence to address time-to-market and time-to-volume challenges, DFM was
included in design flow (1980) as a yield enhancement strategy. It has become an industrial standard
to assess yield/manufacturability of the design. Test chip is used to validate the geometric stack
against resulted specs and models are frozen and distributed to the CAD department for inclusion in
design and DFM kits. This paper proposes a DFM methodology to include geometric measurements,
which could impact significantly electrical test results making it difficult to adapt the target models.
It requires site to site mapping on the wafer, which is not trivial because wafer center is different
than the mask center and the test structures PCS/M (process control/monitoring structure) could be
present in the horizontal or vertical scribe lines. A case study on the interconnect modeling is
performed in a top ranked SMI and an extended methodology to rapidly align local interconnect
models on target the source specs is proposed. BPR (business process reengineering) and IDEF0 are
used for analysis and newly proposed methodology along with a data model, which is implemented
in a tool for R&D engineers.
Keywords: design for manufacturing (DFM), manufacturing for design (MFD), time to market,
(T2M), time to volume (T2V) and time to quality (T2Q), engineering data analysis (EDA)
3. Shahzad M.K., Hubac S., Siadat A., Tollenaere M., An Interdisciplinary FMEA methodology to
find true DFM challenges, 12th European APCM Conference, Grenoble France, 2012
Abstract: The FMEA (failure mode effect analysis) is a widely used and well known approach for
the concept, design and process improvements; however it is limited by the expert’s knowledge and
its scope. In this article we have used the FMEA approach to identify and remove root causes from
the DFM (design for manufacturing) inefficiencies for faster technology improvements and
derivatives, resulting in a quick ramp-up-rate. We propose a 4-step interdisciplinary FMEA (i-
FMEA) methodology to find root causes across the business functions based on the fact that
performance of a function depends on inputs, which are outputs from other functions. A case study is
conducted in a reputed IDM (integrated device manufacturer) and results are the causes that we have
never considered as the limiting factors.
Keywords: deisgn for manufacturing (DFM) challenges, effective root cause analysis, FMEA
approach
4. Shahzad M.K., Hubac S., Siadat A., Tollenaere M., MAM (mapping and alignment model) for
inspection data in semiconductor industry, 12th European APCM Conference, Grenoble France,
2012
Abstract: Increasing model to hardware gaps has turned technology development process into a
high cost R&D activity, hence a new technology is developed in an alliance which is transferred,
aligned and adapted for every product. Our engineers are focused on reducing the technology
adaption and alignment lead times based on an efficient and effective root cause analysis but they
spend significant amount of time in data extraction, mapping and alignment because available
inspection data vary in format and coordinate system depending on tool and vendor. Today we have
a huge volume of data in multiple dimensions but database issues (Shahzad et. al, 2011) limit our
capabilities leading to an opportunity loss. In this article we present MAM (mapping and alignment
model) for inspection data to ensure site to site mapping between PT and Inline data, die to die
200
mapping and alignment between EWS and defectivity data and die to site qualification between
PT/Inline and EWS/Defectivity data. It empowers our engineers to quickly find the root causes,
classify them as systematic or random and transform them into rules and models for the faster ramp-
up-rate.
Keywords: coordinates mapping and alignment, effective root cause analysis, knowledge
capitalization
5. Shahzad M.K., Hubac S., Siadat A., Tollenaere M., ROMMII (referential ontology meta model for
information integration) Architecture for Dynamic Restructuring of DWH models, 12th European
APCM Conference, Grenoble France, 2012
Abstract: Semiconductor industry (SI) is facing difficulties in data/information integration while
deploying DFM (design for manufacturing) methods for an efficient root cause analysis. Recent IT
developments have resulted in the availability of huge volume of data across multiple dimensions,
however we are still unable to fully exploit data for a knowledge discovery. The database
technologies have addressed platform heterogeneity and efficiency BUT unstructured data model
evolution, ontology issue, missing links and the risk of failure of a single source and/or multi source
data analysis tool are the key limitations towards an efficient root cause analysis. In this article we
present a software framework supported with ROMMII architecture to address these key challenges
and ensure dynamic restructuring of the data model against varying needs to ensure the availability
and access of every single data element without any risk of the tool failures.
Keywords: data model evolution, software agility, multi-source root cause analysis
6. Shahzad M.K., Hubac S., Siadat A., Tollenaere M., SPM (spatial positioning model) model to
improve DFM methods, 12th European APCM Conference, Grenoble France, 2012
Abstract: Scribe line and infield test structures are used to model the design and manufacturing
interface complexities resulting from technology scaling, manufacturing and fundamental limitations
(Duane et. al, 1997 and Duane and James, 1996). The biggest challenge in accurately capturing
these model to hardware gaps is our inability to accurately map, align and position the inspection
data (PT, Inline, EWS, Defectivity) collected at die and field levels. The test structures are assumed
to be the true representative of the products and are physically located at different positions, hence
mapping and alignment do not capture the spatial variations which could answer even the random
drifts. In this article we focus the challenge faced by the engineers in implementing DFM with the
spatial variation based on test structures positions and the shortest distance between test structures
measuring inspection parameters. We propose the SPM model that computes the test structure
positions across the wafer and our intelligent algorithm find the target test structure in closest
vicinity for spatial correlations. SPM model shall directly impact the productivity of engineers and
empower engineers for an effective multi source correlation.
Keywords: spatial variation analysis, yield management, design for manufacturing (DFM)
7. Shahzad M.K., Thomas C., Hubac S., Siadat A., Tollenaere M., A yield aware sampling strategy
for inspection tools capacity optimization, International Conference on Artificial Intelligence, USA
2012
Abstract: The product quality in semiconductor manufacturing is ensured with 100% inspection at
each process step; hence inspection tools quickly run out of capacities resulting in the production
cycle delays. To best utilize the production and inspection capacities, existing sampling (static,
dynamic and smart) strategies are based on the risk and delays. These strategies, however do not
guarantee a reliable lot sample that represents a likely yield loss and there is a high risk of moving a
201
bad production lot to next production steps. We present a 3-step yield aware sampling strategy to
optimize inspection capacities based on the likely yield loss with the predictive state (PSM) and
alarm (PAM) models as: (i) classify potentially suspected lots, (ii) cluster and/or populate suspected
lots in the priority queues and (iii) apply last in first out (LIFO) to optimize capacities. This strategy
is implemented with two heuristics. We also present a data model with ASCM (Alarm and State
Control Management) tool for the multisource data extraction, alignment and pre-processing to
support the validation of [PSM, PAM] predictive models.
Abstract: Supply chain partners strive hard for operational business excellence, enhanced
integrated value chain and sustainable competitive advantage under mass-customization/
globalization challenges. In this paper new notions, GBOP (generic bill-of-product: set of product
family variants), GBOP/GSCS (generic supply chain structure) interface and GBOP architectural
constraints have been introduced that shall empower supply chain the flexibility to rapidly
reconfigure under business environmental dynamism and quickly respond to the varying customer
needs with economies of scope. Further a mathematical model is proposed to investigate the
influence of GBOP on supply chain configuration, relationship between GBOP and GSCS
architectures, optimal redefinition of GBOP/GSCS and decisions related to opening or closing of
market segments under cost minimization and profit maximization objectives.
Keywords: supply chain management (SCM); integer linear programming; quantitative modeling
of facility design
2. Shahzad M.K. and Hadj-Hamou K., Integrated supply chain and product family architecture
under highly customized demand, Journal of Intelligent Manufacturing DOI: 10.1007/s10845-012-
0630-0.
Abstract : Mass customization efforts are challenged by an unpredictable growth or shrink in the
market segments and shortened product life cycles which result in an opportunity loss and reduced
profitability; hence we propose a concept of sustainable mass customization to address these
challenges where an economically infeasible product for a market segment is replaced by an
alternative superior product variant nearly at the cost of mass production. This concept provides
sufficient time to restructure the product family architecture for the inclusion of a new innovative
product variant while fulfilling the market segments with the customer delight and an extended
profitability. To implement the concept of sustainable mass customization we have proposed the
notions of generic-bill-Of-products (GBOP: list of product variants agreed for the market segments),
its interface with generic-supply-chain-structure and strategic decisions about opening or closing of a
market segment as an optimization MILP (mixed integer linear program) model including logistics
and GBOP constraints. Model is tested with the varying market segments demands, sales prices and
production costs against 1 to 40 market segments. Simulation results provide us an optimum GBOP,
its respective segments and decisions on the opening or closing of the market segments to sustain
mass customization efforts.
202
Appendix B: Semiconductor Design, Mask and Manufacturing Processes
We briefly introduce the design, mask and manufacturing process flows followed by key challenges and associated
manufacturability and yield limiting phenomenon. It is important because the DFM methods are focused on
modeling and transforming these challenges into rules and/or models. The resulting solutions are also discussed to
provide a better understanding about the design rules (DR), DFM rules and models which are used during the CAD
simulations to pre-assess the manufacturability and yield. In order to provide more clarity and in depth understanding
to the reader, we have taken an example of a simple CMOS inverter and simulated the design, mask and
manufacturing steps. This example is presented in Appendix-C; however the readers with good understanding on
these processes can continue to chapter-2, section 2.2.
203
Figure B.1 – Bipolar and FET transistors structure [Quirk and Serda, 2000]
MODULE
GATE
CIRCUIT
DEVICE
G
S D
n+ n+
It is important to note that these steps are supported by CAD tools from different companies e.g.
Cadence, Mentor, Synopsys etc. These tools are plugged with SPICE Models, design, process and DFM kits.
The netlist is simulated using SPICE models to characterize the integrated circuits against defined timing,
power and delay indicators. The design goes through the DRC simulation and all violations are either
exempted or removed by the designer by optimizing the circuit layout. The critical area (CAA) and Hotspot
analyses are performed on the circuits’ netlist using DFM rules and models to find potential
manufacturability and yield issues. These kits are developed and maintained for each technology but as we
know that a new technology is developed in a technology alliance, hence these kits are defined and managed
at the technology level.
The new technologies are designed and developed in alliances using test products because it is not
possible to test all potentially possible designs; hence these kits are developed based on the data analysis
from these test products. We also know that the same customer specifications can be met with multiple
designs so every new product designed and developed using a given technology might result in drifts,
variations and model to hardware gaps. If we properly analyze these variations and transform them into rules
and/or models then it shall result in our ability to continuously update the design and DFM kits. In the
normal practice, the new technology from the alliance is transformed to the alliance partners’ business
models where it is declared as frozen and changes are not authorized (section 1.3, Figure 1.10). The reason is
that the general purpose and standard cell libraries are developed and qualified for each technology, if this
improvement link is authorized then all these existing design libraries are again qualified and validated,
205
which is expensive and time consuming. It might result in significant delays. The designers are willing to
accept any potential changes recommended through local DFM efforts if they are supported with the gains in
parametric and functional yields.
More often the process integration teams come up with the changes in the design kits based on the
analysis of the parametric data collected from production line. The proposed changes are not supported
because they are not supported with the gains in parametric or functional yields. We need to provide our
engineers the ability to support their proposed changes in the design and DFM kits along with respective
yield improvements or degradation results. It shall extend our knowledge capitalization ability to support the
technology alignment and adoption improvement efforts.
206
filled by chip assembly and RET is performed to improve litho printability and device is verified with Litho
rule (LRC) and mask rule (MRC) checks [Luo, 2010].
207
Figure B.8 - mask manufacturing defects
The mask data preparation is the last step where we can put our efforts to improve the design for
manufacturability and yield issues. This step if not properly addressed shall result in worst yield issues. The
no of masks to be used depends and are defined by the technology being used; hence the rules are applied on
both integrated circuit designs (devices and interconnects) and scribe frames. As we know that the biggest
challenge faced today is the sub wavelength manufacturing so a mask provides us a safeguard against those
yields limiting factors by ensuring the devices against litho and mask rule checks. The SRAF features are
special features, which are added to the hot spots to avoid the effects during lithography, etching or CMP
operations. The reasons to apply OPC due to lithography issues are presented in the Figure B.9a. It is evident
that due to sub wavelength lithography we are not able to exactly fabricate the shapes as described in design
layout; hence OPC rules must be defined to ensure a clean design layout (Figure B.9b). The OPC based
layout is presented in Figure B.9c where bleu shapes are after etch targets, grey shapes are after litho target
and green correspond to actual mask layout [Chiang, 2010 and Orshansky, 2008 and Wong, 2008].
208
120
Count 60
40
20
0
C120 IMG175 IMG140 C110 C090 C065 C055 C045
Technology
209
electrons if accumulated towards the gate cannot jump into gate because of the oxide shield, however depletion
region is extended that ultimately allows the flow of electrons.
The npn MOSFET is presented in Figure B.12 with its two operation modes. The npn transistor shall work
as an electronic switch if we are able to establish a conductive path between the source and drain under the gate. A
battery (3v) is connected to a small lamp with external wires. If there is no voltage applied on the gate then there is
no flow of the current between source and drain resulting in the open gate situation and the bulb remains off,
however a small voltage at the gate is all what is need to turn it on. The gate to source voltage (Vgs) is applied on the
gate that creates an electrical field, which pulls electrons from the source, drain and substrate. These electrons get
collected underneath the gate oxide a.k.a. the depletion region. This depletion region gets extended as Vgs is
increased and when it reaches the threshold voltage (Vth) the depletion regions extends to the source and drain
resulting in the flow of drain to source current (Ids). The flow of current from drain to source results in the completion
if the circuit and the lamp is turned on. The depletion region in npn transistor is referred as n-channel resulting in
drain to source current flow.
Figure B.12 – npn MOSFET structure and current flow [Quirk and Serda, 2000]
The structural and operational mode for the pnp MOSFET is presented in Figure B.13. The open gate
situation prevails until no voltage is applied on the gate and there is no current flow between drain and source. When
we apply a negative voltage on the gate then electrons are pushed back from the gate oxide resulting in the
accumulation of holes underneath the oxide. As soon as the applied voltage reaches the threshold level, the depletion
region gets extended towards drain and source resulting in flow of current from source to drain. This depletion region
is also known as the p-channel and current flow is attributed to the flow of positive charges.
Figure B.13 – pnp MOSFET structure and current flow [Quirk and Serda, 2000]
210
wafer lapping, etching and polishing operations. It follows cleaning of the surface, inspection and
packaging of the wafers for shipment to the SI [Hwaiyu, 2005and Quirk and Serda, 2000 and
Rakesh, 2008].
Wafer Lapping
Crystal Growth and Edge Grind Cleaning
b) Fabrication of the integrated circuit: The major tasks used in the manufacturing of a transistor
are presented in the table B.1 [Michael, 2004].
oxygen This is the first step in the IC manufacturing where an oxide layer
Thik Films
UV light The exposure step transfers the pattern from the mask to photo
resists coating on wafer surface. The mask patterns are 5x or 4x
Mask larger than features being manufactured. These features are
optically shrunk before reaching the surface of wafer. The mask
and misalignment errors are the key challenges faced by the
Photo Lithography
211
oxygen These gate oxidation and poly silicon depositions (thin films) are
gate oxide very critical operations a.k.a. thin films; hence it results in the
formation of gate oxide that acts as a dielectric between gate and
Thin Films h) Oxidation (Gate oxide) wafer substrate. It plays an important role in the formation of
Dopant gas channel between source and drain regions. The poly silicon results
Silane gas in gate formation and is used to switch the transistor on and off.
polysilicon
The most important challenge faced today is the uniformity while
reducing the thickness of these thin films.
i) Poly silicon Deposition
This step is used to transfer the mask pattern for the gate with
Photo and Etch
R&D engineers.
j) Polysilicon Mask and Etch
Scanning
ion beam
The ion implantation step is used to change the substrate so that
they can act as n-type or p-type source and drain regions. It is
alternative to diffusion. In diffusion, dopant atoms are moved from
surface into silicon substrate by thermal means and in ion
Diffusion
G
ox
S D implantation they are forcefully moved into the surface with
k) Ion Implantation ionized beams.
The active regions are those regions on the surface of silicon
G substrate where poly gates are formed along with sources and
S D
drains. The operations used for the active regions are called front-
l) Active Regions end-of-line (FEOL) operations.
silicon nitride By this time the transistor is ready and we move towards
Thin Film
top nitride
metallization process (inter connections) a.k.a. back-end-of-line
S
G
D
(BEOL) operations. This film is deposited to save the interaction
between two layers of operations (FEOL and BEOL).
m) Nitride Deposition
Contact
holes
This step is the start of the BEOL operations and first step
involves creating contacts to the source, drain and the gate so that
potential can be applied at the transistor level to turn it act like a
Photo and Etch
Metal
contacts
These steps involve repeating the photo and etch operations to
construct vias and metal lines starting from the contact (BEOL).
The number of metal lines may vary depending on technology,
drain
S
G
D
however reduction in the geometric specifications are serious
o) Metal Deposition challenges against potential defects.
and Etch
212
Cycle Time per Operation Misprocessing
12
Photo Ion Implant Diffusion
9 3 Production Bay Production Bay Production Bay
6
Rework Scrap
Production
Equipment
Production Equipment
Inspection Production Inspection Inspection
Equipment
Wafer Outs
Time In Time Out
Wafer Moves
Wafer Starts
213
Figure B.16 – Probe interface card and electrical tests
The objectives of EWS test are the chip functionality and sorting good and bad dies, so
evidently this test is performed at die level. The test results are categorized by assigning different bin
numbers to the wafers. It includes continuity test, output checks and functional tests. The only issue
with this type of test is the time required, hence fault models are used to sort the dies and optimize
the time. Few reasons for the bad products has been identified as the larger wafer diameters,
increased die size, increase in number of process steps, shrinking features, process maturity and
crystal defects. It is really interesting to note that R&D engineers are specially struggling in
performing reverse engineering to trace back the root cause for bad product yield and variation in the
drifts.
b) Assembly and Packaging: In this phase, the good dies are separated from the wafer, assembled
and are packaged. The assembly involves backgrind, die separation, die attach and wire bonding.
The backgrind reduces the wafer thickness and die separation cuts each die from the wafer and it is
followed by physical attachment on the substrate. Wirebonding finally attaches wires from die
bonding pads to the lead frame of the substrate for the external connections [Quirk and Serda, 2000
and Scotten, 2012].
The traditional packaging consists of plastic or ceramic packaging. The ceramic packaging is
used for the state of the art ICs that require a maximum reliability or high power. The two main types
of ceramic packaging are either a refractory (high temperature) ceramic or ceramic DIP (CERDIP)
technology. The main objective is to protect the dies from environment, interconnections, physical
support and heat dissipation. The new packaging designs are being introduced to provide more
reliable, faster and higher-density circuits at lower cost. The most common technology is the ball
grid array (BGA). It uses ceramic/plastic substrate with an array of solder balls to connect substrate
to the circuit (Figure B.17).
Figure B.17 – Probe interface card and electrical tests (Scotten, 2012)
c) Final Test: It is likely that during packaging the die may have been damaged and its failure due
to packaging issues is more than 1%. The final test is 100% on all assembled and packaged chips
to insure that any IC improperly packaged is not shipped.
214
B.6 Key Manufacturing Processes and Challenges
Before we move to the role of DFM in the semiconductor industry, let us rapidly summarize the key
manufacturing processes (i) oxidation, (ii) photolithography, (iii) etch, (iv) ion implantation, (v) chemical
vapor deposition (CVD) for thins layer and (vi) polishing a.k.a. CMP.
i) Oxidation: It is a batch process where multiple wafer (200+) are processed together and SiO2 is
grown on the silicon wafer surface between 900ºC and 1200ºC in the oxidation furnace. The wafers
are heated in the furnaces containing oxidant (process gas), usually O2, steam or N2O that result in
the formation of thin deposition of oxide. The schematic of the oxidation furnace is presented in
Figure B.18 [Hwaiyu, 2005 and Michael, 2000].
ii) Photo Lithography: It is one of the mostly used steps in the manufacturing process. The structural
schematic description of litho stepper equipment is presented in Figure B.19. The wafers in this
process undergo the resist coat followed by the soft or hard bake steps depending on resist type being
used. The transfer sections are responsible for the transfer of wafers to/from alignment exposure
sections and resist coat and development sections. These wafers are loaded back to the wafer cassette
(Lot) which is a box that carries 25 wafers at a time [Hwaiyu, 2005].
215
This operation is very critical to insure the manufacturability and yield issues due to mask
misalignment errors. The most common errors associated with this process are grouped into focus
exposure misalignment (Table B.2) and illumination (Table B.3) errors.
Table B.2 – important focus exposure misalignment factors [DFM Dictionary, 2004]
Sub wavelength: Shortly after the turn of the century, a stepper using
248nm illumination wavelength (λ) was able to produce 130nm features
on the silicon wafer and193nm λ illumination produced 90nm features.
Aggressive RET enabled feature sizes less than half of the wavelength
of the illumination, but as silicon features continued to shrink, the sub-
wavelength challenge resulted in added complexity. The 193nm node
has become the end of the road for optical wavelength reduction. The
efforts to build 157nm λ stepper system were abandoned due to the
technical problems and the steppers shifted towards immersion
technology and water instead of air, at the interface between the lens
and wafer.
216
Mask Error Enhancement Factor (MEEF): It is because of the
fact that wavelength of light used to expose resist is larger than the
feature size; hence edge placement errors (EPE) are amplified. So,
during OPC modeling, an edge movement on a mask feature can result
in a larger edge movement on the silicon image. The change in the edge
placement on the mask is reduced by the reduction ratio (R) of the
projection system and amplified by MEEF so that a 2 nm edge shift on
the mask, even reduced by 4X can result in a 1 nm shift on the wafer.
217
b) Plasma as Etch gas
218
Etch Selectivity: Etch selectivity is determined by the ratio
of the etch rates of two different materials. For example, the
selectivity of a chemical etch for silicon dioxide compared to
silicon is 20x times faster. In the example on left, we need to
etch through the silicon dioxide down to silicon substrate.
However, the center area of silicon dioxide is 100nm (= h1 )
higher than the left or right edges. Removing all of the oxide in
the center means the other areas will be over etched. If the etch
rate of silicon is 20X less than silicon dioxide, the silicon
valley V2 (shown on the right) will only be over etched by 5
nm (= 100/20).
Polishing pad
(1) Slurry dispense (3) Mechanical force presses slurry into wafer
- Si
(2) H2O & OH travel to wafer surface Si(OH)4
Si (4) Surface reactions and
mechanical abrasion
Si Si Si Si Si Si Si Si
Si Si Si Si Si Si Si Si Si Si Si Si Si Si Si
SiO2 layer
The most commonly used metal for metallization is copper a.k.a. soft metal. The wide
wires can experience copper dishing in the center of metal lines which is removed during
polishing. A similar problem called erosion occurs on the dielectric oxide that separates wires,
hence the dishing is related with the wire widths and erosion is related to difference in wire
density (Figure B.24).
220
ensures the product functionality against the given specification; hence any deviation from these defined
design rules shall result in product functional yield loss. In comparison to the DRM, the DFM rules are
basically an extension to the design rules and are focused on insuring the printability of the features on
silicon wafer. The DFM models are used to perform CAA and hotspot analyses to identify the printability
issues in the design layout and the compensations are added to result the design layout which is more
manufacturable with less yield issues. These compensations are added to the mask which acts as an interface
between design and manufacturing processes.
An electronic product (chip) is characterized by electrical parameters; hence an unexpected
drift/variation (beyond model corners) is to be investigated for its root causes. The major cause for these
drifts can be traced to uncontrolled variations in the geometric shapes but it is not always the case and such
variations might present the newly emerging (spatial) behaviors, not included in the SPICE models, design
rules, DFM rules or DFM models. Our success lies in the ability to quickly analyze and transform these new
variations into models and/or rules during technology transfer and technology derivative initiatives
(alignment) and product design and development (adoption) efforts to reduce the technology lead times and
associated costs.
Let us summarize different yield and manufacturability issues faced by R&D engineers on daily
basis arising from the key manufacturing variation mechanisms (section B.6). If we do not come up with the
correct response then we are obliged to increase the area of the design layout to insure the product
functionality which is against the industrial slogan smaller, faster and cheaper. The most common DFM
issues are classified as systematic or random defects. The random defects cannot be controlled; however the
systematic defects can be easily controlled and modeled. Before proceeding into the role of DFM challenges
in SI, let us quickly summarize the most well-known DFM defects and respective rules and/or models used
in the CAD simulations.
B.7.1 Manufacturability Issues
The most common manufacturability issues irrespective of the technology are discussed in order to
demonstrate the impact of the key process behaviors on the resulting IC chips. The SPICE models take into
account the effect of geometric feature variations and predict its influence on the electrical parameters of the
product. In the example below, at the F1 E1 condition, the images are printed larger than normal because of
the higher light intensity with greater diffusion exposes more of the resist and the opposite occurs at the F1
E-1 condition (Figure B.25). It is also likely possible that contact over the via is misaligned and totally
misses the metal resulting in a functional failure [Hwaiyu, 2005).
These worst process conditions when supported with varying etch conditions, unreliable depth of
focus and CMP effect; often result in systematic manufacturability issues (Table B.5):
221
Transistor end cap extension: The gate poly of the transistor must
extend beyond the active area to make sure that there are no current leakage
issues. This extension is known as end cap extension and the mask, overlay
and exposure focus error often result in the reduction of end cap extensions.
Gate Length: The Transistor gate length is a critical factor for both
timing and power. The gateLengthMin and gateLengthMax checks measure
the transistor's gate lengths at several places along the transistor's channel
and trigger the detector if the measurements fall within the specified {min
max) measurement range.
Via and contact coverage issues: The coverage is defined as the area
of overlap, which can range from 0% to 100%, depending on the
intersection of the two images on each layer. We can see in example that
these errors often or might result in missing contact and vias causing
serious issues in interconnection and contact formation steps. The tOC and
tLE_E refers to outer corner and line end extension and are known as
feature fragments.
Bridging and necking: These are defined as if the space between two
images or width of an image falls within a specified {min, max} range. It
must be a positive value to avoid opens or shorts that result in functional
product failure. The red lines represent the drawn features after OPC
whereas the bleu lines represent the silicon image printed on the silicon.
Open and short: The open defect is said to be formed if the distance
across an image falls within the range of {0, max} and a short is formed
when the distance between two or more images is greater than 0. It has to
be a positive distance at the nominal focus exposure (FE) conditions.
Space and width defect: The Feature Space is the distance from an edge
fragment outward to nearest feature whereas the feature width is the
shortest distance across an image from an Image Edge to the opposite
Image Edge. The feature edges are composed of several different Fragment
Types e.g. tE (edge), tOC (outer corner), tIC (inner corner) and tLE_E (line
end edge) where tLE_E > tOC > tIC>tE.
Line end edge issues: If image pushes out beyond the feature it is
called Line-End Push Out (LEPO) whereas shrinking is referred as Line-
End Pull Back (LEPB). These issues can become swear if not controlled
specially for the contact and vias and current leakage.
Random particles: The incidental particles are the major causes for
opens and shorts on the layer being fabricated. They can randomly fall
anywhere on the integrated circuit resulting in its functional failure are
called random defects whereas systematic defects (hot spots) can be easily
modeled and controlled.
Table B.5 – Systematic and random manufacturing defects [DFM Dictionary, 2004]
222
The deviations between intended and realized physical pattern(s) on silicon are called physical faults
resulting in the functional failure of the product and deviations in the electrical characteristics are referred as
the electrical faults. The physical deviations often result from the imperfection in manufacturing processes
and the mechanisms behind these are classified as incidental and base line. The incidental mechanisms (gap
fills, linear deposition etc.) can be resolved and controlled whereas the baseline mechanisms (flakes, particles
in/on resist, incomplete develop etc.) are very hard to model.
We have discussed in chapter-1, section 1.2 that technology alignment and adoption lead times are
critical for the success of SI; hence it is our ability to analyze and model the faults and associated
mechanisms (Table B.6) that result in the success or failures. Let us summarize generic approaches
(rules/models) adopted to avoid these manufacturability issues and yield limiters prior to discuss the role of
DFM in the evolution of SI.
223
Process and Proximity Compensation: The model based Optical
Proximity Correction (OPC) and Process and Proximity Compensation
(PPC), both rely upon physical models to identify the faults that occur
during pattern transfer from a mask (reticle) to the wafer. A Black Box
model lumps all these non-idealities in a single, multi-dimensional
model. Unfortunately, the effects of one parameter change (a mask
error, defocus condition, over exposure condition, under etch
condition) can either compensate or multiply the effects of another
parameter change. Physics based models; calculate these effects
separately in the sequential flow. The various process models are
calibrated with both measured and/or simulated data within a
predefined range.
Table B.6 – Generic approaches to address failure mechanisms [DFM Dictionary, 2004]
The above list is just a summary of key failures and remedies (Table B.5 and Table B.6) from an
exhaustive list. It is evident that engineers are always focused on addressing these manufacturing and yield
limiting factors as discussed above. We have also seen that the efforts to address these issues either result in
some rules (design/DFM rules) and/or models (SI2007). Let us take an example of a circuit and follow the
most common techniques as discussed above to improve silicon features printability on the silicon image
using OPC (Table B.7).
a) The end of Line (EOL) extension improves Line b) Hammer Head improves Line End Pull Back
End Push Out (LEPO) (LEPB)
c) Inner Corners (IC) have a negative offset to d) Outer Corners (OC) do not have excessive corner
improve corner spill fill
e) Via Coverage (Two Layer Overlap Area) is f) Long straight lines have been biased correctly so
symmetric OPC is required
224
Table B.7 – Layout improvement efforts using OPC technique [DFM Dictionary, 2004]
The DFM optimization helps to make layouts more robust against both systematic and random yield
loss mechanisms. Robustness against systematic yield loss mechanisms is particularly relevant to help avoid
unforeseen design-process marginalities, especially during technology alignment and adoption efforts. The
ramp-up experience has shown that design-process marginalities occur almost exclusively at layout
configurations that are exactly at minimum design rule; hence they can be strongly reduced by designing just
slightly (5-10%) above minimum design rule, where possible. The minor corrections can make a major
difference for systematic yield.
The robustness against random yield loss mechanisms is relevant in volume production, where a
small yield difference on a large number of lots represents a significant financial value. The defect density
distribution is commonly used in conjunction with critical area models for yield predictions. The practical
DFM implication is that minor corrections also give minor effects where random yield is concerned; hence
the DFM corrections should be fairly large: via doubling, metal width/space > +30%, etc. It must be noted
that the DFM actions for random yield should not lead to an increase in chip area, because an improvement
in random yield will almost never outweigh the gross die decrease.
225
226
Appendix C: CMOS Inverter Design and Manufacturing (An Example)
We have briefly reviewed the SI design and manufacturing process in Chapter-2 to provide users a glimpse
of complexity involved in the integrated circuits. In order to strengthen the IC design and manufacturing
knowledge and for the readers who wish to see those processes in more detail, we took a simple CMOS
inverter (a circuit with 2 transistors) example and completed the design, mask and manufacturing steps. The
important phenomenon that might raise the DFM challenges are also discussed and commented to improve
the understanding of the DFM methods and their evolution.
C.1 The CMOS Inverter Design
The objective of this section is to go through the design and manufacturing steps of a simple CMOS
inverter circuit and identify the major challenges being faced by the engineers at all steps. The inverter is
truly the nucleus of all digital designs and is the basis of logic gates for complex electronics functions for the
digital and analog designs. It is composed of n-MOS and p-MOS transistors together and its basic function is
to inverse the signal. Before proceeding with the example, let us first simplify the operations of n-MOS and p-MOS
transistors (Figure C.1). It is easy to interpret that when a voltage (potential) is applied to the gate, p-MOS results in
an open circuit i.e. no current flows from the drain to source whereas in case of no voltage the current do flow.
1 0
0
1
1 0
1 0
1 0
p-MOS switch: “1”=open circuit
1 0 n-MOS switch: “0”=open circuit
No current No current
0 1
0 1
1 0 1 0
1 0 1 0
p-MOS switch: “0”=pass circuit Idrain n-MOS switch: “1”=pass circuit Idrain
227
minimum dimension allowed in the technology and is to be drawn as indicated in the schematic but W is allowed
vary. The µo, q, K are the process technology constants.
In
NMOS PMOS
Out
1
in 1
out
P-well N-well
ON
1 Log ISD current
Trade-off: Key process parameters:
0 1 Increase Ion Gate oxide thickness
Decrease Ioff Poly gate length
Implants
OFF
current
Sub-Vt
slope
0
Threshold
1 Voltage (Vt)
0
VGS
0 1
1 0
If If
Gate=“1” Gate=“0”
Then Then
NMOS NMOS
we pull down we pull down
Out to “0” Out to “1”
0 0 0 0
N-Well (N-)
Metal-1 P+ Implant
Figure C.5 – Colors and notations for CMOS inverter CAD design
We start with two potential physical designs of the CMOS inverter. The design as presented in
Figure C.6a uses the p and n diffusion layers for the formation of n and p MOS transistors and p-type
substrate whereas the design presented in C.6b uses the diffusion and p-implant in the structure of the
inverter.
VDD VDD
VSS
VSS
229
2 and 4 from left to right are flipped to have reversed their sources and drains to transform the transistor
layout into a compact transistor as shown in the figures C.7b and C.7c respectively.
0.2µm
S D S D S D S D
D
2µm
S D D S S D D S
a) Transistor with 2µm width b) Transistor divided in 4 transistors with 0.5µm width and S-D flipping c) Compact Transistor
DL/2 DL/2
Source Drain
L
DW/2 DW/2
Figure C.8 – Delta in drawn length (DL) and width (DW) of transistor [Lee, 2005]
Let us consider an example where DL is 0.015um and DW is 0.045um. The fast transistor is
simulated with narrower L and wider W (negative DL, -0.015um) whereas the slow transistor is modeled
using a wider L and smaller W (negative DW, -0.045um). The following table (Table C.1) shows L and W of
the transistors from the first and second optimization efforts, and the geometric variations at the slow, typical
and fast corners that emulate the manufacturing process tolerances.
Fast Typical Slow
Maximum L(um) 0.185 0.200 0.215
Contacts W(um) 20.045 20.000 19.955
Compact L(um) 0.185 0.200 0.215
Transistor W(um) 20.180 20.000 19.820
Table C.1 – Model corners for fast, typical and slow devices
230
The impact of layout optimization techniques do not seem to have a strong variation in the two
optimization steps on width; however the length undergoes significant variation due to under or over etch
and a compact transistor formation. The impact of the length variation (0.18um) between slow or fast corner
against typical corner is 4x in comparison to the one without compact transistor. This could pose a circuit
performance deviation for the circuit designers if left unaccounted. The designers perform Monte Carlo
simulation by adding small statistical variation to W and L of every transistor in the circuit with an objective
to center design for the yield improvement. The layout can be challenging for the circuit designer to optimize
it for better yield.
The subsequent layout optimizations methods applied on the circuit layout are (i) Speed up the
transistor with higher frequency response by reducing the parasitic components (resistance and capacitance),
(ii) clean up the substrate disturbances from minority carrier and coupling noise, (iii) balancing area, speed
and noise, (iii) stress relief (insert dummy transistors), (iv) protect the gate (control charge accumulation at
the poly gate a.k.a. antenna effect) and (v) improve yield (avoiding single via or contact, spacing, end-of-line
variations, metal coverage of contacts and vias). All these optimization efforts result in design rules
programmed into design kits.
In this section we have drawn the logical circuit (schematic) for CMOS inverter followed by its
transformation into physical design. The CAD simulation using technology based SPICE Models for the
extraction of electrical parameters ensure manufacturability and functionality. The physical (transistor level
synthesis) requires further simulations based on manufacturability variations. It is based on the fact that we
are not able to manufacture the geometric shapes like they are drawn in the schematic; hence layout
optimization is required followed by the CAD simulation to ensure timing, power and leakage. We have not
physically simulated the logical and physical designs because it shall result in more than 2000+ electrical
parameters; however we have presented and introduced the whole process. We have optimized the layouts
manually whereas other optimization techniques are listed for reference. The CAD simulations follow an
automated process using software tools which are supported with design kits, process kits and DFM kits. It is
also important to note that we have not included discussion on the BEOL interconnections but the CAD
simulation flow is exactly similar with the only difference of SPICE models. The validated designs result in
the generation of netlist which is used as an input for the mask preparation.
3 4
1 Lithography pattern 1 1
2
2 Test pattern 2 5 5
231
The most important issues towards manufacturing and yield are printability (lithography), polishing
and etching (under/over etch) issues that result in geometric shape variations. It directly impacts the electrical
parameters (2000+) resulting in functional or parametric yield losses. The cross section view of the CMOS
inverter fabrication is presented in Figure C.10. In this circuit, Vdd refers to the drain voltage and GND is the
ground connection. The y is the field oxide that separates two transistors and A refers to the poly gate
contacts.
GND VDD
232
C.3 CMOS Inverter Manufacturing Steps
We start with a raw wafer and first manufacturing process step is the oxidation step (Figure C.12)
where SiO2 (dielectric) is grown on the wafer surface at 1200ºC in an oxidation furnace with H 2O and O2.
The principle objective here is to protect the surface of the wafer from the subsequent manufacturing steps.
The wafers are heated in the furnaces containing oxidant, usually O2, steam or N2O resulting in the
deposition layer of an oxide. It is important to note that no mask is used during this step.
p substrate
SiO2
p substrate
Photoresist
SiO2
p substrate
p substrate
233
SiO2
p substrate
n well
p substrate
n well
p substrate
The next step is the polysilicon deposition (Figure C.15) using Mask 2. We start by
depositing a very thin layer of gate oxide < 20 Å (6-7 atomic layers) in the furnace. This step follows
the Chemical Vapor Deposition (CVD) for the deposition of polysilicon. The wafer is placed in
furnace with Silane gas (SiH4) that forms many small crystals called polysilicon.
Polysilicon
Thin gate oxide
n well
p substrate
Polysilicon
Polysilicon
Thin gate oxide
n well
p substrate
The n+ diffusion step is a self aligned process where we repeat the oxide (Figure C.16a) and masking
steps for n+ dopants diffusion to form nMOS source, drain and n-well contact (Figure C.16b). The
polysilicon step is better than the metallization step for self aligned gates because it does not melt during
processing. The n+ diffusion is made (Figure C.16c) using diffusion process where impurities are absorbed
on and beneath the wafer surface using heat. It is replaced with the ion implantation due to better control but
for the source and drain regions it is still called diffusion. The oxide is stripped off to complete the patterning
step (Figure C.16d).
234
n well
p substrate
a) Oxidation layer
n+ Diffusion
n well
p substrate
n+ n+ n+
n well
p substrate
n+ n+ n+
n well
p substrate
The next step is a p+ diffusion that forms the regions for pMOS source, drain and substrate contact
(Figure C.17).
p+ Diffusion
p+ n+ n+ p+ p+ n+
n well
p substrate
The final step includes the deposition of contacts and metallization to wire the devices n-
MOS and p-MOS transistors. The chip is covered with thick field oxide and the contacts are etched
(Figure C.18a). The aluminum is sputtered over the whole wafer and patterned to remove excess
metal leaving behind the wires (Figure C.18b).
Contact
n well
p substrate
a) metal contacts
235
Metal
Metal
Thick field oxide
p+ n+ n+ p+ p+ n+
n well
p substrate
b) line metallization
236
Appendix D: SMA (Spice Model Alignment) Tool
The proposed Spice Model Alignment (SMA) tool is focused on helping R&D engineers for new front-end
(FE) technology transfer, developed in the international semiconductor development alliance (ISDA). The
use case diagram for technology transfer is presented in Figure D.1 where ISDA, technology R&D, process
integration (PI) and design groups are the key stake holders. The ISDA alliance develops new technology
and defines the perimeters. These perimeters are the key technology parameters along with their target and
specification limits. The technology R&D group in the receiving plant is responsible for alignment of these
parameters which is further classified as the technology maturity levels 10, 20, 30 etc. The PI team plays a
key role during the alignment of these parameters which includes process flow, design rule check, and design
rule manual alignment against device and interconnect SPICE model parameters. The design group develops
design kits taking into account the manufacturing variations encountered during the alignment process to
ensure manufacturability and yield during prototyping and normal production.
The scope of SMA tool is around the process flow alignment against the SPICE model parametric
alignment by the ISDA alliance. The device and interconnect SPICE models maturity requires respective
technology parameters alignment against the changes made by the ISDA alliance. It includes the
management of standard and special conditions to be used during design CAD simulations of technology
parameters using test product. The test structures a.k.a. vehicles are used during these simulations and are
specially designed to represent the complexity of the target products to be developed using this technology.
The process starts with the definition of device geometries [W, L] scaling to be tested through simulations
for its compliance against target technology parameters and subsequent validation by manufacturing on
silicon wafers. The simulated SPICE parameters computation often requires normalization in order to format
the simulated results into unified measurement units. These SPICE model parameters are then related with
237
the PT parameters to generate PT specs. These generated PT specs are used for the process alignment. It is
very important to normalize the SPICE models’ simulated parameters because the PT specs measured on the
production line have specific formats depending on the type of metrology equipment used. The lower part of
the use case diagram is the part of proposed system where during process alignment any drifts or variations
from the simulated parameters are further investigated using correlation between geometric shape variations.
The challenges faced within this process are as under:
a) The names of SPICE model parameters from the simulations are not standard; hence PT-SPICE
parameter correlation is to be manually controlled for each technology and each maturity level of
SPICE models. Every technology has almost 2000+ parameters and normally engineers are
allotted with 200 parameters each to manage and control the maturity and respective process
alignment.
b) The computation of SPICE parameters using SPICE Models result in the tab limited data files
(dat, csv etc.) file formats. These simulation files are hard to manipulate and engineers spend
huge amount of time because multiple spice parameters are used and normalized to compute
technology parameters. The varying file format of simulation files is a big issue along with the
computation and normalization.
In order to address these issues and help PI and Technology R&D engineers, the SMA tool is
developed and deployed to achieve following objectives which result in quick technology transfer and
alignment.
Manage SPICE-PT parameter relationship
Generate PT-SPECS applying normalization and choice of corners
Simulate parameters and edit normalization formulas
The sequence diagrams, data model and deployment diagrams are not presented here, however the
tool is explained through different functionalities using GUIs. The reason for this is the confidentiality as
well as space restrictions in the thesis. The SMA tool is versioned (X.Y.Z) upon each modification requested
by the engineers where X Major Version, Y Minor Version and Z Revisions. It is developed using
Visual Basic 6.0 professional edition and MS-Access 2003 for prototyping. It is 2-tier, multi-user database
application where all users are connected to the central shared database for authentication and profile
management. It is a relational database and also serves as the share point for pre-simulated SPICE
parameters among the application users. If the engineers requested SPICE parametric simulation results are
in the parameters which are shared then those parameters are directly extracted and are not simulated. The
functionality of the tool is presented through GUI explanations. The proposed tool is unique in the sense that
based on the functionality selected; the local database is dynamically generated where analysis results are
stored until and unless it is removed by the end users. The end users can share the simulation results in the
central database for the other tool users.
238
Figure D.2 – SPICE Model Maturity and Process Alignment
The first GUI is presented in Figure D.3 (a & b) where end users can browse simulation result files
(SPICE parameters) by selecting the folder. The files are listed with the description of each file in
folder/subfolders with its extensions and size. These files can be selected by double clicking on the files
which are then added to the list box on the right; however user can select all files by clicking the button <<
Select All >>. The PT specs file is required to be browsed by each engineer for his own set of SPICE
parameters which provide relationship between SPICE and PT parameters along with the normalization and
computation formulas. Once selected, users import these relationships by clicking the button << import >>.
The end users are also provided with option to filter Spice parameters, devices and PT parameters by
selecting from the respective list boxes.
239
Figure D.3 (a) – SMA Tool Main GUI
The PT-SPICE relationship file when loaded is presented in the Figure D.4 (a) where model file and
description refers to the SPICE parameters along with respective PT parameter name and units. The Figure
D.4 (b) presents formulae for PT parameters along with normalization. The users can change these formulas
at any time where respective results are computed instantly based on the new formulas. The changes made
can be made permanent for future. The normalization formulas do include the units and can be easily
modified to ensure the final resulting units for compliance with the measure PT parameters. The user can
double click on any of the listed formulas where he is provided option as a text box along with previous
formulae for change. Once the changes are made, they can be made permanent by clicking the enter key. The
users are required to select option “yes’ from the message box to save it permanently, however these changes
are made in the dynamic data base created on the local computer of the user. If user want to keep the
relationship file as an excel sheet, he can extract the whole file by clicking the button on the bottom <<
Export >>.
240
Figure D.4 (a) – Import of PT-SPICE Relationship File
One of the important algorithms implemented in SMA tool, is computation of the PT specs based on
one or multiple simulated SPICE parameters. The implemented algorithm is based on the concept of
“STACK” data structure which is traversed on last in first out (ILFO) principle. The algorithm cannot be
presented, however simulation of the algorithm with an example formulae, is presented below in Figure D.5
(a & b). The formulae is presented in infix format as (1 + (2 * ((3 + (4 * 5)) * 6))) which is first converted
into postfix notation as 1 2 3 4 5 * + 6 * * +. This post fix notation is then loaded into a stack and traversed
for the computation of formulae. This algorithm has two steps as (i) infix to postfix notation and (ii)
computation of postfix notation expression. The infix notation computation is presented in Figure D.5 (a)
where we search for the inner most parentheses and transform it into post fix notation. The algorithm
continues until the parentheses are finished. The parentheses are removed from the final post fix notation
which is used in the expression evaluation presented in Figure D.5 (b).
241
Figure D.5 (a) – Normalization Formulas and Instant Changes
(a) (b)
(c) (d)
242
(d) (e)
(f) (g)
(h) (i)
243
(j) (k)
(l) (m)
(n) (o)
Figure D.5 (b) – Stack Structure based Expression Evaluation
244
The Figure D.5 (b) describes the sequence in which postfix notation based mathematical formula is
read and evaluated. In the Figure D.4 (a) the options << customized corners >> and << default corners >>
provide users to define the scope based on what PT specs are generated. The SPICE parameters are simulated
based on the model corners as SS (slow), FF (fast) and TT (target). The ST prefix stands for
STMicroelectronics where respective model used for the simulation is the company’s internal model. The
notations SSF and FFF stands for slow-fast and fast-fast corners. The user clicks on the button << Show
Results >> to compute PT specs as LSL, Target and USL. The target PT spec is the target computed from the
simulated SPICE parameters whereas LSL and USL specs corresponds to the minimum and maximum model
corners as shown in Figure D.6 (a).
245
246
Appendix E: BEOL (back-end-of-line) Variance Analysis Tool
The BEOL variance analysis tool is developed for T2D aples team who is primarily responsible for the PT-
variance analysis followed by silicon correlation to answer drifts and excursions. The IDEF model at two
levels is presented in Figure E.1 below.
The most import tasks in this process are (i) PT data extraction (A1) followed by (ii) PT variance
analysis a.k.a. BEOL parasitic variability analysis (A3). The physical parametric data extraction and
variability analysis (A3 and A4) refers to the geometric measurements made on the silicon wafer which leads
to the possibility of Si correlation analyses (A5). The BEOL variance analysis tool addresses the challenges
faced by the T2D aples team during the steps A1 and A3. The data extracted using data extraction utilities;
hence we do not find any format issues but data volume is quite high. The end users also filter data upon its
extraction; hence the resulting data is normally distributed and do not exhibit significant variation. The
engineers apply manual filter i.e. delete by hand certain values as outliers which might remove the potential
drifts. The computation of different PT correlations based on scribe line and metal layers, is tedious job and
engineers spend huge amount of time in data extraction and analysis to find the drifts which can be further
investigated to find root causes.
The proposed tool offer following advantages to the R&D engineers:
a) It verifies the data format and completeness issues
b) perform filtered and non filtered data analysis
c) generate PT variance analysis graphs to be used for further investigation
The BEOL-Variance analysis tool is presented with its GUIs due to confidentiality issues, however it
uses multidimensional data model at the back end to improve the computational efficiency. It has reduced the
hectic work of weeks into 5 minutes processing and as a result, engineers focus on the analysis and
interpretation for decision making rather than pre-processing and generating graphs. The tool starts with the
user authentication as presented in Figure E.2.
247
Figure E.2 – Users Login and Authentication Step
The first step is to browse the capacitance and resistance scribe and field files. The users can validate
the selected files against the format, syntax and completeness by clicking on the button << Verify Data >>.
The basic input file characteristics are computed and displayed to user for reference (Figure E.3).
The results are saved at users selected paths in an excel sheet as presented in Figure E.6 (a). The
parasitic element based analysis using excel and built-in functions are presented in the Figure E.6 (b). It can
be seen that the role of algorithms in statistics computation is highly critical. The results provide a
comparison of no filter, LSL-USL filter and LVL-UVL filters and it can be seen that when filters are applied
and number of counts are less the difference between the computed statistics using excel and built-in
functions is quite significant.
Figure E.6 (a) – Computation of Statistics with Excel and Built-in Functions
249
Figure E.6 (b) – Parasitic Elements (Resistance/Capacitance) Analysis
The Figures E.6 (c & d) presents the graphs generated by tool as the box whisker plots for scribe line
or field metal resistances and capacitances. As it is said earlier that significant variation or drifts are further
investigated for their root causes. The first and prime suspect of these drifts is the variation in geometric
shapes of transistors; hence we need to further correlate the PT drifts against these geometric variations. This
PT-Inline correlation needs site level information that shall act as a pivot for multi-source data analysis for
root cause analysis. It is presented in appendix F, however proposed solution (KLA-ACE recipe) takes into
account the site level information for both PT and Inline data and also provides the end users to create new
variable to be used during analyses.
250
Figure E.6 (d) – Parasitic Elements (Resistance/Capacitance) Analysis
251
252
Appendix F: KLA-Ace Recipe for PT-Inline Correlation
The objective of PT-Inline correlation is to find root causes against PT drifts and variations. This section is
the continuation of the appendix-E where we are focused on finding PT variance analysis using our proposed
BEOL-Variance analysis tool followed by root cause analysis against drifts. We have developed KLA-ACE
Recipe that follows the following activity diagram for the mapping and alignment of multi-source data at site
levels. The KLA-ACE is a well-known multi-source data analysis tool used in the semiconductor industry. It
is a commercial product by KLA Tencor which is based on workflow engine concept. The end users create
recipes using different data extraction, pre-processing and alignment nodes. These recipes are created offline
but when executed in one go or steps results in designed analysis.
Input Lots
PT-Inline Data
Extraction
Statistics/Charts Option
Input Inline Data and
Filter Inline Data
R, C Variance Analysis
KLA-ACE Stat
Compute Resis Var Compute Capa Var
Compute RC Data
KLA-ACE Charts
Index based on X,Y f or PT
Compute RC Var
Replace Site No. With
Merge RC X,Y Data
KLA-ACE Stat
KLA-ACE Charts
Input MASK Data
Statistics & Chart Analysis
KLA-ACE Stat
Merge Resis Inline Data
63
3
2 35
1
2 -5 0 2
3 -5 1 3
. . . .
63 4 2 17
We use mask level information to find site level coordinates of the sites tested for geometric
measurements (inline). These x,y coordinates are then helpful in performing site level mapping for
correlation purposes as show in the Figure F.3. The proposed method is almost similar to the hard coded
methods for the mapping and alignment. The need for generic model for the site and die level mapping and
die to site qualification was felt very badly when we observed varying measurement reference frames and
notch positions for different metrology tests which often result in varying site and die level x,y coordinates.
We then propose MAM and SPM models as presented in chapter-5 for generic site/die level mapping and
alignment, and die to site qualification for accurate analyses. The said recipe is still valid and being used for
PT-Inline site level correlation analyses.
3 66
2
1
4
13 2 -5 0 . . . 2
3 -5 1 12 -4 -2 .
12 4 -4 -2 13 -5 0 12
. . . . . . 13
The said core flow as presented in Figure F.1 also provides the end users with the ability to
create new variables from existing variables. It is very important function because the width and
height against a geometric shape are computed from multiple inline parameters. It facilitates users
with variety of correlations.
254
Appendix G: EPP (Equipment, Product, Process) Life Cycle Tool
The EPP tool is primarily developed for the IMPROVE (Implementing Manufacturing Solutions to Increase
Equipment Productivity and FAB Performance) project. The main objective is to provide contextual data to
support effective root cause analyses efforts. It is a 2-tier and multi-user database application developed
using VB6.0 professional and SQL Server 2008. This EPP is developed with a multi-dimensional database,
however in this appendix the data model and detailed UML model are not presented due to confidentiality
and space limitations. The proposed application is presented using the GUIs.
The EPP starts with users’ authentication with id and password. This application during its startup
identifies the best resolution and adjusts all objects on the GUIs accordingly (Figure G.1). The users start by
selecting the duration of the data extraction, however default period is of recent one week. It is important to
note that output of this tool is classified as the product life cycle or equipment life cycle.
The users do have an option to select type of workshop, equipment or module(s). The modules are
considered as the sub-equipments which are associated with the parent equipment in parent-child
relationship. This tool is directly connected with the maintenance data warehouse sp we start by clicking the
<< Import >> button. It fetches data from data warehouse into local multidimensional database (Figure G.2).
The data comprises of “Equipment States”, “Equipment Failure to Work Request/Order” and “Equipment
Checklist Steps”. The << Process >> button aligns data and generate equipment life cycle for the given
period which is presented in the Tab << Equipment Life Cycle Data >>. The users have an option to scroll
and go through other data sets generated in three tabs. The option “Arrange Columns” automatically adjusts
the length and width of the columns for user readability.
The internal algorithm which aligns and generates the equipment life cycle is not presented here
because of the confidentiality reasons as it involves the internal data warehouse schema. The equipment life
cycle data is objectively focused on tracing the equipment states, respective failure modes which lead to the
work request/order and information about the check list tasks performed as corrective or preventive
maintenance against these work orders. This data provides us a possible connection with OOC (out of
control) production lots to assess whether equipment health is related to the yield loss or not. It also helps us
in building models for predictive maintenance to avoid the expensive corrective maintenances.
The << WIP (Lots) >> button generates the list of the production lots which were being processed
when the equipment was undergoing different states, failures and actions. This list of production lots helps us
to extract product life cycle from the process data warehouse. Three KLA-ACE recipes are developed to
extract WIP, Metrology and Out of control data from the process, metrology and OOC databases. The data
extracted from these databases is linked to generate the product life cycle which is highly useful for
modeling different aspects. We start by running the KLA-ACE recipe to extract WIP and Metrology
information of the production lots; however OOC information is selected based on the duration directly from
255
the OOC database. The extracted WIP and Metrology data files (csv extracts) are browsed by the users as
presented in the GUI at Figure G.3. The users select all the files by clicking on button << Select All >> or by
clicking on individual files in the file list. The << Import >> button extracts the information into
multidimensional database to improve processing times. The << Process >> button extracts OOC data and
generates the product life cycle information which is presented to the user in the “Product Life Cycle Data”
tab, however the source data can be found in other three tabs. The “Autosize” option adjust the column
widths and heights for readability and this data can be exported in the csv format.
256
Appendix H: ACM (Alarm Control Management) Tool
The proposed ACM tool is subjected to the equipment engineering teams with an objective to help
them in equipment alarms management to reduce unnecessary equipment stops. The goals of the proposed
tool are to (i) improve equipment engineers’ efficiency and (ii) empower equipment engineers to learn
changing equipment behavior. The objectives targeted with this tool are as under:
a) Analyze, Manage and Control Alarms and Warning coming from TOOL
b) Learn Alarm Patterns linked with Product, Process or Equipment
c) Predict Cost and Yield related risk against significant Alarm Patterns
The ACM tool links the proprietary databases of the individual equipments with the other data
sources as presented in Figure H.1 below. The automation system collects the alarms generated by
equipment and stores them into alarms database. The ACM tools provides slicing, dicing and drill down/up
operations on the data extracted from these proprietary databases. The data collected from the production
line, stored in multiple data sources, is linked with the alarms generated by the equipments to build the PAM
and PSM models as presented in chapter-7. The objective is to learn the alarm patterns linked with bad and
good yields for subsequent use in the prediction to spare the capacities.
FDC
Automation
Alarms
WIP Equipment Database
Maint
Metro
ACM
The ACM tool is developed using VB 6.0 professional and MS SQL server and it is a 2-tier multi-
user database application. The database used at backend is a dimensional data warehouse and OLAP queries
are implemented as simple SQL queries. The query responses are optimized through views and indexing @
the cost of space. The key features of the tools are presented as under:
a) Input Alarm files, pre-process and filter Alarms in the database
b) Provide user Global Pareto and Best/Worst equipment statistics based on weekly and overall
basis
c) Alarm pre-processing prior to analysis on historical data
d) Generate Pareto on Count/Duration
e) Display Missing geometries within each simulation
f) Export the SPICE parameters, PT specs and model output to excel
g) Export selected SPICE/Model output parameter to excel
The tool is not presented as UML model due to space limitation; however a brief description is
presented through key and important GUIs. The users have an option to connect directly to the equipment
alarms proprietary database or input the alarms data from the alarms data extracted through data extraction
utilities. The duplicate records if found are separated during data input step. The main GUI is presented at
Figure H.2(a). The “Litho-Tool” name presents the type of equipment family under processing. The users
browse the input files which are then selected for import. The files with background color as green are
257
already pre-processed files. The button << Import >> results in importing raw data along with pre-processing
into multidimensional data warehouse. The alarms data in its raw format needs processing to find out its start
and end times for the computation of duration. The right side statistics presents the historical evolution of
alarms on the basis of error count and duration. The best and worst tools are highlighted for special
consideration of equipment engineer.
258
Figure H.2 (b) – Input Alarms Data for ACM Tool
259
Figure H.4 – Chart Options with Tool Tip Information
An example of data extraction is presented at Figure H.5 where user is informed upon the data
extraction about alarms exclusions, duplications and deletion. The users also have the option to filter and
further exclude, delete and/or include the alarms using alarm levels as error, warning or others. Te option
export to database is very critical as it allows the engineers to export the data as presented in the flexgrid into
the database whereas the same data can be exported to an excel sheet by clicking on the button << Save to
Excel >>. The data filtered across 4 tabs can be moved around by right clicking and then selecting the
options as delete, exclude, include or filter.
260
Figure H.6 – An Example of Data Extraction in Excel Sheet
The data extracted can be further used during data mining and similar operations using data mining
options as presented below in Figure H.7. The data can be filtered on multiple time dimensions and status of
the alarms.
261
Figure H.8 – Graphical Analysis with Slicing and Dicing Options
262
Figure H.10 – Historical Reporting for Alarms Count and Duration
263
264
References
[A]
[Abdelali et al., 2003] A. Abdelali, J. Cowie, D. Farwell, B. Ogden and S. Helmreich, Cross-Language
Information Retrieval using Ontology. Proceedings of TALN Batz-sur-Mer. France, 2003
[Abernethy and Altman, 1998] N. F. Abernethy and R. B. Altman, “SOPHIA: Providing basic knowledge
services with a common DBMS”, Proceedings of the 5th KRDB Workshop, Seattle, WA, 1998
[Ahmed and Abdalla, 2000] A.M. Ahmed and H.S. Abdalla, Beyond competition: A framework for the
21st century. International Journal of Production Research, 38(15):3677–3709, 2000
[Aitken et al., 2008] R. Aitken, L. Capodieci, F. Klass, C. Hou and V. Singh, DFM in Practice: Hit or Hype?
45th Annual ACM IEEE Design Automation Conference, 2008
[Andrews, 1980] K. R. Andrews. The Concept of Corporate Strategy. R. D. Irwin, 1980
[Anderson, 2006] D.M. Anderson, Design for Manufacturability & Concurrent Engineering; How to Design
for Low Cost, Design in High Quality, Design for Lean Manufacture, and Design Quickly for Fast
Production, 448 pages; CIM Press, 2006
[Appello et al., 2004] D. Appello, A. Fudoli and K. Giarda, Understanding Yield Losses in Logic Circuits,
IEEE Design & Test of Computers, pp.208-215, 2004
[Aparício et al., 2005] A.S. Aparicio, O.L.M Farias, and D.N. Santos, Applying Ontologies in the Integration
of Heterogeneous Relational Databases’, In Proceeding of the Australasian Ontology Workshop (AOW),
Sydney, Australia, pp. 11-16, 2005
[Arens et al., 1996] Y. Arens, C.N. Hsu, and C.A. Knoblock, Query processing in the SIMS information
mediator. In Advanced Planning Technology. AAAI Press, California, USA, 1996
[Astrova et al., 2007] I. Astrova, N. Korda, and A. Kalja, Storing OWL Ontologies in SQL Relational
Databases, World Academy of Science, Engineering and Technology, 2007
[B]
[Bachman, 1973] C.W. Bachman, The programmer as navigator. Communications of the ACM, 16(11):653–
658, 1973
[Ballhaus et al., 2009] W. Ballhaus, A. Pagella, and C. Vogel, A change of pace for the semiconductor
industry? Technical report, German Technology, Media and Telecommunications, November 2009
[Blackman, 1998] K. R. Blackman. Technical note: IMS celebrates thirty years as an IBM product. IBM
Systems Journal, 37(4):596–603, 1998
[Blodgett and Schultz1969] J.H. Blodgett and C.K. Schultz, Data processing pioneer. Journal of the
Amerrican Society for Information Science and Technology, 20(3):221–226, 1969
[Bolz, 1958] Bolz R.W. (Ed.); Metals Engineering Processes, McGraw-Hill, New York, 1958
[Boning et al., 2008] D.S. Boning, K. Balakrishnan, C. Hong, N. Drego, A. Farahanchi, K.M. Gettings,
L. Daihyun, A. Somani, H. Taylor, D. Truque, and X. Xiaolin, Variation. Semiconductor Manufacturing,
IEEE Transactions on, 21(1):63–71, 2008
[Boothroyd and Redford, 1968] G. Boothroyd, A.H. Redford, Mechanized Assembly, McGraw-Hill, London,
1968
[Boothroyd and Dewhurst, 1990] G. Boothroyd, and P. Dewhurst, Product Design for Assembly, Wakefield,
RI, USA, 1990
265
[Boothroyd, 1994] G. Boothroyd, Product Design for Manufacture and Assembly, Computer Aided Design,
Vol. 26, Issue 7, 1994
[Bralla, 1998] J.G. Bralla, Design for Manufacturability Handbook, Second Edition - McGraw-Hill
Professional, 1998
[Braun and MacDonald, 1982] E. Braun and S. MacDonald, Revolution in miniature: The history and impact
of semiconductor electronics. Cambridge University Press, 1982
[Brinkman et al., 1997] W.F. Brinkman, D.E. Haggan and W.W. Troutman, A history of the invention of the
transistor and where it will lead us. IEEE Journal of Solid-State Circuits, 32(12):1858 –1865, 1997
[Buccella and Cechich, 2003] A. Buccella and A. Cechich, An ontology approach to data integration.
JCS&T, Vol. 3 (No.2), 2003
[Buccella et al., 2005] A. Buccella, A. Cechich, N. Brisaboa, Ontology-Based Data Integration Methods: a
Framework for Comparison, Revista Colombiana de Computaci´on, Vol. 6, No. 1, 2005
[Business-Directory, 2007] BNET Business-Directory. Strategic analysis, 2007
[Businessballs, 2006] Businessballs. Pest market analysis tool, 2006
[Busse et al., 1999] S. Busse, R.D. Kutsche, U. Leser, and H. Weber, Federated Information Systems:
Concepts, Terminology and Architectures. Technical Report. Nr. 99-9, TU Berlin, 1999
[C]
[Calvanese et al., 2001] D. Calvanese, G. De. Giacomo, and M. Lenzerini, In Proc. of the 14th Int.
Workshop on Description Logic (DL 2001), volume 49 of CEUR Electronic Workshop Proceedings,
http://ceur-ws.org/, pages 10-19, 2001
[Carrillo and Franza, 2006] J.E. Carrillo and R.M. Franza, Investing in product development and production
capabilities: The crucial linkage between time-to-market and ramp-up time. European Journal of
Operational Research, 171(2):536 – 556, 2006
[Chang and Sze, 1996] C.Y. Chang and S.M. Sze, ULSI technology, McGraw-Hill, Materials science and
process technology series, 1996
[Cho and Hsu, 1997] J. Cho and C. Hsu, Manufacturing driven design, IEEE International Conference on
Systems, Man and Cybernetics, 1997
[Cliff, 2003] M. Cliff, DFM - An Industry Paradigm Shift, International Test Conference (ITC’03), 2003
[Codd, 1970] E.F. Codd, A relational model of data for large shared data banks. Communications of the
ACM, 13(6):377–387, 1970
[Colledani and Tolio, 2009] M. Colledani and T. Tolio, Performance evaluation of production systems
monitored by statistica process control and off-line inspections. International Journal of Production
Research, 120(2):348–367, 2009
[Colledani, 2008] M. Colledani, Integrated analysis of quality and production logistics performance in
asynchronous manufacturing lines. In IFAC World Congress, 2008
[Comai and Tena, 2007] A. Comai, and J. Tena, Early warning systems for your competitive landscape.
Society of Competitive Intelligence Professionals, 10(3), June 2007
[Corbett, 1991] J. Corbett, M. Dooner, J. Meleka, and C. Pym, Design for Manufacture Strategies, Principles
and Techniques, Addison-Wesley, 1991
[Corcho and Gomez-Perez, 2000] O. Corcho, and A.G. Perez, Evaluating knowledge representation and
reasoning capabilities of ontology specification languages. In Proceedings of the ECAI 2000 Workshop on
Applications of Ontologies and Problem-Solving Methods, Berlin, 2000
266
[Cottrell and Grebinski, 2003] D.R. Cottrell, T.J. Grebinski, Interoperability Beyond Design: Sharing
Knowledge between Design and Manufacturing, isqed, pp.214, Fourth International Symposium on Quality
Electronic Design, 2003
[Curé, 2005] O. Curé, Mapping Databases to Ontologies to Design and Maintain Data in a Semantic Web
Environment, Systemic, Cybernetics and Informatics Volume-4, number-4, 2005
[Coutinho, 1964] J.S. Coutinho, Failure effect analysis. Trans. NY Acad. Sci., 26(II):564–584, 1964
[Cui and Brien, 2000] Z. Cui, and P. O’Brien, Domain Ontology Management Environment. In Proceedings
of the 33rd Hawaii International Conference on System Sciences, 2000
[Cui and Brien, 2001] Z. Cui, D. Jones, and P. O’Brien, Issues in Ontology-based Information Integration.
IJCAI – Seattle, USA, 2001
[D]
[Dale, 2012] F. Dale, Intel’s semiconductor market share surges to more than 10-year high in 2011, March
2012
[Darmont et al., 2007] J. Darmont, F. Bentayeb, O. Boussaid, Benchmarking Data Warehouses, International
Journal of Business Intelligence and Data Mining, 2007
[Das and Kanchanapiboon, 2009] S. Das, and A. Kanchanapiboon, A multi-criteria model for evaluating
design for manufacurability. International Journal of Production Research, 49(4):1197–1217, 2009
[Dauzere-Peres et al., 2011] S. D. Peres, J.L. Rouveyrol, C. Yugma, and P. Vialletelle, A smart sampling
algorithm to minimize risk dynamically. In Advanced Semiconductor Manufacturing Conference, 2011
[Henderson, 1998] D. Henderson, Overcoming data integration problems, Info Management Direct, 1998
[Deridder and Wouters, 2000] D. Deridder, B. Wouters, The Use of an Ontology to Support a Coupling
between Software Models and Implementation, European Bibliography Conference on Object-Oriented
Programming (ECOOP’00), International Workshop on Model Engineering, 2000
[DFM Dictionary 2004] Si2, DFM Dictionary, http://www.si2.org/openeda.si2.org/dfmcdictionary/
index.php/Main_Page, retrieved on 12th july 2011
[Dhillon, 1999] B.S. Dhillon, Design Reliability, Fundamentals and Applications. CRC Press LLC, Florida
USA, 1999
[Doan and Halevy, 2005] A. Doan and A.Y. Halevy, Semantic-integration research in the database
community: a brief survey. AI Magazine, 26(1):83–94, 2005
[Dou and Lependu, 2006] D. Dou and P. LePendu, Ontology-based Integration for Relational Databases. In
Proceedings of ACM Symposium on Applied Computing (SAC), pages 461-466, 2006
[Dummer, 1997] G.W.A. Dummer, Electronics Inventions and Discoveries. Taylor & Francis, 1997.
[Duncan, 1956] A.J. Duncan, The economic design of X-bar charts used to maintain current control of a
process, Journal of the American Statistical Association, vol. 51(274), pp. 228-242, 1956
[E]
[Eskelinen, 2001] H. Eskelinen, Improving the productivity of complex electronic systems design by
utilizing applied design methodologies, IEEE AESS Systems Magazine, Vol. 16, Issue 10, 2001
[F]
[Fiedler et al., 2005] G. Fiedler, T. Raak, and B. Thalheim, Database collaboration instead of integration. In
APCCM'05, 2005
267
[G]
[Gilad, 2011] B. Gilad, (President Academy of Competitive Intelligence Boca Raton Florida USA).
Strategy without intelligence, intelligence without strategy. Business Strategy Series, 12(1):4–11, 2011
[Guarino, 2002] N. Guarino, Ontology-Driven Conceptual Modeling, Tutorial at 21st International
Conference on Conceptual Modeling (ER’02). Tampere, Finland, 2002
[Glaser et al., 2004] H. Glaser, H. Alani, L. Carr, S. Chapman, F. Ciravegna, A. Dingli, N. Gibbins, S.
Harris, M.C. Schraefel, N. Shadbolt, CS AKTive Space: Building a Semantic Web Application. In First
European Web Symposium (ESWS’04). Springer Verlag, pp. 417–432, 2004
[Goh, 1997] C.H. Goh, Representing and Reasoning about Semantic Conflicts in Heterogeneous Information
Sources. Phd, MIT, 1997
[Gruber, 1993] T. Gruber, A translation approach to portable ontology specifications. Knowledge Acquisition
1993 –Vol. 5(2), 199–220, 1993
[Guarino and Persidis, 2003] N. Guarino, A. Persidis, Evaluation Framework for Content Standards.
Deliverable 3.5, OntoWeb EU project (IST-2000-29243), 2003
[H]
[Heimbigner and McLeod, 1985] D. Heimbigner and D. McLeod, A Federated Architecture for Information
Management, ACM Transactions on Office Information Systems, Vol. 3, No. 3, pp. 253-278, July 1985
[Herrmann et al., 2004] J.W. Herrmann, J. Cooper, S.K. Gupta, C.C. Hayes, K. Ishii, D. Kazmer, P.A.
Sandborn and W.H. Wood, New Directions in Design for Manufacturing, Proceedings of DETC’04 ASME
2004 design engineering technical conferences and computers and information in engineering conference,
2004
[Hsu, 1977] J.I.S. Hsu, A cost model for skip-lot destructive sampling. IEEE Transactions on Reliability,
26(1):70–72, 1977
[Humphrey, 2005] A.S. Humphrey, SWOT analysis for management consulting, 2005 (newsletter) from
http://www.sri.com/about/alumni
[Hurat and Cote, 2005] P. Hurat and M. Cote, DFM for Manufacturers and Designers, 25th Annual BACUS
Symposium on Photomask Technology, Proceedings of SPIE, Volume 5992 pp. 126-132, 2005
[Hurat et al., 2006] P. Hurat, M. Cote, C.M. Tsai, J. Brandenburg, A Genuine Design Manufacturability
Check for Designers, Proc. of SPIE Vol. 6156, 2006
[Hurson and Bright, 1991] A.R. Hurson, and M.W. Bright, Multidatabase Systems: An Advanced Concept in
Handling Distributed Data. Advances in Computers, 32:149–200, 1991
[Hurtarte, 2007] J. S. Hurtarte, Understanding Fabless IC Technology. San Diego, CA, Elsevier, 2007
[Honma et al., 2009] M. Honma, H. Itoh, N. Iriki, S. Kobayashi, S. Kuwabara, N. Miyazaki, H. Suzuki, N.
Yoshioka, S. Arima and K. Kadota, Organized DFM for total optimization in semiconductor manufacturing,
AEC/ APC Symposium Asia 2009
[I]
[Impagliazzo, 2012] J. Impagliazzo. A brief history of database systems. Retrieved July 20th, 2012 from
http://www.comphist.org/computing_history/new_page_9.htm
[Inmon, 2005] W.H. Inmon, Building the Data Warehouse. Wiley; 4 edition (October 7, 2005), 2005
[International Business Strategies, 2009] International Business Strategies, Key trends in technology and
supply for advanced features within IC industry. Technical report, by International Business Strategies, Inc.,
ibs_inc@ix.netcom.com, October 2009
268
[Ireson et al., 1996] G.W. Ireson, C.F. Coombs, and R.Y. Moss, The Handbook of Reliability
Engineering and Management. McGraw-Hill New York, 1996
[J]
[James, 2009] D. James, Design-for-manufacturing features in nanometer logic processes – a reverse
engineering perspective, IEEE Custom Integrated Circuits Conference (CICC), 2009
[Jones, 2012] S.W. Jones, Introduction to integrated circuit technology, retrieved on 20th July 2012,
http://www.icknowledge.com/freecontent/Introduction%20to%20IC%20technology%20rev%205.pdf
[Johann et al., 2002] E. Johann, K. Christian and M. Tadeusz, The COMET Metamodel for Temporal Data
Warehouses, In Proc. of the 14th Int. Conference on Advanced Information Systems Engineering
(CAISE’02), 2002
[Jordan, 1972] W.E. Jordan, Failure modes, effects and criticality analyses. In Proc. Annu. Reliability
Maintainability Symp., pages 30–37, 1972
[K]
[Kahng, 2010] A.B. Kahng, Scaling: More than Moore’s law, IEEE Design & Test Computers, Vol. 27, pp.
86-87, 2010
[Kasem, 1997] M. Kasem, Integrated development and manufacturing methodology for advanced power
MOSFETs, IEE/CPMT Int'l Electronics Manufacturing Technology Symposium, 1997
[Kumar, 2008] R. Kumar, Fabless Semiconductor Implementation. USA, McGraw-Hill Professional, 2008
[Konstantinou et al., 2006] N. Konstantinou, D. Spanos, M. Chalas, E. Solidakis, and N Mitrou, VisAVis:
An Approach to an Intermediate Layer between Ontologies and Relational Database Contents. International
Workshop on Web Information Systems Modeling (WISM2006) Luxembourg, 2006
[Koren et al., 1999] Y. Koren, U. Heisel, F. Jovane, T. Moriwaki, G. Pritschow, G. Ulsoy and H. Van
Brussel, Reconfigurable Manufacturing Systems, Annals of the CIRP Vol. 48/2/1999, Retrieved from:
http://www-personal.umich.edu/~ulsoy/pdf/RMS_CIRP99_Keynote.pdf
[L]
[Lapedus, 2008] M. LaPedus, TSMC's 32-nm DFM model aims to unify design flow, retrieved from:
http://www.eetimes.com/showArticle .jhtml?articleID=208402578, 2008
[Landers and Rosenberg, 1982] T. Landers, and R.L. Rosenberg, An Overview of MULTIBASE. In Second
International Symposium on Distributed Data Bases, pages 153–184, Berlin, Germany, Holland, 1982
[Lee, 2002] J. Lee, Artificial intelligence-based sampling planning system for dynamic manufacturing
process. Expert Systems with Applications, 22(2):117–133, 2002
[Lefkow, 2012] C. Lefkow, Global electronics sales ‘to surpass 1 trillion in 2012’, January 2012
[Lenzerini, 2002] M. Lenzerini, Data integration: A theoretical perspective. In PODS, pages 233–246, 2002
[Liddle et al., 2003] S.W. Liddle, K.A. Hewett and D.W. Embley, An Integrated Ontology Development
Environment for Data Extraction, In Proceedings of the 2nd International Conference on Information System
Technology and its Applications (ISTA2003), Lecture Notes in Informatics, vol. P-30, pp. 21-33, 2003
[Liebmann, 2008] L. Liebmann, DFM, the teenage years. Proc. of SPIE, page 6925, 2008
[M]
[Maier et al., 2003] A. Maier, H. P. Schnurr and Y. Sure, “Ontology-based Information Integration in the
Automotive Industry”, In Proc. of ISWC 2003, pp. 897-912
[Mason, 2007] M.E. Mason, DFM EDA Technology A Lithographic Perspective, IEEE Symposium on VLSI
Technology, 2007
269
[Matteo et al., 2006] M.G. Stefano and L. Jens, R. Stefano, and V. Gottfried, Schema versioning in data
warehouses: Enabling cross-version querying via schema augmentation. Data & Knowledge Engineering,
59(2):435–459, 2006
[McGregor, 2007] J. McGregor, The common platform technology: A new model for semiconductor
manufacturing. Technical Report, 2007
[Mehrabi et al., 2002] M.G. Mehrabi, A.G. Ulsoy, Y. Koren, P. Heytler, Trends and perspectives in flexible
and reconfigurable manufacturing systems, Journal of Intelligent Manufacturing, 13, 135-146, 2002
[Mena et al., 1996] E. Mena, V. Kashyap, A. Sheth, and A. Illarramendi, Observer: An approach for query
processing in global information systems based on interoperability between pre-existing ontologies. In
Proceedings 1st IFCIS International Conference on Cooperative Information Systems (CoopIS ’96).
Brussels, 1996
[Mindtools, 2007] SWOT analysis: discover new opportunities. Manage and eliminate threats, Retrieved on
20th July 2012, from http://www.mindtools.com/pages/article/newTMC_05.htm.
[Moore, 1998] G.E. Moore, Cramming more components onto integrated circuits. Proceedings of the IEEE,
86(1):82–85, January 1998
[Morinaga et al., 2006] H. Morinaga, H. Kakinuma, T. Ito, and T. Higashiki, Development of a platform for
collaborative engineering data flow between design and manufacturing. In IEEE International Symposium on
Semiconductor Manufacturing, pages 45 –48, 2006
[Mouli and Scott, 2007] C. Mouli, and M. Scott, Adaptive metrology sampling techniques enabling higher
precision in variability detection and control. In Advanced Semiconductor Manufacturing Conference, 2007
[Myint and Tabucanon, 1998] S. Myint, and M.T. Tabucanon, The framework for an expert system to
generate alternative products in concurrent engineering design. The framework for an expert system to
generate alternative products in concurrent engineering design, 37(2):125–134, 1998
[N]
[Nyulas et al., 2007] C. Nyulas, M. O. Connor, and S. Tu. DataMaster – a plug-in for importing schemas and
data from relational databases into Protégé. In Proc. of Protégé 2007, Stanford Medical Informatics, 2007
[O]
[OMG, 1997] Object Management Group, Meta Object Facility (MOF) Specification, Joint Revised
Submission, OMG Document ad/97-08-14, 1997 retrieved from http://www.omg.org
[Ouyang et al., 2000] C. Ouyang, H. T. Heineken, J. Khare, S. Shaikh and M. d’Abreu, Maximizing wafer
productivity through layout optimizations, 13th International Conference on VLSI design, 2000
[P]
[Palopoli et al., 2000] L. Palopoli, G. Terracina and D. Ursino: The System DIKE: Towards the Semi-
Automatic Synthesis of Cooperative Information Systems and Data Warehouses. ADBIS-DASFAA
2000, Matfyzpress, 108-117
[Peters, 2005] L. Peters, DFM Worlds Collide, Then Cooperate, EETimes Magazine, 2005 retrieved from:
http//www. eetimes.com
[Porter, 1996] M. E. Porter, What is strategy? Harvard Business Review, pages 61–78, November-
December 1996
[Porter, 1998] M. E. Porter, Competitive advantage: creating and sustaining superior performance, PB: Free
Press, 1998
270
[Q]
[Quirke et al., 1992] M.T. Quirke, J.J. Barrett and M. Hayes, Planar Magnetic Component Technology A
review, IEEE Transactions on Components, Hybrids and Manufacturing Technology, Vol. 15, No. 5 (884-
892), 1992
[Quirk and Serda, 2000] M. Quirk, and J. Serda, Semiconductor Manufacturing Technology. Prentice Hall; 1
edition (November 19, 2000), 2000
[R]
[Raaij and Verhallen, 2005] V.W.F. Raaij, and T.M.M. Verhallen, Adaptive sampling methodology for in-
line defect inspection, 2005
[Radojcic et al., 2009] R. Radojcic, D. Perry and M. Nakamoto, Design for manufacturability for fabless
manufactuers. Solid-State Circuits Magazine, IEEE, 1(3):24–33, 2009
[Rahm and Bernstein, 2001] E. Rahm and P.A. Bernstein, A survey of approaches to automatic schema
matching. VLDB J., 10(4):334–350, 2001
[Ramakrishnan and Gehrke, 2003] R. Ramakrishnan and J. Gehrke: Database Management Systems 3rd
Edition, McGraw Hill, Chapter 3: The Relational Model, pages 57-99, 2003
[Ramana and Rao, 2002] K.V. Ramana and P.V.M. Rao, Automated manufacturability evaluation of sheet
metal components: a review, ASME International Design Engineering Technical Conferences and
Computers and Information in Engineering Conference (IDETC/ CIE2002), 2002
[Raina, 2006] R. Raina, What is DFM, DFY and why should i care? IEEE International Test Conference,
ITC-2006, Santa Clara, CA, USA, 2006
[Redford, 2009] M. Redford, J. Sawicki, P. Subramaniam, C. Hou, Y. Zorian and K. Michaels, DFM—Don't
Care or Competitive Weapon?, 46th ACM/IEEE Design Automation Conference (DAC’09), pp. 296-297,
2009
[Reid and Sanders, 2005] R.D. Reid and R.N. Sanders, Operations Management: An integrated approach.
2nd Edition © Wiley 2005
[Reddy, 2000] S.V. Reddy, Digital design flow options, Master Thesis submitted at Ohio State University,
retrieved from: http:// www.ohio.edu, 2000
[Reifer, 1979] D.J. Reifer, Software failure modes and effects analysis. IEEE Transactions on Reliability,
R-28(3):247–249, 1979
[Reynolds et al., 1988] M.R. Reynolds, R.W. Amin, J.C. Arnold, and J.A. Nachlas, X charts with variable
sampling intervals. Technometrics, 30(2):181–192, 1988
[Rezayat, 2000] M. Rezayat, Knowledge-based product development using xml and kcs. Computer Aided
Design, 32(5):299–309, 2000
[Ross, 2003] P.E. Ross, Commandments [technology laws and rules of thumb], Spectrum, IEEE,
40(12):30–35, 2003
[S]
[Sahnoun et al., 2011] M. Sahnoun, B. Bettayeb, P. Vialletelle, A. Mili, and M. Tollenaere, Impact of
sampling on w@r and metrology time delay, In Intel European Research & Innovation Conference, 2011
[Schuermyer et al., 2005] C. Schuermyer, K. Cota, R. Madge, B. Benware, Identification of systematic yield
limiters in complex ASIC through volume structural test fail data visualization and analysis, IEEE
International Test Conference Proceedings (ITC), 2005
271
[Shahzad et al., 2011a] M. K. Shahzad, S. Hubac, A. Siadat, and M. Tollenaere, An Extended IDM Business
Model to Ensure Time-to-Quality in Semiconductor Manufacturing Industry, volume 219 of
Communications in Computer and Information Science, pages 118–128. Springer Berlin Heidelberg, 2011.
10.1007/978-3-642-24358-5_12
[Shahzad et al., 2011b] M.K. Shahzad, S. Hubac, A. Siadat, and M. Tollenaere, An extended business model
to ensure time-to-quality in semiconductor manufacturing industry, In International Conference on
Enterprise Information Systems, 2011
[Sheth and Larson, 1990] A.P. Sheth and J.A. Larson, Federated Database Systems for Managing
Distributed, Heterogeneous, and Autonomous Databases, ACM Computing Surveys, Vol. 22, No. 3, pp. 183-
236, 1990
[Shingo and Dillion, 1989] S. Shingo and A. P. Dillion, A Study of the Toyota Production System,
Productivity Press, 1989
[Seino et al., 2009] T. Seino, S. Honda and T. Tanaka, Manufacturing technology management to accelerate
“design for manufacturability”, Portland International Conference on Management of Engineering &
Technology (PICMET) pp. 1137-1142, 2009
[Spanos, 1991] C.J. Spanos, Statistical process control in semiconductor manufacturing. Microelectronics
Engineering, 10(3-4):271–276, 1991
[Spyns et al., 2002] P. Spyns, D. Oberle, R. Volz, J. Zheng, M. Jarrar, Y. Sure, R. Studer, R. Meersman,
OntoWeb - a Semantic Web Community Portal. In: Karagiannis, D., Reimer, U., (eds.): Proceedings of the
Fourth International Conference on Practical Aspects of Knowledge Management (PAKM’02), LNAI 2569,
Springer Verlag. pp. 189–200, 2002
[Stamford, 2012] C. Stamford, Market share analysis: Total semiconductor revenue, worldwide, April 2012.
[Stoffel et al., 1997] K. Stoffel, M. Taylor and J. Hendler, “Efficient Management of Very Large
Ontologies”, Proceedings of American Association for Artificial Intelligence Conference, (AAAI-97),
AAAI/MIT Press, 1997
[T]
[Tzitzikas et al., 2001] Y. Tzitzikas, N. Spyratos, P. Constantopoulos, Mediators over Ontology-based
Information Sources. Proceedings of the Second International Conference on Web Information Systems
Engineering (WISE’01), 2001
[Treat et al., 1996] J. Treat, G. Thibault, and A. Asin, Dynamic competitive simulation: war gaming as a
strategic tool. Strategy and Business, pages 46–54, 1996
[Trybula, 1995] W. J. Trybula, The evolution of design for manufacturing, IEEE Technical Applications
Conference and Workshops (Northcon95), pp. 332-334, 1995
[Turakhia et al., 2009] R. Turakhia, M. Ward, S. K. Goel and B. Benware, Bridging DFM Analysis and
Volume Diagnostics for Yield Learning – A Case Study, 27th IEEE VLSI Test Symposium (VTS’09) pp. 167-
172, 2009
[Tuomi, 2004] I. Tuomi, Industrial structure and policy choice: Notes on the evolution of semiconductors
and open source. Technical report, Joint Research Centre, Institute for Prospective Technological Studies
Spain, 2004
[V]
[Venkataraman, 2007] S. Venkataraman, Does Test Have a Greater Role to Play in the DFM process, IEEE
International Test Conference (ITC), pp. 1-1, 2007
272
[Vliet and Luttervelt, 2004] H. W. V. Vliet, and K. V. Luttervelt, Development and application of a
mixed product/process-based DFM methodology. International Journal of Integrated Manufacturing,
17(3):224–234, 2004
[W]
[Wache et al., 2001] H. Wache, T. Vögele, U. Visser, H. Stuckenschmidt, G. Schuster, H. Neumann, and S.
Hübner, Ontology - based Integration of Information - A Survey of Existing Approaches. In: Proceedings of
IJCAI-01 Workshop: Ontologies and Information Sharing, Seattle, WA, 108-117, 2001
[Wand et al., 1999] Y. Wand, V. Storey, R. Weber, An Ontological Analysis of the relationship Construct in
Conceptual Modeling, ACM Transactions on Database Systems, Vol. 24, No. 4, pp. 494– 528, 1999
[Webb, 2008] C. Webb, Intel design for manufacturing and evolution of design rules. Intel Technology
Journal, 12(2), 2008
[Welty and Smith, 2001] C. Welty and B. Smith (eds.), Formal Ontology and Information Systems, New
York, ACM Press, 2001
[Weihrich, 1982] H. Weihrich, The TOWS matrix – A tool for situational analysis. Long Range Planning,
15:54–66, 1982 retrieved from http://www.usfca.edu/fac_staff/weihrichh/docs/tows.pdf
[White et al., 1995] K.P. White, R.N. Athay and W.J. Trybula, Applying DFM in the Semiconductor
Industry, 17th IEEECPMT International Electronics Manufacturing Technology Symposium, pp. 438-441,
1995
[White and Trybula, 1996] K.P. White, W.J. Trybula, DFM for the next generation, 19th IEEE/CPMT
Electronics Manufacturing Technology Symposium, pp. 109-116, 1996
[Winer, 1993] L. Winer, The Strategic Creative Analysis (SCAN) Process. Case studies Center for Applied
Research, Lubin School of Business, Pace University, 1993
[Wong et al., 2008] B.P. Wong, A. Mittal, G.W. Starr, F. Zach, V. Moroz, and A. Kahng, Nano-CMOS
Design for Manufacturability: Robust Circuit and Physical Design for Sub-65nm Technology Nodes. John
Wiley & Sons, 2008
[Worrall, 1998] L. Worrall, Strategic analysis: a scientific art, Wolverhampton Business School,
Management Research Centre, 1998
[Wu, 2003] S. L. Wu, Industry Dynamics within Semiconductor Value Chain IDM, Foundry and Fabless,
Technical report, USA, 2003
[Y]
[Yoon et al., 2010] M. Yoon, and F. Malerba, Technological interrelatedness, knowledge generality and
economies of scale in the evolution of firm boundaries: A history-friendly model of the fabless ecosystem. In
International Conference on Opening Up Innovation: Strategy, Organization and Technology at Imperial
College London Business School, 2010
[Youn and McLeod, 2006] S. Youn, and D. McLeod, (Eds.). Ontology Development Tools for Ontology-
Based Knowledge Management, In M. Khosrow-Pour (Ed.), Encyclopedia of e-commerce, e-government and
mobile commerce (pp. 858-864) Hershey, PA: Idea Group 2006
[Young and Murray, 2007] D. Young, S. Murray, Industrial Engineering FE Exam Preparation, Kaplan
Education, Dearborn Publishing, Inc. 2007
[Yuyin et al., 1996] S. Yuyin, Z. Bopeng, C. Fuzhi and M. Qingguo, A Knowledge-based Design For
Manufacture System, IEEE International Conference on Systems, Man and Cybernetics, pp. 1220-1224,
1996
273
[Z]
[Zeigler et al., 2000] B.P. Zeigler, H. Praehofer, and T.G. Kim, Theory of Modeling and Simulation,
Academic Press, San Diego, CA, 2000
[Zuling and Hongbing, 2005] K. Zuling and W. Hongbing, Implementation and Application of Ontology
Databases with User-Defined Rules (UDR) Supported, Proceedings of the First International Conference on
Semantics, Knowledge, and Grid, 2005
[Zuozhi and Jami, 2005] Z. Zuozhi, and S.J. Jami, Domain independent shell for DFM and its applications to
sheet metal forming and injection molding, Computer Aided Design,Vol.37, No. 9, pp 881-898, 2005
274
Vita
Mr. Muhammad Kashif Shahzad was born in small town of Kasur in Pakistan. He completed his
bachelor degrees in mechanical engineering and computer science from university of engineering
and technology, Lahore and Allama Iqbal Open University, Islamabad (1995-2000). He started his
career as Lecturer of Computer Science at Petroman Training Institute, Lahore (1999-2001) and
then he joined WAPDA (water & power development authority) as Assistant Director
(Programmer/System Analyst) where he spent 6 years and held diverse responsibilities ranging
from network design and deployment to the programming with 4-GL languages. Besides his
professional responsibilities, he also completed Master degree in total quality management (TQM)
from University of Punjab (2006). He decided to come to France in 2007 for PhD studies in
industrial engineering to effectively merge his knowledge and experience of mechanical
engineering, quality management and computer science. He completed his Master degree in
industrial engineering from ENSGI, INP Grenoble in 2008 and started PhD with
STMicroelectronics titled <<dynamic exploitation of production databases to improve DFM
methods in semiconductor industry>>. He has attended advanced training on engineering data
analysis (EDA) and modeling and design of experiments training (DOE) from MIT, USA and is
also certified green belt of design for six sigma (DFSS). His research interest includes DFM
methods, data warehouse modeling, system modeling, business process reengineering, capacity
optimization semiconductor manufacturing processes.
Permanent address (or email): 20, Yasrib Street, Akbar Shaheed Road, Kot Lakhpat, Lahore
Pakistan, shahzad347@gmail.com
275