Physically-Aware Analysis of Systematic Defects in Integrated Circuits Wing Chiu Tam Nvidia Corporation R. D. Blanton Carnegie Mellon University diagnosis data in the context of the design layout. The Editor’s notes: methodologies developed Design-induced systematic defects are serious threats to the semiconducinclude evaluating the effector industry. This paper develops novel techniques to identify and prevent tiveness of DFM (design-forsuch defects, which facilitate to evaluate the effectiveness of DFM rules and improve the manufacturing process and design for yield enhancement. manufacturability) rules [1] VQiang Xu, City University of Hong Kong in systematic-defect prevention, identifying and understanding systematic defects h THE RELENTLESS SCALING of CMOS technolo- [2] to guide process improvement effort, formulate gies tremendously decreases device sizes and in- new DFM rules [3] or identify targets for test [4], [5], creases complexity of the resulting circuits. As more and creating a framework for fast and accurate defect and more devices are placed in the silicon real simulation [6] to validate these and other emerging estate, our ability to observe and understand defects test/yield learning methodologies (e.g., [7]). Figure 1 shows how this work addresses is diminishing. IC test, traditionally used to screen problems (shaded bubbles with associated acrofor failures for quality control purposes, now serves an added function of understanding and monitoring nyms) in a yield improvement flow in the industrial defects in the process. By diagnosing and mining the setting. Conventionally, test structures and SRAM test fail data, valuable insight can be obtained con- testing are heavily used to debug and monitor the cerning the reasons that chips fail. After the defects manufacturing process for yield learning purpose, are identified and their failure mechanisms are as shown in the left subflow of Figure 1. By understood, actions can be taken to prevent their analyzing test-structure data and SRAM test results, future occurrence to improve yield. In addition, the fabrication process can be characterized to specific tests can be generated to screen for these improve design, test and yield. Unfortunately, both SRAM (being highly regular) and test structures defects to improve product quality. This paper addresses the problem of information (being limited in volume) cannot capture all layout extraction from volume diagnosis data in order to geometries. Therefore, increasing attention has provide feedback to the designers for product yield been turned to the analysis of IC failure data to improvement. More specifically, this work tackles supplement the yield learning process, as shown in issues in this area by performing analysis of IC the right subflow of Figure 1. Here the failure data can be either generated by simulation or gathered from the real fabrication facilities. The advantage of Digital Object Identifier 10.1109/MDT.2012.2211093 having virtual failure data is that the defect Date of publication: 02 August 2012; date of current version: characteristics are known, which enables evalua23 October 2012. tion and debugging of the follow-on methodologies. September/October 2012 Copublished by the IEEE CEDA, IEEE CASS, IEEE SSCS, and TTTC 0740-7475/12/$31.00 B 2012 IEEE 81 © 2012 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works. ITC Special Section DFM rule evaluation In this section, the methodology for DFM rule evaluation is described. Background and related work DFM (design for manufacturability) refers to a wide spectrum of activities that are performed on the design prior to manufacturing in order to improve yield. DFM is necessary in nanoscale technology because of the formidable challenges encountered in manufacturing. To have acceptable yield, potential manufacturing difficulties must be anticipated Figure 1. Yield-improvement flow developed in this work. and then remedied using appropriate design compensations. One approach is to Using the failure data, the effectiveness of the DFM remove/reduce the offending layout features that rules for systematic-defect prevention can be are susceptible to failures. To achieve this goal, evaluated, which allows designers to prioritize these features must be communicated to the future DFM rule adherence. Despite tremendous designers. The common approach is to use DFM DFM effort, systematic defects still can occur [8], rules. DFM rules are very similar to regular design [9]. Therefore, failure data should be analyzed for rules, except that many of them are often not identifying systematic defects as well. For any strictly enforced. This is because the cost-benefit systematic defects that are confirmed to signifi- tradeoff for adhering/violating a particular rule is cantly affect yield, new DFM rules can be formu- not well understood [11]. This prevents designers lated to prevent occurrence of the offending layout from reaping the full benefit of DFM. Prior work [4], features in future designs, a technique which [9], [12], [13] in this area focuses on identifying or augments DFM rule generation based on test testing for systematic issues, our work focuses on understanding the economic benefit of applying a structures [10]. The rest of the paper is organized as follows: DFM particular set of DFM rules and ranking them rule evaluation describes our DFM rule evaluation according to their relative contribution to yield methodology that we call Rule Assessment of enhancement. Defect-Affected Regions (RADAR). This is followed by a description of our approach for systematic Approach and uniqueness defect identification that we call Layout Analysis for To achieve this goal, a methodology (called Systematic Identification using Clustering (LASIC) in RADARVRule Assessment of Defect-Affected ReSystematic defect Identification. In Mixed-signal gions) has been developed to evaluate DFM rules defect simulation, we describe the defect simu- using diagnosis-implicated regions of failed ICs [1]. lation framework Simulation of Layout-Injected Figure 2 shows the flow diagram of the RADAR Defects for Electrical Responses (SLIDER) which is methodology. The flow starts with diagnosing a large used to generate populations of virtual failure data. number of IC failures. This diagnosis-driven apHere we use SLIDER-generated data to evaluate proach is substantially different from traditional LASIC and RADAR. Finally, this work is summarized approach in which test structures are used to evalin Summary. uate DFM rules [14], [15]. The outcome of diagnosis 82 IEEE Design & Test of Computers is a set of suspect signal lines or cells where the defect is believed to reside. The corresponding area in the layout is examined to collect measures of effectiveness (i.e., the number of times a rule is violated/adhered) for each DFM rule of interest. In addition, for control purposes, the nets for the rest of the circuit, where failure does not occur, are sampled and examined in the same way. Using the information collected, statistical analysis is then performed to determine: 1) if there is strong evidence to indicate that the violation of a DFM rule is associated with the observed failures; 2) which rule is more important in terms of their ability to prevent failures (e.g., if violations of rule A cause more failures than violations of rule B, then rule A is more important than rule B); and 3) the amount of yield loss that would be recovered if a certain percentage of violations of a specific rule is corrected. To detect association of rule violations and yield loss, hypothesis testing is used. This technique is commonly used in the medical communities to determine if certain chemicals/activities have an effect on the health of an individual (e.g., is smoking associated with lung cancer?). To determine the relative importance of a rule, the relative risk is estimated for each rule. In the context of RADAR, relative risk is simply the ratio of the failure rate given rule violation to the failure rate given rule adherence. From the definition, it is clear that relative risk measures the increase in likelihood of failure when a rule is violated. Thus, calculating the relative risks can reveal the relative importance of each rule. Additionally, the failure rate given rule violation is also estimated for all the DFM rules of interest using an EM (expectation-maximization) approach [16], [17]. After the relative risk and failure rate given rule violation are estimated, the failure rate given rule adherence can be estimated as well since relative risk is simply a ratio of the two failure rates. Having the estimates of these two failure rates, in turn, allows the estimation of yield loss before and after addressing a certain amount of rule violations. The difference measures the resulting yield impact of addressing the violation and thus can also be used as a measure to indicate importance of a rule in a given set of rules. Results and contributions RADAR helps the designers better understand the cost-benefit tradeoff and allows them to adjust future designs to improve yield at the same technology node. Table 1 shows the result of applying RADAR to a commercial Nvidia GPU that is manufactured using a TSMC 90 nm process. The chip is nearly 9 mm 9 mm, contains over 12 million gates and 700 K flip-flops, and is tested with over 4,000 test patterns. Nearly 9,000 failing GPUs are diagnosed to generate candidates for each failure. For this experiment, via-enclosure rules (e.g., M8.ENC. 1R, M7.ENC.2R) and metal-density rules (e.g., M23. WD.4R) are evaluated. The rule names are shown in column 1. Hypothesis testing using Pearson’s 2 test (columns 2) shows that a majority of the DFM rules Figure 2. Overview of the RADAR methodology. September/October 2012 83 ITC Special Section Table 1 Summary of DFM rule evaluation results for an Nvidia GPU design. is associated with yield loss. (A green entry in column 2 denotes association while a red entry denotes independence). The associated p-values (i.e., the probability of observing more extreme data than the current one) are shown in column 3. Both relative risk and yield impact are then calculated for each rule, which are shown in columns 4 and 5, respectively. The rules have been sorted in decreasing order of relative risk. The order shows that contact enclosure rules are more effective than density rules. This suggests that via issues may cause more yield loss than planarity/density problems for this particular process/design. It is also clear from columns 4 and 5 that the ordering of yield impact for a given set of rules is not the same as the ordering of relative risk. This is expected because yield impact takes into account the amount of rule violations specific to a design while relative risk does not. In practice, the number of rules of interest can be much more than what is used in this experiment and will include frontend rules as well. This experiment setup thus reduces the generality of the results. However, it should be emphasized that this is not a limitation of RADAR itself. Rather, this issue arises due to the unavailability of rules and failure data because of their confidential nature. Systematic defect identification In this section, the methodology for systematic defect identification is described. 84 Background and related work Defects inevitably occur in a semiconductor manufacturing process and can be classified into two categories, namely random and systematic defects. The former arise because of random contamination particles that affect the structure of a circuit, thus altering its functionality. In contrast, systematic defects happen because of design-process interactions, i.e., part of the design is sensitive to manufacturing in certain process corners and has an increased likelihood of failure. Unlike random defects, systematic defects can happen repeatedly in places with similar layout geometries. Identification of systematic defects immediately leads to yield improvement. Since systematic defects are becoming a significant source of failures in nanoscale technologies, the need to identify them becomes more important than ever. In the last several years, there has been a growing amount of work on systematic defect identification. In [12], [18], [19], expected failure rates are computed for layout features/nets. This information is combined with volume diagnosis data to identify outliers (i.e., systematic defects). Work in [20] applies clustering techniques to test-response data collected from a large number of failing ICs to identify common failure signatures. Last, the work described in [21], [22] performs extensive lithography simulation on the entire layout to identify hotspots. Layout snippets containing these hotspots are extracted and clustered to form DFM rules. The aforementioned work has the goal of identifying systematic defects, and undeniably, they have a certain degree of success, as evidenced by experiment results. Unfortunately, past approaches [12], [18]–[20] fall short of automatically extracting similar layout features that may underlie many of the failures. The work in [21], [22], while capable of automatic layout feature extraction, relies entirely on simulation, which is known to be inaccurate and too conservative. Relying on simulation alone also means that no systematic defects are identified for failing mechanisms that are not modeled. Approach and uniqueness To identify systematic defects, a diagnosis-driven methodology (called LASIC) is developed [2]. The LASIC flow is illustrated in Figure 3. LASIC uses the following four steps: 1) volume diagnosis; 2) layoutsnippet extraction; 3) snippet clustering; and 4) defect IEEE Design & Test of Computers validation. Similar to the RADAR flow, LASIC starts by applying diagnosis to a large sample of chips, or is continually applied to failures. The outcome of diagnosis is a set of suspect signal lines or cells where the defect is believed to reside. Snippet images are then extracted from the corresponding layout region for each suspect. Clustering, such as K-means, is then applied to group similar snippet images together for identifying any commonalities. (To speed Figure 3. Overview of the LASIC methodology. up the clustering process, the snippet images are represented using discrete cosine transform (DCT) so that only The layout snippets that correspond to the snipthe dominant DCT coefficients are used in the pet images are fed to Optissimo for lithography clustering process.) Layout features identified by simulation. Figure 5 shows an example of a LASIC can then be analyzed using a process lithography-induced hot spot (i.e., insufficient consimulator (such as the lithography tool Optissimo tact coverage) identified by Optissimo. There are [23]) to confirm the existence of a systematic defect, altogether 20 hotspots identified in the metal-1 layer. and/or physically analyzed using Physical Failure If these hotspots are confirmed to cause significant Analysis (PFA). yield loss, then DFM rules can be formulated to preIn contrast to [12], [18]–[20], LASIC identifies vent their future occurrence. For example, the insufproblematic layout features automatically. In addi- ficient coverage of the contact in Figure 5 can be tion, since the methodology analyzes real IC failures, resolved by increasing the metal-1-to-contact encloit does not suffer from the false negatives that may sure requirement. plague the approaches described in [21], [22]. Mixed-signal defect simulation Results and contributions The methodology has been implemented and tested on two industrial data. Specifically, two commercial designs, namely a LSI testchip in 130 nm technology and an Nvidia graphic processing unit (GPU) in 90 nm technology, with real failure data are used to validate the methodology. Figure 4 shows example images from two different metal-1 clusters after LASIC has been applied. Figure 4a shows four snippet images from a cluster in metal-1, while Figure 4b shows four snippet images from another metal-1 cluster. Figure 4 shows that geometries in the same cluster resemble each other but are not exactly the same while geometries in different clusters exhibit substantial differences. This example clearly demonstrates that LASIC is able to extract from diagnosis data images that have similar layout features. September/October 2012 In this section, the methodology for mixed-signal defect simulation is described. Figure 4. Illustration of clustered snippet images: (a) from one cluster and (b) a second cluster. 85 ITC Special Section Figure 5. Illustration of a hotspot identified using lithography simulation. Background and related work All test-fail driven yield/test learning methodologies should be validated to ensure their correctness. Unfortunately, this is easier said than done because of unavailability of test fail data and the corresponding PFA data, which are often foundry confidential. To circumvent this difficulty, the concept of virtual failure data creation was proposed in [24]. The idea is to simulate failures by injecting faults/defects in the circuit. The virtual data generated can then be used to evaluate the effectiveness and accuracy of the yield/test learning methodology under investigation. This approach also has the advantage that the ‘‘correct’’ answer (i.e., the injected defect) is known and can be precisely controlled and thus it enables debugging and refinement of the methodology under evaluation. Figure 6. Block diagram of the SLIDER framework. 86 Approach and uniqueness While previous work (e.g., [24]–[27]) has considered this topic before, the innovation here centers on three aspects: i) accuracy resulting from defect injection taking place at the layout level; ii) speedup resulting from careful and automatic partitioning of the circuit into digital and analog domains for mixed-signal simulation; and iii) complete automation that involves defect injection, design partitioning, netlist extraction, mixed-signal simulation, and test-data extraction [6]. The defect simulation framework developed is called SLIDER. Figure 6 shows the key components of SLIDER. It is composed of an analog block generator (ABG), a digital block generator (DBG), a mixed-signal simulator (MSS), and a result parser (RP). The blocks in pink are implemented in SLIDER. The MSS used in SLIDER is Cadence AMS designer [28], [29]; however, any other mixed-signal simulator can be easily used instead. The inputs to the framework are the design database, a description of the defects to be injected into the layout or a set of options to direct the defect generation process, and test patterns. The design database must contain both the layout and the logiclevel netlist. The defect description specifies the geometry of one or more defects (i.e., coordinates, size, polygon, layer, etc.). If the defect description is not specified, then options must be specified to direct the defect generation process. The supported defect types include: open, resistive via, signal bridge, supply bridge, and various kinds of cell defects that include the ‘‘nasty polysilicon defect’’ [30], transistor stuckopen, and transistor stuckclosed. The test patterns mimic the tester stimulus that would be applied to the circuit during production testing. Using these inputs, SLIDER starts by automatically partitioning the design into its respective analog and digital blocks, which is accomplished by ABG and DBG, respectively. The analog block contains all the cell instances and nets that are required to produce accurate circuit-level simulation of defect behavior, while the digital block IEEE Design & Test of Computers contains everything that remains. Defects are then injected into the analog block at the layout level. The modified layout is then extracted to produce the circuit-level netlist of the analog block. The analog block, the digital block, and their connections form the design under test (DUT), as illustrated in blue in Figure 6. A transient analysis is then performed on the DUT by the MSS using the test patterns as inputs. To capture the DUT’s response, RP samples the output waveform at the frequency defined by the test clock. During simulation, the MSS automatically takes care of the cross-domain signals through the use of connection modules [29], an analog-to-digital (digital-to-analog) convertor for a signal that crosses from the analog (digital) domain to the digital (analog) domain. Besides allowing the users to directly specify the defects directly, SLIDER supports several defectgeneration modes, as shown in Figure 7. SLIDER can generate defects that mimic random contaminations that follow a defect density and size distribution (DDSD) (Figure 7a). SLIDER can also generate systematic defects that share a common layout pattern (Figure 7b). In this mode, SLIDER provides a layout pattern (which can be user-defined or automatically generated) to a commercial pattern matching tool [31] to find all locations that mimic the given pattern. The identified locations then serve as defect injection sites as illustrated in Figure 7b. Finally, Figure 7c illustrates the third mode of SLIDER. This mode allows the user to explore different defectinjection scenarios at a given location for improving diagnosis. Results and contributions Because of the scalability enhancement, SLIDER can be applied in other settings that include diagnosis resolution improvement, defect localization, and fault model evaluation. Three set of experiments are performed to demonstrate scalability and usage of SLIDER. Run-time experiment. First, a run-time experiment is performed. Specifically, eleven benchmark circuits are placed-and-routed using Cadence Encounter. For each circuit, 100 random defects are layout-injected and simulated. The results are averaged and summarized in Figure 8. Figure 8a compares the total number of gates and the number of gates in the analog block for each circuit. It is clear that the size of the analog block does not scale, even when the circuit size increases tremendously. This property provides the basis for the scalability of SLIDER, i.e., the amount of effort needed for circuitlevel simulation remains relatively constant. Figure 8b shows the total runtime normalized by the pattern count. It is clear that the normalized runtime is in the order of seconds for all the benchmark circuits. Last, Figure 8c shows that SLIDER adds a negligible Figure 7. Three defect generation modes are supported in SLIDER that include: (a) random; (b) systematic; and (c) diagnostic. September/October 2012 87 ITC Special Section Figure 8. Runtime results that demonstrate the scalability of the SLIDER framework. overhead to the total runtime, which is dominated by the simulation time. LASIC evaluation. In the second experiment, the accuracy of LASIC is evaluated using the virtual failure data generated by SLIDER. Specifically, two failure populations, each of size 100 are generated for both random and systematic defects, using DDSD and systematic defect generation modes of SLIDER, respectively. The failures are then randomly sampled from both populations to form a single population. This process is repeated to get failure populations with different proportions of random and systematic defects. The goal is to evaluate how much random ‘‘noise’’ can be tolerated by LASIC. The experiment is performed using eight ISCAS benchmarks. Four types of defects are injected into these benchmarks: metal-2 bridge, metal-2 open, metal-3 bridge, and metal-3 open. The result is averaged over the eight benchmarks. In addition, to isolate the ‘‘noise’’ that occurs due to inaccuracies/ ambiguities in diagnosis, this experiment is performed for two scenarios: ideal and actual diagnosis. Ideal diagnosis means that diagnosis correctly pinpoints the faulty line with no ambiguity while 88 real diagnosis can be inaccurate (i.e., omitting the defect completely) or low in resolution (i.e., additional suspects are reported along with the actual defect location). Comparing the two scenarios provides insight about how LASIC performs under various scenarios. After applying LASIC to the virtual data, clusters of layout images of the suspect net are generated. The cluster that contains the underlying pattern of the injected systematic defects is called the correct cluster. For each scenario, the averaged rank of the correct cluster is plotted against the percentage of systematic defects injected. The data is summarized in Figure 9 for both ideal and real diagnoses. It is clear from Figure 9 that the average rank of the correct cluster improves (a smaller numeric value of the rank indicates that the cluster is ranked higher) as the percentage of systematic defects increases for both ideal and real diagnoses. This result is expected because the size of the correct cluster should increase as the number of systematic defects present increases. It is also evident that the technique is effective in both ideal and real diagnoses because the correct cluster is present in the top 40 ranked clusters (top 3%) even when the IEEE Design & Test of Computers population only consists of 20% systematic defects. As expected, performance of LASIC is worse for actual diagnosis due to inaccuracies/ambiguities in diagnosis. To mimic how LASIC should be used in practice, another experiment involving multiple systematic sources of defects and random defects are also performed. Specifically, five different systematic populations are generated to mimic the presence of five different sys- Figure 9. LASIC tematic issues for each benchmark circuit. The random population is also included to mimic the presence of random defects. The defects from these populations are randomly sampled and mixed in different proportions. Table 2 summarizes the amount of sampling from each population for each scenario (40 systematic defects are sampled from the ‘‘Systematic 2’’ population under the linear scenario, for example). The same sampling procedure is performed for each benchmark circuit. Four different scenarios are considered in this experiment that includes: 1) ‘‘Equal’’ (column 2) where the number of defects from each systematic population is equal; 2) ‘‘Linear’’ (column 3) where the number of defects is linearly decreasing for each systematic population; 3) ‘‘Dom1’’ (column 4) where one systematic population contributes substantially more defects than the rest; and 4) ‘‘Dom2’’ (column 5) where two of the systematic populations contribute substantially more defects than the rest. In each scenario, 70 random defects are included to mimic the presence of random defects (row 7 of Table 2). LASIC is applied to each of the four scenarios for each benchmark circuit. Here, the goal is to evaluate whether LASIC is able to identify all five systematic defect types. Therefore, the number of systematic issues correctly identified is plotted against the number of top-ranked clusters considered, which is varied from 5 to 55. In contrast to the previous experiment, only real diagnosis is used in this experiment in order to mimic how LASIC would be deployed in practice. Again, the experiment is repeated four times to inject four types of defects into September/October 2012 evaluation for (a) ideal and (b) real diagnoses. the benchmark circuits: metal-2 bridge, metal-3 bridge, metal-2 open, and metal-3 open. The results for each of the four defect types are averaged over the eight circuits and are shown in Figure 10a–d. It is clear from Figure 10 that the number of systematic issues correctly identified increases rapidly to the ideal value of five as the number of top-ranked clusters considered increases. As demonstrated in Figure 10, LASIC is able to identify over 90% (i.e., 4.5/5.0) systematic defects in the top 55 clusters for all defect types considered under all four scenarios. The results clearly show that LASIC is effective even when there are multiple systematicdefect types and random defects present. RADAR evaluation. In the third experiment, the accuracy of RADAR is evaluated using the SLIDERgenerated virtual failure data. A set of eight virtual DFM rules is handcrafted. In addition, for the evaluation of each DFM rule, two patterns are defined, namely the adherence pattern and the Table 2 Multiple systematic defect creation for LASIC evaluation. 89 ITC Special Section Figure 10. LASIC evaluation for multiple systematic defects for: (a) metal-2 bridge; (b) metal-3 bridge; (c) metal-2 open; and (d) metal-3 open. Figure 11. Hypothesis testing evaluation showing the trend for each DFM rule in terms of: (a) statistic T in Pearson 2 test and (b) the relative risk when the amount of violated-rule failures varies. 90 violation pattern. The adherence pattern defines the scenario when the rule is adhered, while the violation pattern defines the scenario when the rule is violated (i.e., the rule itself). Both patterns have to be defined for virtual failures to be created in both scenarios. Similar to the LASIC experiment, two failure populations, each of size 100, are generated for both adherence and violation patterns for each rule, using the systematic-defect generation mode of SLIDER. Again, the failures are then randomly sampled from both populations to form a single population. This process is repeated to generate failure populations with a different proportion of adherence and violation patterns. The goal is to evaluate how the statistic used in hypothesis testing and the relative risk change as the proportion of violated-rule defects varies. Seven ISCAS designs are used in the evaluation. The results are averaged over the designs. To examine how the effectiveness of RADAR changes as the amount of violated-rule failures varies, the statistic T (for Pearson 2 test) and the relative risk are plotted against the percentage of violated-rule failures in the failure population, as shown in Figure 11a and b, respectively. Figure 11a clearly shows that, as the percentage of violated-rule failures increases, the statistic T increases, indicating the increasing association of rule violations and failures, as expected. The associated p-value (not shown in Figure 11) also decreases, indicating the increasing confidence that the IEEE Design & Test of Computers association is valid. Figure 11b Table 3 Multiple violated-rule failures creation for RADAR evaluation. shows that the relative risk increases as the amount of violated-rule failure increases, which indicates that the failure rate given violation is increasingly higher than the failure rate given adherence, as expected. Similar to LASIC evaluation, another experiment involving different violated-rule failures is performed to evaluate the accuracy of the rule known, the exact yield impact can be evaluated to ranking. Using a similar experiment setup, several provide the ‘‘golden’’ reference point for comparison. different sets of virtual failure populations are Again, the results are averaged over the seven created and summarized in Table 3. Each row benchmark circuits used in this experiment. Figure 12 shows the results in the form of a norcorresponds to a scenario while each column specifies the number of injected rule-violation malized bar chart for the yield impact. Each bar in failures. For example, in the ‘‘Linear’’ scenario, 90 Figure 12 has a label in the y axis that consists of two violated-rule failures are selected for rule 1, 80 parts, namely the scenario and the data source. The failures are selected for rule 2, etc. In addition, there scenario refers to one of the failure scenarios defined are 80 adhered-rule failures (ten for each of the eight in Table 3 while the data source can be ‘‘golden,’’ rules) for each scenario. The adhered-rule failures ‘‘ideal diagnosis,’’ and ‘‘real diagnosis.’’ Figure 12 are labeled as unknown because in the estimation shows that the yield impact, agrees very well with the of Pk ðLjR 0 Þ, only rule-violations found in failing actual data for both ideal and real diagnosis. The layout regions are considered. The adhered-rule percentage error (i.e., the absolute difference failures instead of random failures are included to between the estimate value and the actual value ensure a nonzero entry in the adherence count, divided by the actual value) are 18.61% and 26.44% which is necessary for relative risk estimation. In this for ideal and real diagnoses, respectively. The experiment, five scenarios are considered and defined in a similar manner as in the case of LASIC evaluation experiment. Since failures due to different rule violations exist in the same population, diagnosis accuracy and resolution become very important. (A misdiagnosed candidate can lead to a rule violation being incorrectly attributed as the cause, for example.) To isolate the ‘‘noise’’ that can happen due to inaccuracies/ambiguities in diagnosis, this experiment is performed under two controls: ideal diagnosis and real diagnosis, which are defined in the same manner as the LASIC evaluation experiment. Since the exact numbers of violated-rule Figure 12. Rule ranking evaluation in terms of yield impact under and adhered-rule failures are five different failure scenarios. September/October 2012 91 ITC Special Section accuracy of yield impact is expectedly worse in the real-diagnosis scenario than in the ideal-diagnosis scenario because of the diagnosis inaccuracy. As a side note, the initial results were different from what are shown Figures 10 and 12 when LASIC and RADAR were first evaluated. Bugs were discovered and both techniques were improved. This would be difficult without the use of the virtual failure data generated by SLIDER. [8] B. Kruseman et al. ‘‘Systematic defects in deep sub-micron technologies,’’ in Int. Test Conf., 2004, pp. 290–299. [9] R. Turakhia et al. ‘‘Bridging DFM analysis and volume diagnostics for yield learningVA case study,’’ in Proc. IEEE VLSI Test Symp., 2009, pp. 167–172. [10] C. Tabery et al. ‘‘Process window and device variations evaluation using array-based characterization circuits,’’ in Proc. Int. Symp. Quality Electron. Des., THIS WORK DESCRIBES a comprehensive methodology that addresses the prevention and identification of systematic defects. Particularly, RADAR measures the effectiveness of the DFM rules in systematic defect prevention, and therefore enables designer to prioritize DFM rule adherence in future designs. Identification of systematic defects using LASIC enables an understanding of the systematic issues for improving the process and the design. The SLIDER framework enables fast and accurate defect simulation so that virtual failure data can be created to evaluate test/yield learning methodologies. 2006, pp. 6–265. [11] M. Nowak and R. Radojcic, ‘‘Are there economic benefits in DFM?’’ in Proc. Des. Automat. Conf., 2005, pp. 767–768. [12] J. Jahangiri and D. Abercrombie, ‘‘Value-added defect testing techniques,’’ IEEE Des. Test of Comput., vol. 22, no. 3, pp. 224–231, 2005. [13] K. Dongok et al. ‘‘Defect diagnosis based on DFM guidelines,’’ in Proc. IEEE VLSI Test Symp., 2010, pp. 206–211. [14] A. A. Kagalwalla et al. ‘‘Design-of-experiments based design rule optimization,’’ in Proc. SPIE, 2012, pp. 832709-8. Acknowledgment The authors would like to thank Mr. Bruce Cory from Nvidia for providing the GPU design and failure data, and the LSI Corporation for providing design and failure data. [15] Y. Zhang et al. ‘‘32 nm design rule and process exploration flow,’’ in Proc. SPIE, 2008, pp. 71223Z-12. [16] B. Benware et al. ‘‘Determining a failure root cause distribution from a population of layout-aware scan diagnosis results,’’ IEEE Des. Test of Comput. vol. 29, no. 1, pp. 8–18, 2012. [17] C. M. Bishop, ‘‘Mixture models and EM,’’ in Pattern h References [1] W. C. Tam and R. D. Blanton, ‘‘To DFM or not to DFM?’’ in Proc. Des. Automat. Conf., 2011. [2] W. C. Tam, O. Poku, and R. D. Blanton, ‘‘Systematic defect identification through layout snippet clustering,’’ in Proc. Int. Test Conf., 2010, pp. 13.2. [3] W. C. Tam, O. Poku, and R. D. Blanton, ‘‘Automatic DFM rule discovery through layout snippet clustering,’’ in Proc. Int. Workshop on Des. Manufacturabil. Yield, 2010. [4] K. Dongok et al. ‘‘Testing for systematic defects based on DFM guidelines,’’ in Proc. Int. Test Conf., 2007, pp. 1–10. [5] X. Yu et al. ‘‘Controlling DPPM through volume diagnosis,’’ in Proc. IEEE VLSI Test Symp., 2009, pp. 134–139. [6] W. C. Tam and R. D. Blanton, ‘‘SLIDER: A fast and New York: Springer-Verlag, 2006, pp. 423–460. [18] K. Martin et al. ‘‘A rapid yield learning flow based on production integrated layout-aware diagnosis,’’ in Proc. Int. Test Conf., 2006, pp. 1–10. [19] M. Sharma et al. ‘‘Efficiently performing yield enhancements by identifying dominant physical root cause from test fail data,’’ in Proc. Int. Test Conf., 2008, pp. 1–9. [20] L. M. Huisman, M. Kassab, and L. Pastel, ‘‘Data mining integrated circuit fails with fail commonalities,’’ in Proc. Int. Test Conf., 2004, pp. 661–668. [21] J. Ghan et al. ‘‘Clustering and pattern matching for an automatic hotspot classification and detection system,’’ in Proc. SPIE, 2009, pp. 727516-11. [22] N. Ma et al. ‘‘Automatic hotspot classification using accurate defect simulation framework,’’ in Proc. IEEE pattern-based clustering,’’ in Proc. SPIE, 2008, VLSI Test Symp., 2011, pp. 172–177. pp. 692505-10. [7] Y.-T. Lin et al. ‘‘Physically-aware N-detect test pattern selection,’’ Des. Autom. Test Eur., pp. 634–639, 2008. 92 Recognition and Machine Learning, 1st ed. [23] The Optissimo User Manual, 6.0 ed. San Jose, CA: PDF Solutions Inc., 2001. IEEE Design & Test of Computers [24] T. Vogels et al. ‘‘Benchmarking diagnosis algorithms with a diverse set of IC deformations,’’ in Proc. Int. Test Conf., 2004, pp. 508–517. [25] A. Jee and F. J. Ferguson, ‘‘Carafe: An inductive fault analysis tool for CMOS VLSI circuits,’’ in Proc. IEEE VLSI Test Symp., 1993, pp. 92–98. Wing Chiu (Jason) Tam is a senior hardware engineer with Nvidia Corporation working on inhouse DFT tools development. His research interests are in the areas of CAD for design for manufacturing and IC testing. He is a student member of the IEEE and ACM. [26] W. Meyer and R. Camposano, ‘‘Fast hierarchical multi-level fault simulation of sequential circuits with switch-level accuracy,’’ in Proc. Des. Automat. Conf., 1993, pp. 515–519. [27] W. C. Tam, O. Poku, and R. D. Blanton, ‘‘Automated failure population creation for validating integrated circuit diagnosis methods,’’ in Proc. Des. Automat. Conf., 2009, pp. 708–713. [28] The Virtuoso AMS Designer Environment User Guide. San Jose, CA: Cadence Design Systems Inc., 2009. R. D. (Shawn) Blanton is a professor of electrical and computer engineering at Carnegie Mellon University, Pittsburgh, PA, where he also serves as a Director at the Center for Silicon System Implementation, and leader of the Advanced Chip Testing Laboratory (ACTL). His research interests include test and diagnosis of integrated heterogeneous systems. [29] K. Kundert and O. Zinke, The Designer’s Guide to Verilog-AMS, 1st ed. New York: Springer, 2004. [30] R. D. Blanton et al. ‘‘Fault tuples in diagnosis of deep-submicron circuits,’’ in Proc. Int. Test Conf., 2002, pp. 233–241. [31] The Calibre Pattern Matching User Manual, 1.0 ed. Wilsonville, OR: Mentor Graphics Corporation, 2010. September/October 2012 h Direct questions and comments about this article to Wing Chiu Tam, Nvidia Corporation, Nvidia Corporation Building D, 2701 San Tomas Expressway, Santa Clara, CA 95050; phone (408)-486-4981; jtam@nvidia.com. 93