SlideShare a Scribd company logo
1 of 10
Solutia
From: Richard W. Miller (850) 968-8358 Fax: (850) 968-8732
“Solutia Fellow” [rwmill@solutia.com]
Date: September 24, 2008
Subject: Analysis of the 41LDC Manufacturing Process and the AEG Measurement System with an
Emphasis on Resampling Out-of-Spec Lots
Summary
Greenwood’s (GWD) AEG measurement system, validated utilizing the 5/3 strategy (five
routine samples split into thirds with their variance estimates pooled), accounted for ~12% of
41LDC’s total AEG variation. While this process was judged capable (Ppk=0.99), the long-term
component of variation was significantly greater than its short-term component. Finally, the routine
re-measure of out-of-spec material is a good strategy when the measurement system variation
component is small enough; otherwise, the β risk (customer risk of accepting out-of-spec material)
increases dramatically.
Jeff Wang, Gordon Mayfield and I are in the process of setting up continuous measurement
system validation for important polymer characteristics (AEG, RV, TiO2, H2O) at Greenwood.
This report summarizes the AEG characterization of 41LDC made on CP 45 between 6/11 &
8/19/08. In addition to summarizing the process data, a probabilistic argument for resampling out-
of-spec material is presented.
Major Observations & Recommendations
• AEG measurement system accounted for ~12% of GWD’d 41LDC (CP 45) variability;
• While GWD’s 41LDC manufacturing process is capable (Ppk=0.99 for in process
sampling), its capability could be improved significantly (Ppk=1.34) by reducing long-term
variation: campaign-to-campaign centering and variation fluctuation, and general process
drift;
• Initiate measurement system validation (GWD and Pensacola(PNS)) for other important
product properties (e.g., RV, TiO2 and H2O for 41LDC);
• Assess risks of out-of-spec sample re-measure through estimation of each measurement
system’s sigma; a sample measuring USL+1σ Meas Sys or greater than LSL - 1σ Meas Sys
will have a 16% chance of falling within specs on re-measure (even without analytical
error), a 32% chance on two re-measures, a 48% chance on three re-measures;
• Not all in-spec measurement distributions are equal; the level of variation can be more
important than the targeting because customer satisfaction is maximized at the minimum
in the Taguchi Loss Function {min(TLF) = min (σLT
2
+ (Target – μ)
2
};
• When a customer identifies a spec range as USL – LSL, that means the producer’s
effective spec range must be redefined tighter as USLeff=USL-xσ Meas Sys and
LSLeff=LSL+xσ Meas Sys where x might vary between 0 and 6 depending on the level of
customer risk (calling out-of-spec product good) desired.
Greenwood’s 41LDC AEG Process Capability
Chart 1 plots single AEG measures made over this two-month production period. While
this is in-process data—data used to control the CP line—the pack-out (P/O) characteristics were
similar. Red lines mark the population mean and upper and lower control limits (μ +/- 3σST), while
the blue lines mark specification limits (45 +/- 1.2). The first represents the “voice of the process”;
the second restrictions imposed on the process—ideally by customer needs. One can expect
99.73% of the population to fall between the outside red lines when there’s no long-term
component of variation (process drift or campaign-to-campaign centering). When a point falls
beyond the red lines, it’s deemed a special cause, i.e., likely to come from a different population
because of some special cause.
Often, the special cause for an outlier data can be tracked down and changes made to re-
center a shifted process, but just as often the reasons for the shift go unresolved and
compensatory adjustments are made anyway. The BCF process is a good example of the latter
where temperature adjustments are routinely made to re-center positions that continually drift as
process components age and foul.
There’s also the very real possibility of a questionable measurement, either because of
outright error or because of the distributional nature of the measurement system itself. Because
this producer risk (risk of off-grading acceptable product) is often judged more important to the
producer than customer risk (risk of not off-grading unacceptable product), off grade material is
given every opportunity to prove itself before being committed to a rework strategy. The question
then becomes: If the same lot is re-measured how likely is the new measure to fall within specs?
If a point falls beyond the specification limits (beyond the blue lines), especially for P/O data, it’s
common practice to re-measure that lot before reworking it to meet specifications. I’ll address the
resample question and its risk more fully in a later section.
This process’ capability matrix is shown in Chart 2. Specification limits, population mean
and its two measures of variation (short-term sigma and long-term sigma) are used to calculate
the capability metrics (Ppk, Pp, Cpk, Cp and % out-of-spec material) as defined in the chart. The
first column houses capability estimates for the current process, while the other three columns
calculate these metrics for other choices of specification limits. Note, the basic difference
between the two product capability metrics (Ppk and Pp) and the two process capability metrics
(Cpk and Cp) lies in the nature of their respective variation components: the first uses long-term
sigma and characterizes the product the customer receives, while the second uses short-term
sigma and characterizes the ultimate capability of the process.
These metrics confirm Chart 1’s picture: GWD’s 41LDC process is pretty capable (Ppk
nearly 1.0) with blue specification lines well outside the red control ones. The difference between
Ppk and Pp (and Cpk and Cp) indicates a slight centering problem, while the difference between
Ppk and Cpk indicates a significant long-term variation component. If this process could be re-
centered and its long-term component of variation reduced, it’s capable of producing a 1.49 Ppk
(making virtually nothing out-of-spec), i.e., spec range / 6σ = 1.49 means the spec range would
contain ~9 σLTs. Note, a Ppk=1 means the spec range would contain 6 σLTs.
5/3 Strategy: AEG Measurement System Validation
The following table summaries AEG data collected from the 5/3 testing where five routine
samples were split into thirds then characterized blindly on three different shifts. The idea is to
get routine measurement system data which reflect true lab variation. Assuming the subsample
compositions are identical (a good blend), their pooled variance estimate provides our estimate of
the measurement system’s variance. Here, σ
2
Meas Sys is estimated to be 0.016 for the AEG
measure, which in turn, accounts for ~12% of the total variation (σ
2
LT) evident in the Chart 1 data.
Note, the 5/3 strategy follows this initial 2-week sampling with another single sample split
every two weeks. Continuous validation of the AEG measurement system is thus provided at the
cost of a single extra sample each week.
Chart 1 from a Probabilistic View and the Resampling Question
The data presented in Chart 1 can also be looked at through the prism of probabilistic
models, which then enable us to ask questions: How does the difference between short-term and
long-term variance impact the frequency of control limit (μ+/-3σST) and specification limit
violations? How does a process shift impact quality? What is the impact of resampling a lot after
it has experienced an out-of-spec signal the first time? And finally, how does measurement
system variation influence the risk of resampling? Note, this modeling assumes the data are
normally distributed—a good assumption for these data.
The black distribution shown in Chart 3 depicts the probability distribution represented in
the AEG data plotted sequentially in Chart 1 (inset in Chart 3 as well). It describes the frequency
of AEG measures as a function of AEG level. The red distribution is that black distribution
hypothetically shifted to the high side by two long-term standard deviations, and the blue
distribution is the measurement system’s distribution expected for a lot which has measured an
out-of-spec 46.5 AEG in its routine sampling. The green vertical lines mark the process’ target
and specification limits, while the blue vertical lines mark control limits for the black distribution.
Note, the black distribution, called process State A is slightly off center to the high side of target
(hence the Ppk-Pp capability differences).
Chart 4 shows the percentage of each of the above distribution’s population that lies either
below or above the specification, control and target limits. A further distribution, the measurement
system centered at the upper specification limit, is included as well. Several points become
clearer.
First, short-term sigma are used in Shewhart charts to define control limits (μ +/- 3σST)
where the sigma are calculated for single-point measures from a 2-point moving range and for
multi-point averages from their within sample group variances. The State A distribution, on the
other hand, reflects the long-term sigma variation. It’s the variation one gets by throwing all 174
data in Chart 1 into a hat and calculating its variance. The net effect, since 3σST <= 3σLT, is that
the routine data can be expected to violate its μ+3σST limits more frequently than expected. For
example here, State A violates the lower control limit (LCL) at a 1.40% frequency (1 in 71 data)
and the upper control limit (UCL) at a 1.38% frequency (1 in 72 data); whereas a typical Shewhart
chart is cited to violate either control limit at a 0.135% frequency (1 in 740 data). Obviously, the
producer’s risk (calling good product bad) increases appreciably when σLT >>> σST.
The second question asks how a process shift (the red distribution, a +2 σLT, in Chart 3)
impacts quality. First, is there really any difference between two lots measuring a 46.75 AEG
where the first one comes from un-shifted (black) distribution and the second from the shifted
(red) distribution? No, as far as we can measure, the two are identical. What’s different is the
shifted distribution can be expected to produce 17.09% of its lots above the upper spec limit,
while the original distribution can be expected to place only 0.16% there. That’s why one should
be concerned about a shifted process. Note too, the shifted process is far more likely to produce
a runs-rule signal, e.g., the next 8 points in the black distribution would be expected to fall on one
side of its mean 0.39% of the time {(0.5)^8} or 1 time in 256 groups; whereas, the shifted red
distribution would be expected to signal that 8-point grouping to the high side of the original (black
distribution) mean 83.2% of the time.
The final two questions focus on the measurement system. Now that we can estimate the
measurement system’s variance, we can better assess both the risk of a re-measure and our true
spec limits. Remember, we re-measure a given lot of the red or black distribution in Chart 3;
whereas both of these distributions reflect the breadth of their respective lot measurements. The
blue distribution in Chart 3 outlines the frequency of AEG measures that one could expect from
just taking the original sample (AEG=46.5) and re-measuring it a number of times. Note, 46.5 is
2.4σMeas Sys above the 45.93 USL. We’re assuming there’s no difference in the sample, just this
level of variation in the measurement system itself (similar to the assumption made in the 5/3
sampling strategy). In the particular measurement distribution shown, virtually 99.18% of the
distribution lies above the 46.2 USL so any re-measure is likely to remain out-of-spec. Of course,
that assumes that no error occurred in the original measure. That means if the re-measure falls
within spec limits, there’s likely been an error made in the original measure.
Let’s look at another sample. Say we measure AEG for a lot at 46.2, exactly on the USL
and decide to resample. Given the variation in the measurement system, a single re-measure will
produce an out-of-spec signal 50% of the time and an in-spec signal 50% of the time. Moreover,
if I were to resample two times, there would be a 75% chance of getting an in-spec signal, three
times a 87.5% chance. In this case, producer’s risk (calling good product bad) has been judged
far more important than customer’s risk (calling bad product good). So now let’s reformulate spec
limits to reflect the measurement system’s variation.
Assume that customer’s risk is most important to us. To ensure that no product leaves the
building outside spec limits, one needs to use effective spec limits: USLeff = USL - 3σMeas Sys, here
45.83 (46.2 – 3x0.125) and LSLeff = LSL + 6σMeas Sys or 44.18 (43.8+3x0.125). Obviously,
managing variation is critical to quality.
Variation and Customer Preferences
Traditionally, quality has been viewed as a step function where product is judged either
good or bad depending on whether it falls within specification limits or not. Taguchi put forth the
idea that the customer becomes increasingly dissatisfied as a product drifts from target, even if
that product stays within spec limits. The customer desires a product that is constant part-to-part,
lot-to-lot or bobbin-to-bobbin. The producer, on the other hand, desires to make and sell as much
A-grade material as possible. Each lot contributes to the (Target – μlot)
2
, while their aggregate
variation determines the σLT
2
term.
Customer Satisfaction: Taguchi Loss Function {min(TLF) = min (σLT
2
+ (Target – μlot)
2
}
Clearly, in some cases—like the two distributions in the chart below—variance (even with in-spec
material) is more important to customer satisfaction than targeting.
The following chart further demonstrates the point. In these two distributions is there really
any difference among lots A, B and C? Don’t all differ significantly from lot D? Which of the two
distributions do you think will produce more consistent product for the customer? The answers
are obvious: the right-side distribution, though it has a targeting problem, would both be easier to
use by the customer and easier to control by the producer. The left-side distribution has a
variation problem. Even though everything falls inside spec limits, it will be harder to use by the
customer.
This is essentially the problem that the low-melt yarn program has run into over the past
few years. EMS produces a low melt resin that met melt viscosity specs, though their lot-to-lot
variation looked much like the left-side distribution below. The problem came when we (the
customer) tried to spin their resin: those lots from the high-side of the distribution required far
different (sometimes unknown) process settings than those lots from the low-side of the
distribution required. As a consequence, much of the yarn ends up on the floor as the process re-
equilibrates, and yields suffered. Again, the message is clear: variation is critical to quality.
080924 Measurement System Analysis Re Sampling

More Related Content

What's hot

Statistical process control
Statistical process controlStatistical process control
Statistical process control
jsembiring
 
AIAA-Aviation-2015-Mehmani
AIAA-Aviation-2015-MehmaniAIAA-Aviation-2015-Mehmani
AIAA-Aviation-2015-Mehmani
OptiModel
 
PEMF2_SDM_2012_Ali
PEMF2_SDM_2012_AliPEMF2_SDM_2012_Ali
PEMF2_SDM_2012_Ali
MDO_Lab
 
KnowledgeFromDataAtScaleProject
KnowledgeFromDataAtScaleProjectKnowledgeFromDataAtScaleProject
KnowledgeFromDataAtScaleProject
Marciano Moreno
 

What's hot (20)

A parsimonious SVM model selection criterion for classification of real-world ...
A parsimonious SVM model selection criterion for classification of real-world ...A parsimonious SVM model selection criterion for classification of real-world ...
A parsimonious SVM model selection criterion for classification of real-world ...
 
Tuning the model predictive control of a crude distillation unit
Tuning the model predictive control of a crude distillation unitTuning the model predictive control of a crude distillation unit
Tuning the model predictive control of a crude distillation unit
 
Publications
PublicationsPublications
Publications
 
Statistical process control
Statistical process controlStatistical process control
Statistical process control
 
D05842227
D05842227D05842227
D05842227
 
Spc
SpcSpc
Spc
 
AIAA-Aviation-2015-Mehmani
AIAA-Aviation-2015-MehmaniAIAA-Aviation-2015-Mehmani
AIAA-Aviation-2015-Mehmani
 
Analytical control strategy 1
Analytical control strategy 1Analytical control strategy 1
Analytical control strategy 1
 
Robust design of a 2 dof gmv controller a direct self-tuning and fuzzy schedu...
Robust design of a 2 dof gmv controller a direct self-tuning and fuzzy schedu...Robust design of a 2 dof gmv controller a direct self-tuning and fuzzy schedu...
Robust design of a 2 dof gmv controller a direct self-tuning and fuzzy schedu...
 
Statistical Process Control Part 2
Statistical Process Control Part 2Statistical Process Control Part 2
Statistical Process Control Part 2
 
Optimum penetration depth of cantilever sheet pile walls in dry granular soil
Optimum penetration depth of cantilever sheet pile walls in dry granular soil Optimum penetration depth of cantilever sheet pile walls in dry granular soil
Optimum penetration depth of cantilever sheet pile walls in dry granular soil
 
Statistical Process Control Part 1
Statistical Process Control Part 1Statistical Process Control Part 1
Statistical Process Control Part 1
 
Qa.spc
Qa.spcQa.spc
Qa.spc
 
PEMF2_SDM_2012_Ali
PEMF2_SDM_2012_AliPEMF2_SDM_2012_Ali
PEMF2_SDM_2012_Ali
 
Statistical process control (spc)
Statistical process control (spc)Statistical process control (spc)
Statistical process control (spc)
 
Boost model accuracy of imbalanced covid 19 mortality prediction
Boost model accuracy of imbalanced covid 19 mortality predictionBoost model accuracy of imbalanced covid 19 mortality prediction
Boost model accuracy of imbalanced covid 19 mortality prediction
 
Analytical control strategy 4
Analytical control strategy 4Analytical control strategy 4
Analytical control strategy 4
 
5. spc control charts
5. spc   control charts5. spc   control charts
5. spc control charts
 
Analytical control strategy 2
Analytical control strategy 2Analytical control strategy 2
Analytical control strategy 2
 
KnowledgeFromDataAtScaleProject
KnowledgeFromDataAtScaleProjectKnowledgeFromDataAtScaleProject
KnowledgeFromDataAtScaleProject
 

Viewers also liked (7)

Acs Awards Talk
Acs Awards TalkAcs Awards Talk
Acs Awards Talk
 
080324 Miller Opportunity Matrix Asq Presentation
080324 Miller Opportunity Matrix Asq Presentation080324 Miller Opportunity Matrix Asq Presentation
080324 Miller Opportunity Matrix Asq Presentation
 
080811 Miller Inventory Pilot Study
080811 Miller Inventory Pilot Study080811 Miller Inventory Pilot Study
080811 Miller Inventory Pilot Study
 
Seo010610
Seo010610Seo010610
Seo010610
 
070927 Miller Low Melt Us Patent7261849
070927 Miller Low Melt Us Patent7261849070927 Miller Low Melt Us Patent7261849
070927 Miller Low Melt Us Patent7261849
 
Dr BMN - Finishes for appearance__hand_and_performance
Dr BMN - Finishes for appearance__hand_and_performanceDr BMN - Finishes for appearance__hand_and_performance
Dr BMN - Finishes for appearance__hand_and_performance
 
090528 Miller Process Forensics Talk @ Asq
090528 Miller Process Forensics Talk @ Asq090528 Miller Process Forensics Talk @ Asq
090528 Miller Process Forensics Talk @ Asq
 

Similar to 080924 Measurement System Analysis Re Sampling

Cpk guide 0211_tech1
Cpk guide 0211_tech1Cpk guide 0211_tech1
Cpk guide 0211_tech1
Piyush Bose
 
Comparison of PID Controller with Model Predictive Controller for Milk Pasteu...
Comparison of PID Controller with Model Predictive Controller for Milk Pasteu...Comparison of PID Controller with Model Predictive Controller for Milk Pasteu...
Comparison of PID Controller with Model Predictive Controller for Milk Pasteu...
journalBEEI
 
Assessment of a diagnostic procedure for the monitoring and control of indust...
Assessment of a diagnostic procedure for the monitoring and control of indust...Assessment of a diagnostic procedure for the monitoring and control of indust...
Assessment of a diagnostic procedure for the monitoring and control of indust...
Manuel Stefanato
 
Maintainability analysis of an offshore gas compression train system: A case ...
Maintainability analysis of an offshore gas compression train system: A case ...Maintainability analysis of an offshore gas compression train system: A case ...
Maintainability analysis of an offshore gas compression train system: A case ...
Irfan Ullah Khan
 
Hybrid Stochastic Search Technique based Suboptimal AGC Regulator Design for ...
Hybrid Stochastic Search Technique based Suboptimal AGC Regulator Design for ...Hybrid Stochastic Search Technique based Suboptimal AGC Regulator Design for ...
Hybrid Stochastic Search Technique based Suboptimal AGC Regulator Design for ...
Dr. Omveer Singh
 
Computational intelligence systems in industrial engineering
Computational intelligence systems in industrial engineeringComputational intelligence systems in industrial engineering
Computational intelligence systems in industrial engineering
Springer
 
Six sigma-in-measurement-systems-evaluating-the-hidden-factory (2)
Six sigma-in-measurement-systems-evaluating-the-hidden-factory (2)Six sigma-in-measurement-systems-evaluating-the-hidden-factory (2)
Six sigma-in-measurement-systems-evaluating-the-hidden-factory (2)
Bibhuti Prasad Nanda
 

Similar to 080924 Measurement System Analysis Re Sampling (20)

Cpk guide 0211_tech1
Cpk guide 0211_tech1Cpk guide 0211_tech1
Cpk guide 0211_tech1
 
Comparison of PID Controller with Model Predictive Controller for Milk Pasteu...
Comparison of PID Controller with Model Predictive Controller for Milk Pasteu...Comparison of PID Controller with Model Predictive Controller for Milk Pasteu...
Comparison of PID Controller with Model Predictive Controller for Milk Pasteu...
 
100018718.ppt
100018718.ppt100018718.ppt
100018718.ppt
 
TOCbw I&ECPDD Oct67
TOCbw I&ECPDD Oct67TOCbw I&ECPDD Oct67
TOCbw I&ECPDD Oct67
 
Principal component analysis in modelling
Principal component analysis in modellingPrincipal component analysis in modelling
Principal component analysis in modelling
 
COMPARATIVE ANALYSIS OF CONVENTIONAL PID CONTROLLER AND FUZZY CONTROLLER WIT...
COMPARATIVE  ANALYSIS OF CONVENTIONAL PID CONTROLLER AND FUZZY CONTROLLER WIT...COMPARATIVE  ANALYSIS OF CONVENTIONAL PID CONTROLLER AND FUZZY CONTROLLER WIT...
COMPARATIVE ANALYSIS OF CONVENTIONAL PID CONTROLLER AND FUZZY CONTROLLER WIT...
 
Assessment of a diagnostic procedure for the monitoring and control of indust...
Assessment of a diagnostic procedure for the monitoring and control of indust...Assessment of a diagnostic procedure for the monitoring and control of indust...
Assessment of a diagnostic procedure for the monitoring and control of indust...
 
P Chart Tutorial
P Chart TutorialP Chart Tutorial
P Chart Tutorial
 
Maintainability analysis of an offshore gas compression train system: A case ...
Maintainability analysis of an offshore gas compression train system: A case ...Maintainability analysis of an offshore gas compression train system: A case ...
Maintainability analysis of an offshore gas compression train system: A case ...
 
Hybrid Stochastic Search Technique based Suboptimal AGC Regulator Design for ...
Hybrid Stochastic Search Technique based Suboptimal AGC Regulator Design for ...Hybrid Stochastic Search Technique based Suboptimal AGC Regulator Design for ...
Hybrid Stochastic Search Technique based Suboptimal AGC Regulator Design for ...
 
IoT Device Intelligence & Real Time Anomaly Detection
IoT Device Intelligence & Real Time Anomaly DetectionIoT Device Intelligence & Real Time Anomaly Detection
IoT Device Intelligence & Real Time Anomaly Detection
 
A new scaled fuzzy method using PSO segmentation (SePSO) applied for two area...
A new scaled fuzzy method using PSO segmentation (SePSO) applied for two area...A new scaled fuzzy method using PSO segmentation (SePSO) applied for two area...
A new scaled fuzzy method using PSO segmentation (SePSO) applied for two area...
 
Computational intelligence systems in industrial engineering
Computational intelligence systems in industrial engineeringComputational intelligence systems in industrial engineering
Computational intelligence systems in industrial engineering
 
Data-driven adaptive predictive control for an activated sludge process
Data-driven adaptive predictive control for an activated sludge processData-driven adaptive predictive control for an activated sludge process
Data-driven adaptive predictive control for an activated sludge process
 
Nuclear Power Plant Electronic System Reliability Study
Nuclear Power Plant Electronic System Reliability StudyNuclear Power Plant Electronic System Reliability Study
Nuclear Power Plant Electronic System Reliability Study
 
Six sigma pedagogy
Six sigma pedagogySix sigma pedagogy
Six sigma pedagogy
 
Six sigma
Six sigma Six sigma
Six sigma
 
TQM
TQMTQM
TQM
 
report
reportreport
report
 
Six sigma-in-measurement-systems-evaluating-the-hidden-factory (2)
Six sigma-in-measurement-systems-evaluating-the-hidden-factory (2)Six sigma-in-measurement-systems-evaluating-the-hidden-factory (2)
Six sigma-in-measurement-systems-evaluating-the-hidden-factory (2)
 

Recently uploaded

Cloud Frontiers: A Deep Dive into Serverless Spatial Data and FME
Cloud Frontiers:  A Deep Dive into Serverless Spatial Data and FMECloud Frontiers:  A Deep Dive into Serverless Spatial Data and FME
Cloud Frontiers: A Deep Dive into Serverless Spatial Data and FME
Safe Software
 

Recently uploaded (20)

Apidays New York 2024 - APIs in 2030: The Risk of Technological Sleepwalk by ...
Apidays New York 2024 - APIs in 2030: The Risk of Technological Sleepwalk by ...Apidays New York 2024 - APIs in 2030: The Risk of Technological Sleepwalk by ...
Apidays New York 2024 - APIs in 2030: The Risk of Technological Sleepwalk by ...
 
Exploring Multimodal Embeddings with Milvus
Exploring Multimodal Embeddings with MilvusExploring Multimodal Embeddings with Milvus
Exploring Multimodal Embeddings with Milvus
 
ICT role in 21st century education and its challenges
ICT role in 21st century education and its challengesICT role in 21st century education and its challenges
ICT role in 21st century education and its challenges
 
DEV meet-up UiPath Document Understanding May 7 2024 Amsterdam
DEV meet-up UiPath Document Understanding May 7 2024 AmsterdamDEV meet-up UiPath Document Understanding May 7 2024 Amsterdam
DEV meet-up UiPath Document Understanding May 7 2024 Amsterdam
 
Navigating the Deluge_ Dubai Floods and the Resilience of Dubai International...
Navigating the Deluge_ Dubai Floods and the Resilience of Dubai International...Navigating the Deluge_ Dubai Floods and the Resilience of Dubai International...
Navigating the Deluge_ Dubai Floods and the Resilience of Dubai International...
 
Connector Corner: Accelerate revenue generation using UiPath API-centric busi...
Connector Corner: Accelerate revenue generation using UiPath API-centric busi...Connector Corner: Accelerate revenue generation using UiPath API-centric busi...
Connector Corner: Accelerate revenue generation using UiPath API-centric busi...
 
Strategies for Landing an Oracle DBA Job as a Fresher
Strategies for Landing an Oracle DBA Job as a FresherStrategies for Landing an Oracle DBA Job as a Fresher
Strategies for Landing an Oracle DBA Job as a Fresher
 
[BuildWithAI] Introduction to Gemini.pdf
[BuildWithAI] Introduction to Gemini.pdf[BuildWithAI] Introduction to Gemini.pdf
[BuildWithAI] Introduction to Gemini.pdf
 
Biography Of Angeliki Cooney | Senior Vice President Life Sciences | Albany, ...
Biography Of Angeliki Cooney | Senior Vice President Life Sciences | Albany, ...Biography Of Angeliki Cooney | Senior Vice President Life Sciences | Albany, ...
Biography Of Angeliki Cooney | Senior Vice President Life Sciences | Albany, ...
 
Manulife - Insurer Transformation Award 2024
Manulife - Insurer Transformation Award 2024Manulife - Insurer Transformation Award 2024
Manulife - Insurer Transformation Award 2024
 
Spring Boot vs Quarkus the ultimate battle - DevoxxUK
Spring Boot vs Quarkus the ultimate battle - DevoxxUKSpring Boot vs Quarkus the ultimate battle - DevoxxUK
Spring Boot vs Quarkus the ultimate battle - DevoxxUK
 
Artificial Intelligence Chap.5 : Uncertainty
Artificial Intelligence Chap.5 : UncertaintyArtificial Intelligence Chap.5 : Uncertainty
Artificial Intelligence Chap.5 : Uncertainty
 
EMPOWERMENT TECHNOLOGY GRADE 11 QUARTER 2 REVIEWER
EMPOWERMENT TECHNOLOGY GRADE 11 QUARTER 2 REVIEWEREMPOWERMENT TECHNOLOGY GRADE 11 QUARTER 2 REVIEWER
EMPOWERMENT TECHNOLOGY GRADE 11 QUARTER 2 REVIEWER
 
2024: Domino Containers - The Next Step. News from the Domino Container commu...
2024: Domino Containers - The Next Step. News from the Domino Container commu...2024: Domino Containers - The Next Step. News from the Domino Container commu...
2024: Domino Containers - The Next Step. News from the Domino Container commu...
 
AWS Community Day CPH - Three problems of Terraform
AWS Community Day CPH - Three problems of TerraformAWS Community Day CPH - Three problems of Terraform
AWS Community Day CPH - Three problems of Terraform
 
Repurposing LNG terminals for Hydrogen Ammonia: Feasibility and Cost Saving
Repurposing LNG terminals for Hydrogen Ammonia: Feasibility and Cost SavingRepurposing LNG terminals for Hydrogen Ammonia: Feasibility and Cost Saving
Repurposing LNG terminals for Hydrogen Ammonia: Feasibility and Cost Saving
 
AXA XL - Insurer Innovation Award Americas 2024
AXA XL - Insurer Innovation Award Americas 2024AXA XL - Insurer Innovation Award Americas 2024
AXA XL - Insurer Innovation Award Americas 2024
 
Strategize a Smooth Tenant-to-tenant Migration and Copilot Takeoff
Strategize a Smooth Tenant-to-tenant Migration and Copilot TakeoffStrategize a Smooth Tenant-to-tenant Migration and Copilot Takeoff
Strategize a Smooth Tenant-to-tenant Migration and Copilot Takeoff
 
Rising Above_ Dubai Floods and the Fortitude of Dubai International Airport.pdf
Rising Above_ Dubai Floods and the Fortitude of Dubai International Airport.pdfRising Above_ Dubai Floods and the Fortitude of Dubai International Airport.pdf
Rising Above_ Dubai Floods and the Fortitude of Dubai International Airport.pdf
 
Cloud Frontiers: A Deep Dive into Serverless Spatial Data and FME
Cloud Frontiers:  A Deep Dive into Serverless Spatial Data and FMECloud Frontiers:  A Deep Dive into Serverless Spatial Data and FME
Cloud Frontiers: A Deep Dive into Serverless Spatial Data and FME
 

080924 Measurement System Analysis Re Sampling

  • 1. Solutia From: Richard W. Miller (850) 968-8358 Fax: (850) 968-8732 “Solutia Fellow” [rwmill@solutia.com] Date: September 24, 2008 Subject: Analysis of the 41LDC Manufacturing Process and the AEG Measurement System with an Emphasis on Resampling Out-of-Spec Lots Summary Greenwood’s (GWD) AEG measurement system, validated utilizing the 5/3 strategy (five routine samples split into thirds with their variance estimates pooled), accounted for ~12% of 41LDC’s total AEG variation. While this process was judged capable (Ppk=0.99), the long-term component of variation was significantly greater than its short-term component. Finally, the routine re-measure of out-of-spec material is a good strategy when the measurement system variation component is small enough; otherwise, the β risk (customer risk of accepting out-of-spec material) increases dramatically. Jeff Wang, Gordon Mayfield and I are in the process of setting up continuous measurement system validation for important polymer characteristics (AEG, RV, TiO2, H2O) at Greenwood. This report summarizes the AEG characterization of 41LDC made on CP 45 between 6/11 & 8/19/08. In addition to summarizing the process data, a probabilistic argument for resampling out- of-spec material is presented. Major Observations & Recommendations • AEG measurement system accounted for ~12% of GWD’d 41LDC (CP 45) variability; • While GWD’s 41LDC manufacturing process is capable (Ppk=0.99 for in process sampling), its capability could be improved significantly (Ppk=1.34) by reducing long-term variation: campaign-to-campaign centering and variation fluctuation, and general process drift; • Initiate measurement system validation (GWD and Pensacola(PNS)) for other important product properties (e.g., RV, TiO2 and H2O for 41LDC); • Assess risks of out-of-spec sample re-measure through estimation of each measurement system’s sigma; a sample measuring USL+1σ Meas Sys or greater than LSL - 1σ Meas Sys will have a 16% chance of falling within specs on re-measure (even without analytical error), a 32% chance on two re-measures, a 48% chance on three re-measures; • Not all in-spec measurement distributions are equal; the level of variation can be more important than the targeting because customer satisfaction is maximized at the minimum in the Taguchi Loss Function {min(TLF) = min (σLT 2 + (Target – μ) 2 }; • When a customer identifies a spec range as USL – LSL, that means the producer’s effective spec range must be redefined tighter as USLeff=USL-xσ Meas Sys and LSLeff=LSL+xσ Meas Sys where x might vary between 0 and 6 depending on the level of customer risk (calling out-of-spec product good) desired.
  • 2. Greenwood’s 41LDC AEG Process Capability Chart 1 plots single AEG measures made over this two-month production period. While this is in-process data—data used to control the CP line—the pack-out (P/O) characteristics were similar. Red lines mark the population mean and upper and lower control limits (μ +/- 3σST), while the blue lines mark specification limits (45 +/- 1.2). The first represents the “voice of the process”; the second restrictions imposed on the process—ideally by customer needs. One can expect 99.73% of the population to fall between the outside red lines when there’s no long-term component of variation (process drift or campaign-to-campaign centering). When a point falls beyond the red lines, it’s deemed a special cause, i.e., likely to come from a different population because of some special cause. Often, the special cause for an outlier data can be tracked down and changes made to re- center a shifted process, but just as often the reasons for the shift go unresolved and compensatory adjustments are made anyway. The BCF process is a good example of the latter where temperature adjustments are routinely made to re-center positions that continually drift as process components age and foul. There’s also the very real possibility of a questionable measurement, either because of outright error or because of the distributional nature of the measurement system itself. Because this producer risk (risk of off-grading acceptable product) is often judged more important to the producer than customer risk (risk of not off-grading unacceptable product), off grade material is given every opportunity to prove itself before being committed to a rework strategy. The question then becomes: If the same lot is re-measured how likely is the new measure to fall within specs?
  • 3. If a point falls beyond the specification limits (beyond the blue lines), especially for P/O data, it’s common practice to re-measure that lot before reworking it to meet specifications. I’ll address the resample question and its risk more fully in a later section. This process’ capability matrix is shown in Chart 2. Specification limits, population mean and its two measures of variation (short-term sigma and long-term sigma) are used to calculate the capability metrics (Ppk, Pp, Cpk, Cp and % out-of-spec material) as defined in the chart. The first column houses capability estimates for the current process, while the other three columns calculate these metrics for other choices of specification limits. Note, the basic difference between the two product capability metrics (Ppk and Pp) and the two process capability metrics (Cpk and Cp) lies in the nature of their respective variation components: the first uses long-term sigma and characterizes the product the customer receives, while the second uses short-term sigma and characterizes the ultimate capability of the process. These metrics confirm Chart 1’s picture: GWD’s 41LDC process is pretty capable (Ppk nearly 1.0) with blue specification lines well outside the red control ones. The difference between Ppk and Pp (and Cpk and Cp) indicates a slight centering problem, while the difference between Ppk and Cpk indicates a significant long-term variation component. If this process could be re- centered and its long-term component of variation reduced, it’s capable of producing a 1.49 Ppk (making virtually nothing out-of-spec), i.e., spec range / 6σ = 1.49 means the spec range would contain ~9 σLTs. Note, a Ppk=1 means the spec range would contain 6 σLTs.
  • 4. 5/3 Strategy: AEG Measurement System Validation The following table summaries AEG data collected from the 5/3 testing where five routine samples were split into thirds then characterized blindly on three different shifts. The idea is to get routine measurement system data which reflect true lab variation. Assuming the subsample compositions are identical (a good blend), their pooled variance estimate provides our estimate of the measurement system’s variance. Here, σ 2 Meas Sys is estimated to be 0.016 for the AEG measure, which in turn, accounts for ~12% of the total variation (σ 2 LT) evident in the Chart 1 data. Note, the 5/3 strategy follows this initial 2-week sampling with another single sample split every two weeks. Continuous validation of the AEG measurement system is thus provided at the cost of a single extra sample each week.
  • 5. Chart 1 from a Probabilistic View and the Resampling Question The data presented in Chart 1 can also be looked at through the prism of probabilistic models, which then enable us to ask questions: How does the difference between short-term and long-term variance impact the frequency of control limit (μ+/-3σST) and specification limit violations? How does a process shift impact quality? What is the impact of resampling a lot after it has experienced an out-of-spec signal the first time? And finally, how does measurement system variation influence the risk of resampling? Note, this modeling assumes the data are normally distributed—a good assumption for these data. The black distribution shown in Chart 3 depicts the probability distribution represented in the AEG data plotted sequentially in Chart 1 (inset in Chart 3 as well). It describes the frequency of AEG measures as a function of AEG level. The red distribution is that black distribution hypothetically shifted to the high side by two long-term standard deviations, and the blue distribution is the measurement system’s distribution expected for a lot which has measured an out-of-spec 46.5 AEG in its routine sampling. The green vertical lines mark the process’ target and specification limits, while the blue vertical lines mark control limits for the black distribution. Note, the black distribution, called process State A is slightly off center to the high side of target (hence the Ppk-Pp capability differences).
  • 6. Chart 4 shows the percentage of each of the above distribution’s population that lies either below or above the specification, control and target limits. A further distribution, the measurement system centered at the upper specification limit, is included as well. Several points become clearer. First, short-term sigma are used in Shewhart charts to define control limits (μ +/- 3σST) where the sigma are calculated for single-point measures from a 2-point moving range and for multi-point averages from their within sample group variances. The State A distribution, on the other hand, reflects the long-term sigma variation. It’s the variation one gets by throwing all 174 data in Chart 1 into a hat and calculating its variance. The net effect, since 3σST <= 3σLT, is that the routine data can be expected to violate its μ+3σST limits more frequently than expected. For example here, State A violates the lower control limit (LCL) at a 1.40% frequency (1 in 71 data) and the upper control limit (UCL) at a 1.38% frequency (1 in 72 data); whereas a typical Shewhart chart is cited to violate either control limit at a 0.135% frequency (1 in 740 data). Obviously, the producer’s risk (calling good product bad) increases appreciably when σLT >>> σST. The second question asks how a process shift (the red distribution, a +2 σLT, in Chart 3) impacts quality. First, is there really any difference between two lots measuring a 46.75 AEG where the first one comes from un-shifted (black) distribution and the second from the shifted (red) distribution? No, as far as we can measure, the two are identical. What’s different is the shifted distribution can be expected to produce 17.09% of its lots above the upper spec limit,
  • 7. while the original distribution can be expected to place only 0.16% there. That’s why one should be concerned about a shifted process. Note too, the shifted process is far more likely to produce a runs-rule signal, e.g., the next 8 points in the black distribution would be expected to fall on one side of its mean 0.39% of the time {(0.5)^8} or 1 time in 256 groups; whereas, the shifted red distribution would be expected to signal that 8-point grouping to the high side of the original (black distribution) mean 83.2% of the time. The final two questions focus on the measurement system. Now that we can estimate the measurement system’s variance, we can better assess both the risk of a re-measure and our true spec limits. Remember, we re-measure a given lot of the red or black distribution in Chart 3; whereas both of these distributions reflect the breadth of their respective lot measurements. The blue distribution in Chart 3 outlines the frequency of AEG measures that one could expect from just taking the original sample (AEG=46.5) and re-measuring it a number of times. Note, 46.5 is 2.4σMeas Sys above the 45.93 USL. We’re assuming there’s no difference in the sample, just this level of variation in the measurement system itself (similar to the assumption made in the 5/3 sampling strategy). In the particular measurement distribution shown, virtually 99.18% of the distribution lies above the 46.2 USL so any re-measure is likely to remain out-of-spec. Of course, that assumes that no error occurred in the original measure. That means if the re-measure falls within spec limits, there’s likely been an error made in the original measure. Let’s look at another sample. Say we measure AEG for a lot at 46.2, exactly on the USL and decide to resample. Given the variation in the measurement system, a single re-measure will produce an out-of-spec signal 50% of the time and an in-spec signal 50% of the time. Moreover, if I were to resample two times, there would be a 75% chance of getting an in-spec signal, three times a 87.5% chance. In this case, producer’s risk (calling good product bad) has been judged far more important than customer’s risk (calling bad product good). So now let’s reformulate spec limits to reflect the measurement system’s variation. Assume that customer’s risk is most important to us. To ensure that no product leaves the building outside spec limits, one needs to use effective spec limits: USLeff = USL - 3σMeas Sys, here 45.83 (46.2 – 3x0.125) and LSLeff = LSL + 6σMeas Sys or 44.18 (43.8+3x0.125). Obviously, managing variation is critical to quality.
  • 8. Variation and Customer Preferences Traditionally, quality has been viewed as a step function where product is judged either good or bad depending on whether it falls within specification limits or not. Taguchi put forth the idea that the customer becomes increasingly dissatisfied as a product drifts from target, even if that product stays within spec limits. The customer desires a product that is constant part-to-part, lot-to-lot or bobbin-to-bobbin. The producer, on the other hand, desires to make and sell as much A-grade material as possible. Each lot contributes to the (Target – μlot) 2 , while their aggregate variation determines the σLT 2 term. Customer Satisfaction: Taguchi Loss Function {min(TLF) = min (σLT 2 + (Target – μlot) 2 } Clearly, in some cases—like the two distributions in the chart below—variance (even with in-spec material) is more important to customer satisfaction than targeting. The following chart further demonstrates the point. In these two distributions is there really any difference among lots A, B and C? Don’t all differ significantly from lot D? Which of the two distributions do you think will produce more consistent product for the customer? The answers are obvious: the right-side distribution, though it has a targeting problem, would both be easier to use by the customer and easier to control by the producer. The left-side distribution has a variation problem. Even though everything falls inside spec limits, it will be harder to use by the customer. This is essentially the problem that the low-melt yarn program has run into over the past few years. EMS produces a low melt resin that met melt viscosity specs, though their lot-to-lot variation looked much like the left-side distribution below. The problem came when we (the customer) tried to spin their resin: those lots from the high-side of the distribution required far
  • 9. different (sometimes unknown) process settings than those lots from the low-side of the distribution required. As a consequence, much of the yarn ends up on the floor as the process re- equilibrates, and yields suffered. Again, the message is clear: variation is critical to quality.