Download Presentation-Statistical Process Control

Document related concepts

History of statistics wikipedia , lookup

Data mining wikipedia , lookup

Misuse of statistics wikipedia , lookup

Time series wikipedia , lookup

Transcript
Statistical Process Control
0011 0010 1010 1101 0001 0100 1011
1
2
4
Statistical Process Control
0011 0010 1010 1101 0001 0100 1011
1
2
4
Why do we Need SPC?
To Help Ensure Quality
0011 0010 1010 1101 0001 0100 1011
Quality means fitness for use
- quality of design
- quality of conformance
Quality is inversely proportional to variability.
1
2
4
What is Quality?
0011 0010 1010 1101 0001 0100 1011
• Transcendent definition: excellence
– Excellence in every aspect
• PERFORMANCE
1
2
How well the output does what it is supposed to do.
• RELIABILITY
4
The ability of the output (and its provider) to
function as promised
What is Quality
0011 0010 1010 1101 0001 0100 1011
• CONVENIENCE and ACCESSIBILITY
How easy it is for a customer to use the product or
service.
• FEATURES
1
2
The characteristics of the output that exceed the output’s
basic functions.
• EMPATHY
4
The demonstration of caring and individual attention to
customers.
What is Quality
0011 0010 1010 1101 0001 0100 1011
• CONFORMANCE
The degree to which an output meets specifications or
requirements.
• SERVICEABILITY
1
2
How easy it is for you or the customer to fix the output
with minimum downtime or cost.
• DURABILITY
How long the output lasts.
4
What is Quality
0011 0010 1010 1101 0001 0100 1011
• AESTHETICS
How a product looks, feels, tastes, etc.
• CONSISTENCY
1
2
The degree to which the performance changes over
time.
• ASSURANCE
4
The knowledge and courtesy of the employees and
their ability to elicit trust and confidence.
What is Quality
0011 0010 1010 1101 0001 0100 1011
• RESPONSIVENESS
Willingness and ability of employees to help
customers and provide proper services.
• PERCEIVED QUALITY
1
2
4
The relative quality level of the output in the eyes of
the customers.
What is Quality
0011 0010 1010 1101 0001 0100 1011
• Product-based definition: quantities of
product attributes
1
2
– Attributes are non-measurable types of data
• What are all the different features
4
What is Quality
0011 0010 1010 1101 0001 0100 1011
• User-based definition: fitness for intended
use
1
2
– How well does the product meet or exceed the
expected use as seen by the user
4
What is Quality
0011 0010 1010 1101 0001 0100 1011
• Value-based definition: quality vs. price
– How much is a product or service going to
cost and then how much attention to quality
can we afford to spend.
– Cheap product, little quality
1
2
4
What is Quality
0011 0010 1010 1101 0001 0100 1011
• Manufacturing-based definition:
conformance to specifications
1
2
– The product has both variable specifications
(measurable) and attributable specifications
(non-measurable) that manufacturing monitors
and ensures conformance
4
0011 0010 1010 1101 0001 0100 1011
1
2
4
What is Statistical Process
Control?
What is SPC
0011 0010 1010 1101 0001 0100 1011
• SPC, Statistical Process Control, is a
process that was designed in the 1930’s to
characterize changes in process variation
from a standard.
1
2
– It can be used for both attributes and variables
4
0011 0010 1010 1101 0001 0100 1011
• The basic tool used in SPC is the control
chart
1
– There are various types of control charts
•
•
•
•
•
•
Mean chart
Range chart
Median chart
Mean and range chart (X and R)
c chart
p chart, etc.
2
4
0011 0010 1010 1101 0001 0100 1011
• Control charts
– a graphical method for detecting if the
underlying distribution of variation of some
measurable characteristic of the product seems
to have undergone a shift
– monitor a process in real time
– map the output of a production process over
time and signals when a change in the
probability distribution generating observations
seems to have occurred
– are based on the Central Limit Theory
1
2
4
0011 0010 1010 1101 0001 0100 1011
• Central Limit Theorem says that the
distribution of sums of Independent and
Identically Distributed (IID) random
variables approaches the normal
distribution as the number of terms in the
sum increases.
1
2
4
– Things tend to gather around a center point
– As they gather they form a bell shaped curve
0011 0010 1010 1101 0001 0100 1011
• The center of things are described in various ways
– Geographical center
– Center of gravity
1
2
• In statistics, when we look at groups of numbers,
they are centered in three different ways
– Mode
– Median
– Mean
4
Mode
0011 0010 1010 1101 0001 0100 1011
• Mode
– Mode is the number that occurs the most
frequently in a group of numbers
• 7, 9, 11, 6, 13, 6, 6, 3,11
– Put them in order
– 3, 6, 6, 6, 7, 9, 11, 11, 13
– 3, 6, 6, 6, 7, 9, 11, 11, 13
• The mode is 6
1
2
4
Median
0011 0010 1010 1101 0001 0100 1011
~ is like the geographical center,
• Median (X)
it would be the middle number
• 7, 9, 11, 6, 13, 6, 6, 3,11
– Put them in order
– 3, 6, 6, 6, 7, 9, 11, 11, 13
– 3, 6, 6, 6, 7, 9, 11, 11, 13
• 7 is the median
1
2
4
Mean
0011 0010 1010 1101 0001 0100 1011
• Mean is the average of all the numbers and
_
is designated
by the symbol μ for
X
population mean and for sample mean
1
2
– The mean is derived by adding all the numbers
and then dividing by the quantity of numbers
– X1 + X2 + X3 + X4 + X5 + X6 + X7 +…+Xn
n
4
… to the nth number …
0011 0010 1010 1101 0001 0100 1011
… multiplied by 1 over n
The mean …
The sum of …
n
_
X
=
… is equal to …
… from the first number …
1
n
Σ
i=1
Xi
1
2
… all the numbers …
4
If we had the numbers, 1,2,3,6 and 8, you can
see below that they “balance” the scale. The
0011 0010 1010 1101 0001 0100 1011
mean is not geometric center but like the center
of gravity
1
1
2
3
6
2
4
8
0011 0010 1010 1101 0001 0100 1011
• As the numbers are accumulated, they are
put in order, smallest to largest, and the
number or each number can then be put into
a graph called a Histogram
1
2
4
45
0011 0010 1010 1101 0001 0100 1011
40
35
30
1
25
20
15
10
5
40
45
50
55
2
4
60
45
0011 0010 1010 1101 0001 0100 1011
40
35
30
1
25
20
15
10
5
40
45
50
55
2
4
60
0011 0010 1010 1101 0001 0100 1011
1
2
4
0011 0010 1010 1101 0001 0100 1011
• A normal curve is considered “normal” if
the following things occur
1
– The shape is symmetrical about the mean
– The mean, the mode and the median are the
same
2
4
0011 0010 1010 1101 0001 0100 1011
_
X
1
Variation
2
4
_
X
0011 0010 1010 1101 0001 0100 1011
1
2
4
Workshop I
0011 0010 1010 1101 0001 0100 1011
Central Tendency
1
2
4
Variation
0011 0010 1010 1101 0001 0100 1011
• The numbers that were not exactly on the
mean are considered “variation”
– When weighing the candy, the manufacturer is
targeting a specific weight
– Those that do not hit the specific weight are
variations.
2
4
• Will there always be variation?
• There are two types of variation
– Common cause variation
– Special cause variation
1
Common Cause
0011 0010 1010 1101 0001 0100 1011
• Common cause variation is that normal
variation that exists in a process when it is
running exactly as it should
– eg. In the production of that candy
1
2
4
• When the operator is running the machine properly
–
–
–
–
Within cycle time allotted for each drop of candy
Candy is properly placed on trays
Temperatures are where they need to be
Mixture is correct
0011 0010 1010 1101 0001 0100 1011
• When the machine is running properly
–
–
–
–
Tooling is sharp and aligned correctly
All components are properly maintained
Voltage is correct
Safety interlocks are properly set
• When the material is correct
–
–
–
–
Hardness
Size
Thickness
Blend
1
2
4
0011 0010 1010 1101 0001 0100 1011
• When the method is correct
– Right tonnage machine
– Proper timing
• When the environment is correct
–
–
–
–
Ambient temperature
Ambient humidity
Dust and dirt
Corrosives
1
4
• When the original measurements are correct
– Die opening dimensions
2
0011 0010 1010 1101 0001 0100 1011
• As we have just reviewed, common cause
variation cannot be defined by one
particular characteristic
1
2
– It is the inherent variation of all the parts of the
operation together
• Voltage fluctuation
• Looseness or tightness of bearings
4
0011 0010 1010 1101 0001 0100 1011
1
2
4
0011 0010 1010 1101 0001 0100 1011
• Common cause variation must be optimized
and run at a reasonable cost
– Don’t spend a dollar to save a penny
1
2
4
Special Cause
0011 0010 1010 1101 0001 0100 1011
• Special cause is when one or more of the
process specifications/conditions change
–
–
–
–
–
–
Temperatures
Tools dull
Voltage drops drastically
Material change
Stops move
Bearings are failing
1
2
4
0011 0010 1010 1101 0001 0100 1011
• Special cause variations are the variations
that need to be corrected
1
– But how do we know when these problems
begin to happen?
2
4
Statistical Process Collection and Control!
Workshop II
0011 0010 1010 1101 0001 0100 1011
1
2
4
Range and Mean Variation
0011 0010 1010 1101 0001 0100 1011
• Variation in the process occurs two major
ways.
– The range changes
– The mean changes
1
2
4
0011 0010 1010 1101 0001 0100 1011
• Range is the smallest data point subtracted
from the largest data point
1
2
– It represents the total data spread that has been
sampled
– If the range gets smaller, or more significantly
if it gets larger, something has changed in the
process.
4
_
X
0011 0010 1010 1101 0001 0100 1011
1
Changed Process
2
4
Normal Process
Changed Range
Normal Range
0011 0010 1010 1101 0001 0100 1011
_
X
Changed Process
1
2
4
Normal Process
Changed Range
Normal Range
0011 0010 1010 1101 0001 0100 1011
• Give me some examples that you would
think would cause the range to tighten up
• How about loosen up?
1
2
4
0011 0010 1010 1101 0001 0100 1011
• If you remember, the mean is the point
around which the data is centered
1
2
– If the mean changes, then it would mean that
the central point has changed
4
0011 0010 1010 1101 0001 0100 1011
• Now lets see what the curve would look like
if the mean changed
1
2
4
0011 0010 1010 1101 0001 0100 1011
_
X
1
2
4
0011 0010 1010 1101 0001 0100 1011
_
X’
_
X
1
2
4
0011 0010 1010 1101 0001 0100 1011
• What might cause the mean to shift?
1
2
4
0011 0010 1010 1101 0001 0100 1011
• Most of the time we see both happen to
some degree. In the previous example, you
may have noticed that the range also
became smaller
1
2
4
0011 0010 1010 1101 0001 0100 1011
_
X’
_
X
1
2
4
0011 0010 1010 1101 0001 0100 1011
• It is important that when parts are being
sampled from an entire population of
production, that they are randomly selected
1
– This is called statistical collection of data
2
4
0011 0010 1010 1101 0001 0100 1011
• Why do you think the sampling should be
done randomly?
1
2
4
Workshop III
0011 0010 1010 1101 0001 0100 1011
1
2
4
Black beads and White beads
Statistical Process Control
0011 0010 1010 1101 0001 0100 1011
1
2
4
Gathering Meaningful Data
Why do we need Data?
0011 0010 1010 1101 0001 0100 1011
• Assessment
– Assessing the effectiveness of specific
techniques or corrective actions
• Evaluation
1
2
4
– Determine the quality of a process or product
• Improvement
– Help us understand where improvement is
needed
0011 0010 1010 1101 0001 0100 1011
• Control
– To help control a process and to ensure it does
not move out of control
• Prediction
1
2
– Provide information and trends that enables us
to predict when an activity will fail in the future
• Characterization
4
– Help us understand weaknesses and strengths
of products and processes
Deciding on the Data
0011 0010 1010 1101 0001 0100 1011
• Before data is collected, a plan needs to be
put in place
1
– If there is no clear objective, then the data is
very likely not going to be useful to anyone
– Do not collect data for just because you are
supposed to
2
4
Deciding on the Data
0011 0010 1010 1101 0001 0100 1011
• Base the data being collected on the
hypothesis of the process to be examined
1
– Know what the expected results should be
– Know what parts of the process can give
meaningful data
– Understand the process as a whole
2
4
Deciding on the Data
0011 0010 1010 1101 0001 0100 1011
• Consider the impact of the data collection
process on the whole organization
2
– Data collection can be very expensive and time
consuming
– Data collection imparts a “Big Brother is
Watching” feeling on employees
– Data collection causes employees to work
harder and more focused than they really would
– Data gathering is tedious and requires a
commitment
1
4
Deciding on the Data
0011 0010 1010 1101 0001 0100 1011
• Data gathering must have complete, total,
100% support from management
1
2
– Data gathering inherently takes time, analysis
even more time, and positive results even more
time than that.
– Management can be very impatient some times
4
Guidelines for Useful Data
0011 0010 1010 1101 0001 0100 1011
• Data must contain information that allows for
identification of types of errors or changes made
• Data MUST include cost to making changes
• Data must be compared in some way to
expectation or specification
• Benchmarks must be available from historical data
if possible
• Data must be clear enough to allow for
repeatability in the future, or on other projects
1
2
4
0011 0010 1010 1101 0001 0100 1011
• Data should cover the following
–
–
–
–
What, by whom and for what purpose
What are the specifications
Who will gather the data
Who will support the data gathering
• And have they agreed
– How will the data be gathered
– How will the data be validated
– How will the data be managed
1
2
4
Tips on data collection
0011 0010 1010 1101 0001 0100 1011
• Establish goals and the data gathering and
define and be prepared for questions that
may occur during gathering
• Involve all the people who are going to be
affected by the gathering, analysis and
corrective/preventive actions
• Keep goals small to start with, because the
data could be almost overwhelming
1
2
4
0011 0010 1010 1101 0001 0100 1011
• Design the data collection form to be
simple, so that comparison is easy
1
2
– Provide training in the process to be followed
• Include any validation criteria (check
against known calibrated gauge, etc.) and
record results.
• Automate as much as possible
4
0011 0010 1010 1101 0001 0100 1011
• The data will be collected onto a collection
form that will then be transferred to a
control chart
1
– In many cases the two forms are the same
2
4
• These control charts will help us to decide if
the process is running as we want, and
hopefully tell us that it is beginning to
change so we can avoid producing bad parts
0011 0010 1010 1101 0001 0100 1011
• These charts can be grouped into two major
categories
1
– They are grouped based on the type of data
being collected
• Attributes data
• Variables data
2
4
Attributes
0011 0010 1010 1101 0001 0100 1011
• Attributes data
– Non-measurable characteristics
• Can be very subjective
1
2
– We must develop specific descriptions for each
attribute
• Called operational definitions
–
–
–
–
–
Blush
Splay
Scratched
Color
Presence, etc.
4
Attributes
0011 0010 1010 1101 0001 0100 1011
1
2
4
Attributes
0011 0010 1010 1101 0001 0100 1011
– Observations are counted
• Yes/No
• Present/Absent
• Meets/Doesn’t meet
– Visually inspected
• Go/no-go gauges
• Pre-control gauges
– Discreet scale (has limits)
1
2
4
Attributes
0011 0010 1010 1101 0001 0100 1011
• If color happens to be an attribute that is being
inspected for
– Typically, the expected color sample is given
– Maybe a light and dark sample is given
• The acceptable range is in between.
1
2
4
– A reject is not measured, just counted as one
• Scratches might be given by location, length of
scratch, depth of scratch, width of scratch, visible
at a certain distance, etc.
Attributes
0011 0010 1010 1101 0001 0100 1011
• The description becomes more detailed as
the quality becomes more critical.
1
2
– It is very important that inspection is not used
to sort out chronic problems
– Inspection is used to collect data
4
• Inspection should be as close to the process as
possible and by the operators
Attributes
0011 0010 1010 1101 0001 0100 1011
• On an attributes chart, the characteristic
name is used
1
2
– Scratched, wrong color, specks, bubbles, etc.
– Every time an unacceptable characteristic is
found, it is marked on the data collection sheet
using a method such as counting marks ( IIII )
4
Variables
0011 0010 1010 1101 0001 0100 1011
• Developed through measuring
– Is very objective
– Can be temperature, length, width, weight,
force, volts, amps, etc.
• Uses a measuring tool
– Scale
– Meter
1
2
4
• Scale increments are specified by design
engineers, customer, etc.
Variables
0011 0010 1010 1101 0001 0100 1011
1
2
3
4
5
1
6
2
4
Variables
0011 0010 1010 1101 0001 0100 1011
– As with attributes, variables inspection should
be done as close to process as possible,
preferably by the operator.
1
2
• Inspection should not be done to sort but for data
collection and correction of the process
• This will allow for quick response and rapid
correction, minimizing defect quantities
4
Variables
0011 0010 1010 1101 0001 0100 1011
• Important considerations for data collection
– The measuring tools must be accurate and
precise
1
• Accurate means that it is can produce similar
measurements as a standard
• Precise means repeatability
2
4
Variables
0011 0010 1010 1101 0001 0100 1011
– Sampling program used
• Sampling involves removing a representative
quantity of components from the normal production
and inspecting them
• The sample size and frequency is very important in
remaining confident that any negative process
changes are being uncovered.
• Sampling is also dependent on the type of charting
you will be doing, and will be discussed with each
chart type.
1
2
4
Variables
0011 0010 1010 1101 0001 0100 1011
• After the variables are being collected, they
can then be visualized as a Normal Curve
1
2
– The mean can be calculated and another
characteristic of the normal curve, the standard
deviation, can also be calculated
– There are many low priced and free programs
to allow input and automatic calculation
4
Standard Deviation of Variables
0011 0010 1010 1101 0001 0100 1011
• Standard deviation is characteristic of all normal
curves
– One standard deviation on each side of the mean would
represent 68.26% of the area beneath the curve (or
68.26% of all data)
– Two standard deviations on each side of the mean
would represent 95.44% of the area beneath the curve
– Three standard deviations on each side of the mean
would represent 99.74% of the area.
– Six standard deviations on each side of the mean would
represent 99.9999966% of the area (3.5 defects per
million).
1
2
4
0011 0010 1010 1101 0001 0100 1011
_
X
1
68.26%
.240
.242
-3σ
-2σ
.244
.246
-σ 95.44%
99.74%
.248
.250
2
4
σ
2σ
3σ
.252
.254
.256
.258
.260
0011 0010 1010 1101 0001 0100 1011
1
σ 2σ 3σ
-3σ -2σ -σ
.240
.242
.244
.246
-3
-2
.248
-1
0
.250
1
2
.252
3
2
4
.254
.256
.258
.260
Standard Deviation
0011 0010 1010 1101 0001 0100 1011
• Whenever we talk about standard deviation,
there are two types
– Population standard deviation
– Sample standard deviation
1
2
4
Population Standard Deviation
0011 0010 1010 1101 0001 0100 1011
• The population is considered the greater lot
that the sample is taken from
– A shipment
– A days production
– A shifts production
1
2
4
• The symbol for the population standard
deviation is the Greek letter sigma (σ)
Sample Standard Deviation
0011 0010 1010 1101 0001 0100 1011
• The sample standard deviation is taken from
the specific sample data, which is
considerably smaller than the sample that
would be taken if sampling the entire
population
• The symbol for the sample standard
deviation is the lower case “s”
1
2
4
0011 0010 1010 1101 0001 0100 1011
• The formula for the standard deviation is
σ =
√
Σ (X – μ )2
p-1
s =
√
Σ (X – X )2
n-1
1
2
4
0011 0010 1010 1101 0001 0100 1011
s =
√
Σ (X – X )2
n-1
1
2
4
Workshop IV
0011 0010 1010 1101 0001 0100 1011
Sampling Plans
1
2
4
Statistical Process Control
0011 0010 1010 1101 0001 0100 1011
Pareto Diagrams
1
2
4
0011 0010 1010 1101 0001 0100 1011
• Vilfredo Pareto
– Italy’s wealth
• 80% held by 20% of people
• Used when analyzing attributes
1
2
4
– Based on results of tally numbers in specific
categories
0011 0010 1010 1101 0001 0100 1011
• What is a Pareto Chart used for?
– To display the relative importance of data
– To direct efforts to the biggest improvement
opportunity by highlighting the vital few in
contrast to the useful many
1
2
4
Constructing a Pareto Chart
0011 0010 1010 1101 0001 0100 1011
• Determine the categories and the units for
comparison of the data, such as frequency, cost, or
time.
• Total the raw data in each category, then determine
the grand total by adding the totals of each
category.
• Re-order the categories from largest to smallest.
• Determine the cumulative percent of each
category (i.e., the sum of each category plus all
categories that precede it in the rank order, divided
by the grand total and multiplied by 100).
1
2
4
Constructing a Pareto Chart
0011 0010 1010 1101 0001 0100 1011
• Draw and label the left-hand vertical axis with the
unit of comparison, such as frequency, cost or
time.
• Draw and label the horizontal axis with the
categories. List from left to right in rank order.
• Draw and label the right-hand vertical axis from 0
to 100 percent. The 100 percent should line up
with the grand total on the left-hand vertical axis.
• Beginning with the largest category, draw in bars
for each category representing the total for that
category.
1
2
4
Constructing a Pareto Chart
0011 0010 1010 1101 0001 0100 1011
• Draw a line graph beginning at the righthand corner of the first bar to represent the
cumulative percent for each category as
measured on the right-hand axis.
• Analyze the chart. Usually the top 20% of
the categories will comprise roughly 80% of
the cumulative total.
1
2
4
0011 0010 1010 1101 0001 0100 1011
• Lets assume we are listing all the rejected products
that are removed from a candy manufacturing line
in one week
1
– First we put the rejects in specific categories
•
•
•
•
•
•
•
No wrapper
No center
Wrong shape
Short shot
Wrapper open
Underweight
Overweight
2
4
0011 0010 1010 1101 0001 0100 1011
– Then we tally how many of each category we
have
•
•
•
•
•
•
•
No wrapper - 10
No center - 37
Wrong shape - 53
Short shot - 6
Wrapper open - 132
Underweight - 4
Overweight – 17
– Get the total rejects - 259
1
2
4
0011 0010 1010 1101 0001 0100 1011
– Develop a percentage for each category
•
•
•
•
•
•
•
No wrapper – 10/259 = 3.9%
No center – 37/259 = 14.3%
Wrong shape – 53/259 = 20.5%
Short shot – 6/259 = 2.3%
Wrapper open – 132/259 = 51%
Underweight – 4/259 = 1.5%
Overweight – 17/259 = 6.6%
1
2
4
• Now place the counts in a histogram, largest to
smallest
100
0011 0010
901010 1101 0001 0100 1011
80
70
60
1
51%
50
40
30
20.5%
14.3%
20
6.6%
10
0
Wrapper
Open
No Center
3.9%
No Wrapper
2
4
2.3%
1.5%
Underweight
• Finally, add up each and plot as a line diagram
100
92.4%
96.3%
98.6%
100.1%
90 1010 1101 0001 0100
85.8%
0011 0010
1011
80
71.5%
70
60
1
51%
50
40
30
20.5%
14.3%
20
6.6%
10
0
Wrapper
Open
No Center
3.9%
No Wrapper
2
4
2.3%
1.5%
Underweight
100
92.4%
96.3%
98.6%
100.1%
90
85.8%
0011 0010 1010 1101 0001 0100
1011
80
71.5%
70
60
1
51%
50
40
30
20.5%
14.3%
20
6.6%
10
0
Wrapper
Open
No Center
3.9%
No Wrapper
2
4
2.3%
1.5%
Underweight
Workshop V
0011 0010 1010 1101 0001 0100 1011
Pareto Charting
1
2
4
• After the problem is corrected, continue the data collection
65
0011 0010 1010 1101 0001 0100 1011
60
55
50
41.7%
1
45
40
35
29.1%
30
25
20
13.4%
15
7.8%
10
0
4
4.7%
5
Wrong Shape
No Center
Overweight
No Wrapper
2
Short Shot
3.1%
Underweight
0011 0010 1010 1101 0001 0100 1011
• Sometimes other information would be
better
1
– Use the scrap cost of each rejected part
– Use rework cost of each rejected part
2
4
• This can be especially useful if the rejects
are all at about the same quantity
Here are some tips
0011 0010 1010 1101 0001 0100 1011
• Create before and after comparisons of Pareto
charts to show impact of improvement efforts.
• Construct Pareto charts using different
measurement scales, frequency, cost or time.
• Pareto charts are useful displays of data for
presentations.
• Use objective data to perform Pareto analysis
rather than team members opinions.
1
2
4
0011 0010 1010 1101 0001 0100 1011
• If there is no clear distinction between the
categories -- if all bars are roughly the same height
or half of the categories are required to account for
60 percent of the effect -- consider organizing the
data in a different manner and repeating Pareto
analysis.
• Pareto analysis is most effective when the problem
at hand is defined in terms of shrinking the
product variance to a customer target. For
example, reducing defects or elimination the nonvalue added time in a process.
1
2
4
0011 0010 1010 1101 0001 0100 1011
• The accumulative curve can also be
removed, especially if there is not distinct
shift in the slope of the curve.
1
2
4
Statistical Process Control
0011 0010 1010 1101 0001 0100 1011
Control Charts
1
2
4
X and R
0011 0010 1010 1101 0001 0100 1011
• When are they used
– When you need to assess variability
– When the data can be collected on an ongoing
basis
– When it can be collected over time
– When we are using variables
– Subgroups must be more than 1
1
2
4
0011 0010 1010 1101 0001 0100 1011
• How is it made
– First, complete the header information
1
• This is important so that each collection can be
properly understood and separated from others.
– Record the data
2
4
• Not just data but significant observations.
0011 0010 1010 1101 0001 0100 1011
– Calculate the mean of each subgroup of
information
– Calculate the range for each subgroup
1
2
4
0011 0010 1010 1101 0001 0100 1011
• When the first chart is finished, calculate
the grand average, X
– This is the average of the averages
• Calculate the average of the ranges
1
2
4
0011 0010 1010 1101 0001 0100 1011
• Calculate the control limits
– The formula is
– UCLx = X + ( A2 x R )
– LCLx = X – (A2 x R )
1
2
4
0011 0010 1010 1101 0001 0100 1011
• Calculate the Range limits
– UCLR = D4 x R
– LCLR = D3 x R
1
2
4
0011 0010 1010 1101 0001 0100 1011
Weighting Factors
Subgroup Size
A2
D3
D4
2
1.880
0
3.267
3
1.023
0
2.574
4
0.729
0
2.282
5
0.577
0
2.114
1
2
4
0011 0010 1010 1101 0001 0100 1011
• Scale the charts
– Use a scale that will ensure the numbers will all
fit in the chart and also that the new numbers
will also fit, even outside the control limits
1
2
4
0011 0010 1010 1101 0001 0100 1011
• Find the largest X value and compare to the UCL.
Use the larger
• Find the smallest X value and compare to the
LCL. Use the smaller
• Subtract the smaller from the larger and write
down the difference
• Divide the difference by 2/3 the number of lines
on the chart (30 lines on this chart) and round
upward if needed.
1
2
4
Control Chart Interpretation
0011 0010 1010 1101 0001 0100 1011
• Interpret the Data
–
–
–
–
Any point lying outside the control limits
Seven points in a row above or below the average
Seven points in a row going in one direction
Any non-random patterns
• Should look like the normal curve
– Too close to average
– Too far from average
– Cycling
1
2
4
0011 0010 1010 1101 0001 0100 1011
• Declare whether in control or out of control
• Respond to the information
1
2
4
PreControl
0011 0010 1010 1101 0001 0100 1011
• PreControl is a form of X bar and R chart…
without the chart.
1
2
– It is to be used only when a process has been
proven to be in control
– It is to be used to stop production before bad
parts are produced
4
0011 0010 1010 1101 0001 0100 1011
• A definitive variable must be selected that
will tell if the process is moving out of
control
– ie. weight, length, warp, etc.
1
2
4
• A PreControl gauge must be made to
measure the variable
– Dial indicator, weigh scale, taper gauge, etc.
0011 0010 1010 1101 0001 0100 1011
• The total specified range that is acceptable
is to be calculated for standard deviation,
and plus or minus two standard deviations
are the control limits for the PreControl
gauge
1
2
4
– The third standard deviation is red
– The second standard deviation is yellow
– The first standard deviation is green
0011 0010 1010 1101 0001 0100 1011
• Rules of PreControl
– At start up, after the process has been set to
know parameters that produces good parts
1
2
• Collect 5 samples in a row from each cavity, mold,
core, station, whatever.
• Each are checked to the definitive PreControl gauge
• All five of each must be in the green
4
– If not, the process must be brought under control
0011 0010 1010 1101 0001 0100 1011
• Once all five are in the green the process is
started up
• At selected intervals (typically one an hour
or once every half hour) two samples in a
row are selected from each cavity, etc.
1
2
4
0011 0010 1010 1101 0001 0100 1011
– If the both are green, process is ok
– If the first is green, but the second is yellow, ok
but be alert
– If both are yellow, stop the process and go back
to start up
– If either is red, stop the process and go back to
start up
1
2
4
0011 0010 1010 1101 0001 0100 1011
• We can determine how far data that is being
gathered is out of specification
1
– Let us look at some make believe data
2
4
Using Control Charting
0011 0010 1010 1101 0001 0100 1011
1
2
4
0011 0010 1010 1101 0001 0100 1011
• We are investigating the time it will take for
a process to pick up a part, move it to
another area and place it on a conveyor
1
– It is unacceptable if it takes more than 14
seconds
2
4
0011 0010 1010 1101 0001 0100 1011
• Data collection has given us the following
information
1
2
– The grand mean is calculated to be 10.00
seconds
– Our sample size has been 5 observations each
hour
– Our calculated average range is 4.653
4
Estimated Standard Deviation
0011 0010 1010 1101 0001 0100 1011
• The formula to find the estimated standard
deviation is as follows
_
^ = R/d
σ
2
1
2
4
• This standard deviation is calculated to one
more decimal point than the original data
0011 0010 1010 1101 0001 0100 1011
• Now we know the grand mean, and we
know the estimated standard deviation
1
2
– From this we can calculate the location of the
“tails” of the distribution curve
– Add three of the estimated standard deviations
to the grand mean for the upper tail
– Subtract three of the estimated standard
deviations to the grand mean for the lower tail
4
0011 0010 1010 1101 0001 0100 1011
1
2
4
0011 0010 1010 1101 0001 0100 1011
• Now, let’s add the upper and lower
specification limits to the curve
1
– We know that the upper limit is 14 seconds
– We can assume the lower limit is 0 seconds
2
4
0011 0010 1010 1101 0001 0100 1011
1
2
4
0011 0010 1010 1101 0001 0100 1011
• Now we can see that some of the data is
outside the limits that have been set, but
how much?
1
2
4
Z scores
0011 0010 1010 1101 0001 0100 1011
• We are going to analyze the data and
determine how badly we are out of
specification.
1
– Z scores will help us to determine that
2
4
0011 0010 1010 1101 0001 0100 1011
Zupper = USL - X
σ^
1
2
4
0011 0010 1010 1101 0001 0100 1011
• Zupper = 14 – 10.00 / 2.00 = 2.00
1
2
• Now we will look at the following table to
see what percentage that is
• We see that the Z score is .0228, which
when changed to a percentage is 2.28%
4
0011 0010 1010 1101 0001 0100 1011
1
2
4
0011 0010 1010 1101 0001 0100 1011
• The Lower is done in a similar manner
Zlower = X - LSL
σ^
1
2
4
0011 0010 1010 1101 0001 0100 1011
• Zlower = 10.00 - 0 / 2.00 = 5.00
1
2
• Now we will look at the following table to
see what percentage that is
• We see that the Z score is so small that it is
insignifican, or it is 0%
4
0011 0010 1010 1101 0001 0100 1011
• Total percent outside the limits is
– 2.28% + 0% = 2.28%
1
2
• From this information, we can be pretty
sure that if we look at this process
continually, we will probably find that
2.28% of the time, we will be over the 14
seconds we are limiting the process to.
4
0011 0010 1010 1101 0001 0100 1011
• Now lets see if the process itself can
provide the limited time allowed
1
2
– We are going to look at the width of the
calculated spread vs. the specified spread
– The allowed spread is the difference between
the lower specification limit and the upper
specification limit
4
0011 0010 1010 1101 0001 0100 1011
• USL – LSL = 14
• Now we will take that specification spread
and divide it by six times the estimated
standard deviation
2
4
– Why six times the standard deviation
• 6 x 2.00 = 12
• 14 / 12 = 1.17
1
0011 0010 1010 1101 0001 0100 1011
• Cp = USL - LSL
6 σ^
1
2
4
0011 0010 1010 1101 0001 0100 1011
1
2
4
0011 0010 1010 1101 0001 0100 1011
• From the previous slide, it is obvious that
we want to make sure that the actual spread
is less than the specified spread
1
– A Cp of more than one is desirable, and the
higher the better
2
4
0011 0010 1010 1101 0001 0100 1011
• The only problem with this is we can have a
Cp of 2, but still be outside the limits of the
specification
• So there is another calculation that will
measure this. It is called Cpk
1
2
4
0011 0010 1010 1101 0001 0100 1011
• Cpk is simple. It is the smallest of the Z
scores, or Zmin, divided by 3
• In our case, that would be 2.00 / 3 = .67
• As in the Cp, we would want the result to be
greater than 1.
1
2
4
0011 0010 1010 1101 0001 0100 1011
• So in our example we had a Z score of 2.00,
which told us that 2.28% of the time we
were more than 14 seconds
• We also know that the process spread was
1.17, which means we can maintain the
proper time
• But we also found that the Cpk was .67,
which tells us we are outside of the limits
1
2
4
0011 0010 1010 1101 0001 0100 1011
• Our conclusion?
– We have a good process, but we need to find a
way to correct it so that we are well within the
limits as specified.
1
2
4