Survey
* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project
* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project
The DMAIC Process Detail The Measure Phase © Max Zornada (2005) Slide 1 DMAIC Process Storyboard DEFINE TEAM FORMATION Objective: Select problem/ opportunity theme, select team members MEASURE Objective: Identify and implement the measures required to establish baseline performance and quantify the opportunity. Key Steps:Cause and Effect Diagram Run chart or Objective: Define the Problem/Opportunity, Customers, Customer Requirements, and Process. Team charter Key Steps:Flowchart •Determine what to measure •Understand the measures •Understand Variation •Assess measurement system •Assess process performance •Develop business case •Develop project team charter •Understand Customer Requirements •Understand the Process. Output: Problem/Opportunity selected, Team members selected. Output: Team Project Charter, Work Plan, Measurable Output: A quantified picture of the current process Customer Requirements, Process Map/Process Analysis performance, problem impact. The process sigma rating. ANALYSE IMPROVE - I : Generate Potential Solutions Objective: Identify and verify the root cause(s) of the problem. Key Steps:•Analyse data •Analyse process •Determine potential root causes •Hypothesis Testing •Verify root causes Cause and Effect Diagram (Fishbone) Checksheet Pareto Chart Output: Root cause(s) identified. IMPROVE - II: Implement and Check Objective: Implement the preferred solution. Confirm that the problem and its root cause(s) have been reduced or eliminated. Key steps:•Implement preferred After solution •Verify effectiveness •Apply comparative methods if necessary. Output: Confirmation that the best solution to eliminate the problem & its root cause(s) has been implemented. Before control chart Objective: Determine possible solutions that will address the identified root cause(s) of the problem. Key Steps:Potential Solutions Action Plan •Generate potential solutions •Assess potential solutions •Select preferred solution •Test/Pilot preferred solution •Develop implementation plan Output: Preferred solution or countermeasures Step 1 Step 2 Step 3 Step 4 Step 5 Step 6 CONTROL - Standardise Objective: Prevent the problem and its root cause from recurring. Flowchart Key steps:Standard procedure •Standardise the solution (standards & procedures) •Document project •Implement scorecard •Implement controls Output: Solution embedded and “routinised” in relevant process, procedures and standards. © Max Zornada (2005) FUTURE PLANS Objectives: Review team effectiveness, plan to address remaining issues and institutionalise the learning. D C Key Steps:Define, I M •Review remaining project opportunities •Review other applications •Review learnings Measure, Analyse, Improve, Control A Output: Recommendations for future projects and improvements to team processes. Project documentation and learnings “pack” Slide 2 Overview of the Measure Phase Determining what to Measure; Understanding and Describing Data; Understanding and Managing Variation; Statistical Process Control; Process Capability and Sigma Level; Overview of Sampling © Max Zornada (2005) Slide 3 Determining what to Measure © Max Zornada (2005) Slide 4 Y = f(x) The fundamental equation that drives Six Sigma; Output (Y) is a function of the Inputs and the Process Examples of Y Examples of x’s - x1, x2, x3 …. xn Output Inputs to the process Outcomes Leading indicators/Drivers Effect Problems and their causes Symptom “Noise” factors A Dependent Variable Complexity A Key Performance Indicator Independent Variables Control and “levers” © Max Zornada (2005) Slide 5 Approaches to identifying measures (and what data to collect) The process scorecard “generic” template; Cause and effect (fishbone diagram) around the CTQ outcome of interest; The Critical-to-Quality (CTQ) Tree approach; Correlation analysis between measures and outcomes; The Measurement Assessment. © Max Zornada (2005) Slide 6 Consider a Generic Process Real Work Stream Complexity Stream Complexity Stream Customer © Max Zornada (2005) Slide 7 Measuring Process Performance Supplier Interface Measures? Process Performance Measures ? Customer Customer Interface Measures? © Max Zornada (2005) Slide 8 Consider HTLC: Typical Measures Things like …. Supplier Interface Measures Incoming stock delivery performance Credit check turnaround time Backlog of items on order Outputs Process Inputs Process Measures Time to process an order Customer Interface Measures Customer Satisfaction Cost Number of orders in the system (work in progress) Complaints Orders not delivered on time/Late Order backlog Overdue orders Number of orders filled. No. of Orders Received © Max Zornada (2005) Slide 9 A Generic Template for Developing Process Based Performance Measures Overall Organisation/Business Unit End-to-End Core Process Internal Process Outcomes Measures •Time •Cost •Quality/Waste © Max Zornada (2005) Customer Outcomes Measures •Delivery •Quality •Value External Customer Slide 10 A Scalable Concept But the specific measures developed will be different in each case. Workgroup A Subprocess Internal Supplier Internal Process Outcomes Measures •Time •Cost •Quality/Waste Customer Outcomes Measures •Delivery •Quality •Value Workgroup B Sub-process Customer Outcomes Measures •Delivery •Quality •Value Internal Customer Internal Process Outcomes Measures •Time •Cost •Quality/Waste © Max Zornada (2005) External Customer Slide 11 Inputs, Process and Outcomes Measures A draft “generic” template for process measurement Input Measures Customer Outcomes Measures (Inputs) •Delivery •Quality •Value Customer outcomes the supplier(s) to the process work to, in order to meet the input requirements for the process Customer Outcomes Measures •Delivery •Quality •Value Process Internal Process Outcomes Measures •Time •Cost •Quality/Waste © Max Zornada (2005) External Customer Slide 12 What can you do about …. A late delivery to ensure its delivered on time - after its already been delivered late? A cost over run - after it has already been incurred? Avoiding dissatisfying a customer - after they have already been dissatisfied? We need some predictive measures as well; These are referred to as leading indicators or drivers. © Max Zornada (2005) Slide 13 Consider our process The Real Work Stream gives us the optimum: Real Work Stream Customer Processing time/order (& cycle time) and hence delivery time; Work input and hence resources & sost/order; Nothing goes wrong. This will be the best this process can do. © Max Zornada (2005) Slide 14 Causes of Outcomes Drivers or Leading Indicators What would cause an order to take longer to be processes? What would cause an order to cost more to be processed? What would cause an order to be delivered late? What would cause the process to operate other than perfectly (only real work). Complexity! © Max Zornada (2005) Slide 15 Understanding complexity provides insight into what the leading indicators should be Real Work Stream Example: HTLC Eg. item in stock measure Backlog of items on Order Eg. technician availability measure % order rescheduled due to “technician not available” Customer © Max Zornada (2005) Slide 16 Shortcut method for identifying leading indicators (x’s) What are all of the things that could stop the process ( internally)? What external factors could stop the process from meeting its customer and process outcomes? Can you measure these? © Max Zornada (2005) Slide 17 The Process Scorecard A “generic” template for process measurement: Inputs, Process, Leading Indicator and Outcomes Measures Customer Outcomes Measures •Delivery •Quality •Value Input Measures Customer Outcomes Measures (Inputs) •Delivery •Quality •Value Leading Indicators (causes of the outcomes) Customer Outcomes Measures Process Customer Leading Indicators •Process specific issues •Key Complexity Issues •External to process issues Internal Process Outcomes Measures •Time •Cost •Quality/Waste © Max Zornada (2005) Internal Process Outcomes Measures Slide 18 Process Measurement: Y = f(x) Customer Outcomes Measures •Delivery •Quality •Value Input Measures Customer Outcomes Measures (Inputs) •Delivery •Quality •Value Usually the Y’s Process Customer Leading Indicators •Process specific issues •Key Complexity Issues •External to process issues Internal Process Outcomes Measures •Time •Cost •Quality/Waste Potential x’s found here © Max Zornada (2005) Slide 19 Fishbone Diagram HTLC Example Inputs Customer Outcomes Complexity Issues Item not in stock Customer complaints Order incorrectly specified Promised date too soon Refund/Penalty claims Technician not available No transport Delivery Performance Supplier didn’t supply Return trips to same customerTraffic conditions Processing time too long Process Lead Time too long Other External Impacts © Max Zornada (2005) Can we get measures for these things ? Process Outcomes Slide 20 Identifying Measures We can potentially generate lots of measures; Only a small number of measures may actually matter; Do we measure everything? Problem is identifying which ones? We can focus our measurement and data collection by considering the relationships between various potential measures and the CTQ outcome which is the focus of the improvement effort; Build a measures correlation matrix. © Max Zornada (2005) Slide 21 Measures Correlation Matrix Potential Input/Process/LI Measure (x’s) Outcome Measure (Y) On-time delivery Right Equip Equipment Delivered Total Works No. of Order with item Out of Stock 10 9 9 3 7 0 12 No. of Orders for Non-Stocked Item 1 3 3 15 No. of Order on Backlog 9 0 1 10 9 0 0 9 No. of Orders with no transport allocated 9 0 0 9 Time of day order delivered 1 0 0 1 Quoted Lead Time per Order 3 0 0 3 No. of Orders incorrectly specified 3 9 3 15 Processing time 1 0 0 0 Importance No. of Orders with no tech allocated © Max Zornada (2005) Slide 22 Measures Correlation Matrix Potential Input/Process/LI Measure (x’s) Outcome Measure (Y) On-time delivery Right Equip Equipment Delivered Total Works No. of Order with item Out of Stock 10 9 9 3 7 0 12 No. of Orders for Non-Stocked Item 1 3 3 15 No. of Order on Backlog 9 0 1 10 9 0 0 9 No. of Orders with no transport allocated 9 0 0 9 Time of day order delivered 1 0 0 1 Quoted Lead Time per Order 3 0 0 3 No. of Orders incorrectly specified 3 9 3 15 Processing time 1 0 0 0 Importance No. of Orders with no tech allocated © Max Zornada (2005) Slide 23 The Tree Diagram Approach to Identifying Measured © Max Zornada (2005) Slide 24 Identifying Measures using a CTQ Tree HTLC Example Ontime delivery performance The Outcome Measure (Y) © Max Zornada (2005) Slide 25 Identifying Measures using a CTQ Tree HTLC Example Brainstorm things that could affect the outcome. Refer previous analysis. Ontime delivery performance The Outcome Measure (Y) Credit check turnaround time Stock Availability Technician Availability Transport Availability © Max Zornada (2005) Slide 26 Identifying Measures using a CTQ Tree HTLC Example Brainstorm things that could affect the outcome. Refer previous analysis. Ontime delivery performance The Outcome Measure (Y) Credit check turnaround time Orders received Stock Availability Technician Availability Transport Availability © Max Zornada (2005) Orders filled from stock Order placed on backlog Identify measures for each of the things that can affect the outcome (the x’s) Slide 27 Deciding what data to collect Two basic uses of data: Monitoring: Aggregate data: data used to tell you at what level the process is operating and to indicate when something has changed; Improvement: Disaggregate or Stratify: need to be able to identify specific linkages between data elements and sources; © Max Zornada (2005) Slide 28 Using a Measurement Assessment Tree Questions we want to answer about the process Output (Y) Stratification Factors (x variables) Measures Y By time period N Y # late, by hour of day Y # late, regular contract By type of customer # late, casual sales Y How many are late? Are there trends of patterns? # late, by day of week Y No. of late deliveries How much is it costing? Does this metric potentially help predict the output Y? By location By value Does data exist to obtain this metric N Y Y Y Y © Max Zornada (2005) # late, by region # late, by suburb Y Y Y Y # late, by distance from w/h # late, small orders # late, large orders Slide 29 N N N Operational Definitions A measurement must give consistent results no matter who does the measuring; An Operational Definition gives a description of what something is and how it is measured; Therefore, before data is collected, we must agree on the operational definition of all terms and on the measurement criteria to be used. © Max Zornada (2005) Slide 30 Normalisation Correcting for different scales of measurement is called normalisation; Normalisation allows us to compare two groups of data, where the raw data may have been collected in different ways; e.g. different time frame, different units, different sample size. Data are usually normalised by time, volume or task. © Max Zornada (2005) Slide 31 Understanding Data and Variation © Max Zornada (2005) Slide 32 Hi-Tech Leasing Corporation Monitoring the performance of the Order Fulfillment Process © Max Zornada (2005) Slide 33 The Funnel Experiment H A G F E Flip chart paper B Target C Radius of circle = 4 cm Defines on time delivery zone Result of an individual pencil drop D Late deliveries zone (outside the circle) Distance from centre in cm = days to fill the order © Max Zornada (2005) Slide 34 The funnel experiment The Funnel experiment is used to simulate the performance and an order fulfillment process; The company promise customers delivery within 4 days; Each pencil drop through the funnel represents an order going through the process; The distance the drop lands from the target is measured in centimeters. This represents how long that particular order took to fill in days. Orders landing inside the inner circle of radius 4 cm represent orders delivered within the service standard; Order landing outside the inner represent late deliveries; The zones labeled A, B, C, D, E, F, G, H represent the different reasons for which the delivery was late. © Max Zornada (2005) Slide 35 The funnel experiment The Rules Group 1: Aim for the target, lock in the settings and take 50 shots for the target without re-targeting. Group 2: Same as group 1, except if you miss, you can re-target to improve your chances next time. Group 2: Aim each shot where the last one landed. Measure the 1st 25 shots and record the measurement on the worksheet. Note: Measurements must be made in the order in which they occur. © Max Zornada (2005) Slide 36 Data Collection Sheet: The Funnel Experiment Drop Value (X) 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 Drop 16 17 18 19 20 21 22 23 24 25 Value (X) Type the information directly into an Excel Worksheet to simplify calculations. Total Average © Max Zornada (2005) Slide 37 Types of Data Continuous or variables data is data that is measured on a continuous scale and can take any value i.e. can take values that are not whole numbers (e.g. 3.1, 7.5, 8.9 etc.); Examples include measure such as:- time, weight, temperature, length, money etc.; Discrete data is data that is measured by counting things. It can only assume a countable number of values. Can also be referred to as attribute data when used to count items with a particular characteristic or attribute; Examples include: Percent defective, number of errors, customer satisfaction ratings on 1-5 scale, models of car. © Max Zornada (2005) Slide 38 Populations and Samples A population is the body of data that we want information about; In our case, all of the orders processed by HTLC; From the population we select a sample; In our case, our sample is the number of orders for which we collected data about (25 orders); The idea is to be able to analyse the sample to get information about the population. © Max Zornada (2005) Slide 39 Collecting and Displaying Discrete Data © Max Zornada (2005) Slide 40 The Anatomy of a Check Sheet Example: Lost Time Injury Data for a Security Firm Progressive “Tally” count of each occurrence, collected from raw data or directly from the field Note: = 1, =5 The categories in which the data occurs Category Numerical frequency count or summary of the “tally” or occurrences in each category Frequency Tally 1 Allergy 8 2 Broken limb 12 3 Back/Neck 6 4 Concussion 14 5 Contusion 7 6 Heart Attack 1 7 Sprains 23 8 Other 3 Total Overall Total © Max Zornada (2005) 78 Slide 41 Frequency/ Number of Occurrences Check Sheet Data can be displayed using a Pareto Chart 30 23 20 14 12 10 8 7 6 3 © Max Zornada (2005) 1 Slide 42 The Pareto Principle (Also know as the 80/20 Rule): “A few causes account for most of the effect” Trivial Many 20% of the causes account for 80% of the effect! Critical Few Causes Effect © Max Zornada (2005) Slide 43 Check Sheet for Funnel Experiment Reason category 1 A 2 B 3 C 4 D 5 E 6 F 7 G 8 H Tally © Max Zornada (2005) Frequency Slide 44 Pareto Chart Proforma © Max Zornada (2005) Slide 45 Displaying Continuous Data © Max Zornada (2005) Slide 46 Representing Data Graphically Data can be represented graphically by using tools such as Dot Plot Histogram Individual data points 20 17 15 13 10 10 1 2 3 4 5 6 7 5 Scale of allowable values 0 2 2 1-12 13-24 Y-Axis: Magnitude scale 3 25-36 3 37-48 49-60 61-72 73-84 Individual data points Run Chart © Max Zornada (2005) X- Axis: Time based scale Slide 47 Histogram 20 17 15 13 10 10 5 2 0 1-12 2 3 3 13-24 25-36 37-48 49-60 61-72 73-84 Determining the number of class intervals Class intervals Number of data values Number of Class Intervals Under 50 5 to 7 50 to 100 6 to 10 100 to 250 7 to 12 Over 250 1 0 to 20 © Max Zornada (2005) Slide 48 Exercise: Displaying Data Construct a Dot Plot of your Data manually; Use Excel to construct a Histogram and a Run chart. © Max Zornada (2005) Slide 49 Order Fulfillment Process Performance – Dot Plot © Max Zornada (2005) Slide 50 Describing Data Location – where is it? Measured by: Spread – how spread out are the points? (Variation) Measured by: Mean Median Mode Range Interquartile Range Standard Deviation Shape – what does it look like? Bell shape Skewed Uniform © Max Zornada (2005) Slide 51 Some Common Data Shapes Bell Shaped Skewed "Right" "Toothlike" Bi-Modal © Max Zornada (2005) Slide 52 Measures of Location Refer Tools and Techniques Guide Book Appendix Mean or Average Median Mode Exercise: Calculate the above three measures of location for your data. © Max Zornada (2005) Slide 53 The Median The Median is the Middle Value in an Ordered Sequence If odd number of values = middle value of sequence If even number of values = average of 2 middle values Position of Median in Sequence is = (n+1)/2 For an odd number of data points Raw Data: 9, 6, 4, 2, 4, 2, 4, 7, 2, 4, 3 Ordered Data: 2, 2, 2, 3, 4, 4, 4, 4, 6, 7, 9 Position: 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11 Median = the 6th point = 4 Even number of data points Raw Data: 9, 6, 4, 2, 3, 2, 4, 7, 2, 3, 3, 9 Ordered Data: 2, 2, 2, 3, 3, 3, 4, 4, 6, 7, 9, 9 Position: 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12 Median = the average of 6th & 7th points = (3+4)/2 = 3.5 © Max Zornada (2005) Slide 54 Mean/Median - Advantages/Disadvantages Advantages of the mean: Easy to calculate and interpret Based on all of the data Disadvantages of the mean: Only gives good indication of location for symmetrical data Not a good indicator of location if there are extreme points in the data or is the data distribution is skewed. Advantages of the median: Easy to obtain and interpret Unaffected by extreme points and skewed distributions Disadvantages of the median: Only based on the “middle” data point(s) © Max Zornada (2005) Slide 55 Measures of Spread or Variation Range Inter-quartile Range Variance Standard Deviation © Max Zornada (2005) Slide 56 Understanding process performance I Calculate the range for your data. Review the graphical representations and measures of location from previous exercises. What performance would you be prepared to guarantee your customers? © Max Zornada (2005) Slide 57 Consider the outputs from 3 different systems Range Mean The systems producing these outputs are obviously different; However, the mean, median and range of each are the same; These summary measures do not allow us differentiate between these distributions; We need another measure that does. The Standard Deviation! Median © Max Zornada (2005) Slide 58 The Standard Deviation Mean The standard deviation can be thought of as a measure that represents the average of the distance of all of the points from the mean. © Max Zornada (2005) Slide 59 Getting a little more sophisticated: The Standard Deviation Is a measure of the degree to which average process performance represents typical process performance n Standard Deviation (S) i 1 The Greek symbol Sigma is used to refer to the population standard deviation. (X X) 1 2 (X 2 (X i X) 2 n 1 X) 2 .... (X n X) 2 n 1 s 2 Where S2 is called the Variance © Max Zornada (2005) Slide 60 The Interquartile Range The interquartile range - IQR, (or hinge spread) is the spread of the middle 50% of the data; IQR = UQ - LQ where LQ is the lower quartile, or the middle of the lower half of the data, and, UQ is the upper quartile, or middle of the upper half of the data. © Max Zornada (2005) Slide 61 Calculating the Interquartile Range for an Even Number of Data Points Lower Half of Data Ordered Data: Position: Upper Half of Data 2, 2, 2, 3, 3, 3, 4, 4, 6, 7, 9, 9 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12 LQ = Middle or Median of the lower half of the data = (2+3)/2 = 5/2 LQ = 2.5 UQ = Middle or Median of the upper half of the data = (6+7)/2 = 13/2 UQ = 6.5 IQR = UQ - LQ = 6.5 - 2.5 IQR = 4 © Max Zornada (2005) Slide 62 Calculating the Interquartile Range for an Odd Number of Data Points Upper Half of Data Lower Half of Data Ordered Data: 2, 2, 2, 3, 4, 4, 4, 4, 6, 7, 9 Position: 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11 UQ = Middle or Median of the upper half of the data = (4+6)/2 = 10/2 UQ = 5 LQ = Middle or Median of the lower half of the data = (2+3)/2 = 5/2 LQ = 2.5 IQR = UQ - LQ = 5.0 - 2.5 = 2.5 © Max Zornada (2005) Slide 63 Exercise From your data calculate of the performance of the order fulfillment process. Calculate the following: Variance Standard Deviation Lower Quartile Upper Quartile Interquartile Range © Max Zornada (2005) Slide 64 The Box Plot Displaying measures of location and spread as a complete package A boxplot is a graphical way of displaying information about the spread and location of data; A boxplot focuses attention on certain features of the data without having to plot all the values. E.g. the presence of extreme points or skewness in the data; Box can provide a quick way of assessing data for which the team may be considering developing a control chart. © Max Zornada (2005) Slide 65 Box Plot Example Waiting times at a medical clinic Waiting time (in minutes) at a medical clinic is measured by taking 5 samples at random during the mid-morning of each day for a week. Cumulative sum Frequency Time Tally 1 2 3 4 5 6 7 8 9 10 12 15 Total 1 2 2 3 5 2 2 3 2 1 1 1 25 1 3 5 8 13 15 17 20 22 23 24 25 Sample 1 Sample 2 Sample 3 Sample 4 Sample 5 Mon Tues Wed Thurs Fri 6 4 5 15 5 1 10 2 7 3 2 6 9 5 8 12 8 7 4 3 4 9 4 8 5 Total time = 153 minutes Lower quartile (Q1) = 7th point = 4 Median = 13th point = 5 Upper quartile (Q3) = 19th point = 8 IQR = UQ - LQ = 8 - 4 = 4 minutes Mean = 153/25 = 6.1 minutes © Max Zornada (2005) Slide 66 The Anatomy of a Box Plot The Box Plot for Waiting time data Lowest data point inside the Lower Inner Fence Highest data point inside the Lower Inner Fence Whiskers IQR “Outlier” * -2 -1 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 1.5 X IQR = 6 1.5 X IQR = 6 LQ = 4 Median = 5 Lower Inner Fence = LQ - (1.5 X IQR) = 4 - 6 = -2 15 UQ = 8 Mean = 6.1 Upper Inner Fence = UQ + (1.5 X IQR) = 8 + 6 = 14 © Max Zornada (2005) Slide 67 The Box Plot for Waiting time data Conclusions * -2 -1 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 Box Plot suggests that the data distribution is skewed to the right Evidence: 15 Mean is not equal to the Median The Median is less than the Mean The Whiskers are not equal, the right one is longer than the left one. There is an outlier - the 15 minute point. This represents an unusual point which is not typical for the system. © Max Zornada (2005) Slide 68 Presenting Data with Box Plot Vertical Presentation Horizontal Presentation * -2 -1 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 © Max Zornada (2005) 15 14 13 12 11 10 9 8 7 6 5 4 3 2 1 0 -1 -2 * Slide 69 Exercise: Box Plots Construct a Box Plot the Order Fulfilment delivery data simulated by the funnel experiment. © Max Zornada (2005) Slide 70 Box plots Are a graphical representation of data that show, location, spread, symmetry or skewness and whether there are any outliers; Need to have at least 5 distinct data values to draw a box plot. Box plots can be used to compare two or more samples of data; If the boxes do not overlap, then there are statistically significant differences between samples (need at least 10 data points for each box plot for comparisons). © Max Zornada (2005) Slide 71 The Normal Distribution Average = Median 2s = 68.26 % of Data 4s = 95.44 % of Data s = Sample Standard Deviation 6s = 99.73 % of Data σ = Population Standard Deviation © Max Zornada (2005) Slide 72 The Box Plot interpretation The Normal Distribution The Whiskers are equal in length. The Mean = Median and is in the Middle of the Box A box plot can be used as a test of whether data is normally or close enough to normally distributed. © Max Zornada (2005) Slide 73 Understanding Variation Dr. W.A. Shewhart 1920's, found: All processes display variation; Some display controlled variation; Some display uncontrolled variation. © Max Zornada (2005) Slide 74 Common vs Special Causes Dr. W. Edwards Deming called this: Variation due to common causes. Due to the random interaction of the many variables in the system or process. These were purely random and are built in to the system. People working in the system have no control over these. They are effectively "prisoners" of the process. Variation due to special causes. Not a natural part of the system. These could be traced to a specific event, person, machine or localised condition. © Max Zornada (2005) Slide 75 Sources of Variation Common Causes: Variation caused by the random interaction of all of the variables that are built in to or are a “normal” part of the process. This represents “business as usual” performance for the process; Adjusting the process increases this type of variation. Special Cause: Non-random variation. May exhibit a pattern; Due to something happening that is not a normal for the process - a problem, event, a specific cause that can be found and explained; Adjusting the process decreases this type of variation. © Max Zornada (2005) Slide 76 The Normal Distribution Average = Median 2s = 68.26 % of Data 4s = 95.44 % of Data s = Sample Standard Deviation σ = Population Standard Deviation 6s = 99.73 % of Data © Max Zornada (2005) Slide 77 The Path to Continuous Improvement The path to continuous improvement in quality and productivity, is to: Train and empower employees to identify and remove special causes from work processes; Train management in Variation principles and management change or reengineer processes and systems to reduce variation from common causes; Alternately: - Management empowers teams to work on changing the system and processes to remove common causes - continuously. © Max Zornada (2005) Slide 78 Characteristics of Stable Processes Processes that are free of special causes are referred to as stable or in statistical control; Future outcomes of the process can be predicted to occur within the defined (natural) limits of the process, based on past results. Therefore, the process capability to meet customer expectation can be predicted; At any given time, the outcome is random within the defined limits; Because process performance is known and stable, process costs can be reliably predicted. © Max Zornada (2005) Slide 79 Characteristics of Stable Processes Effects resulting from changes to the processes can be measured reliably and quickly; Once in control, some processes are capable of meeting the customer requirements, while others are incapable; Once in control, process improvement activities should focus on making processes capable by reducing the common causes; If specification limits are to be changed, the necessary data is available to understand the likely implications. © Max Zornada (2005) Slide 80 A Control Chart Upper Control Limit (UCL) Process Mean or Average Lower Control Limit (LCL) UCL = Mean + 3 Standard Deviations LCL = Mean - 3 Standard Deviations © Max Zornada (2005) Slide 81 Stable & Unstable Processes A Stable Process An Unstable Process UCL Mean LCL © Max Zornada (2005) Slide 82 Process Improvement by Managing Variation Customer Requirement Process Performance Special cause Special cause USL = Upper Specification Limit LSL = Lower Specification Limit (These are customer specified) USL % of output meeting customer requirements (unpredictable because of presence of special causes) Special cause Process Unstable and Incapable Process Unstable - special causes present Variation too wide Off-target (not “centred”) Target LSL Bad Good © Max Zornada (2005) Slide 83 Process Improvement by Managing Variation Step 1. Make the process stable i.e. remove special causes Process Performance Customer Requirement USL Target LSL Process Stable but Incapable Variation too wide Off-target (not “centred”) © Max Zornada (2005) % of output meeting customer requirements (predictable because of absence of special causes) Slide 84 Process Improvement by Managing Variation Step 2. Re-target to aim at customer target (shift the average) Process Performance Customer Requirement USL Target LSL Process Stable but Incapable Variation too wide Now On-target (“centred”) © Max Zornada (2005) % of output meeting customer requirements (predictable because of absence of special causes) Slide 85 Process Improvement by Managing Variation Step 3. Reduce variation to make process capable Process Performance Customer Requirement USL Target LSL Process Stable and Capable Variation Less than Customer limits On-target (“centred”) © Max Zornada (2005) 100 % of output meets customer requirements I.e. a “Zero Defects” process! Slide 86 Exercise :The Funnel Experiment Construct a control chart of the performance of the order fulfilment data. What can you say about the performance of the process. © Max Zornada (2005) Slide 87 The Shewhart Tests: Test 1-to-4 Test 1. One point beyond zone A Test 2. Nine points in a row in Zone C or beyond A B C C B A A B C C B A Test 3. Six points in a row steadily increasing or decreasing A B C C B A Test 4. Fourteen points in a row alternating up and down A B C C B A © Max Zornada (2005) Slide 88 The Shewhart Tests: Test 5-to-8 Test 5. Two out of three point in a row in zone A or beyond. Test 6. Four out of five points in a row in zone B or beyond. A B C C B A A B C C B A Test 7. Fifteen point in a row in zone C (above & below the centreline) A B C C B A Test 8. Eight points in a row on both sides of centreline with none in zone C. A B C C B A © Max Zornada (2005) Slide 89 A 6 Sigma Control Chart? Customer Specification A Six Sigma Process Upper Specification Limit (USL) UCL Mean LCL Target Level of Performance Lower Specification Limit (LCL) UCL = Mean + 6σ LCL = Mean - 6σ Probability of getting a defect when process in control = 3 in a Million © Max Zornada (2005) Slide 90 Process Capability and Sigma © Max Zornada (2005) Slide 91 Process Capability Cp = Specification Width Process Width = Process Standard Deviation = USL - LSL 6 USL = Upper Specification Limit LSL = Lower Specification Limit A measure of a process's ability to meet or exceed the customer's specification; A capable process must have a Cp of at least 1.0 Does not look at how well the process is centered in the specification range; Target value of Cp = 1.33. Allows for off-center processes Six Sigma quality requires a Cp = 2.0 © Max Zornada (2005) Slide 92 Process Capability For processes know to be not centred C pk = Min (USL - m , m - LSL) 3 3 USL = Upper Specification Limit LSL = Lower Specification Limit = Process Standard Deviation m Process Average A capable process must have a Cpk of at least 1.0 A capable process is not necessarily in the center of the specification, but it falls within the specification limit at both extremes. © Max Zornada (2005) Slide 93 Interpreting Cpk Cpk = negative number Cpk = zero Cpk = between 0 and 1 Cpk = 1 Cpk > 1 © Max Zornada (2005) Slide 94 A Capable and Stable Process Customer target Upper Specification Limit Lower Specification Limit Process mean 0.135% 1,350 ppm 0.135% 1,350 ppm 3 3 = 99.73% of data inside the limits (Cp=1) 0.27% of points will be outside of the specification limits ie. defects (= 3/1000 or 2,700 parts per million (ppm) out of spec.) © Max Zornada (2005) Slide 95 A Capable, Stable, 6 Sigma Process Customer target Upper Specification Limit Lower Specification Limit 0.00017% 1.7 ppm 0.00017% 1.7 ppm 6 6 = 99.99966% of data inside the limits (Cp = 2) 0.00034% of points will be outside of the specification limits ie. defects (= 3.4 parts per million out of spec.) © Max Zornada (2005) Slide 96 A closer look at Defects Unit: The entity that is transformed by value-adding activities. Defect: A countable failure associated with a single unit. A single unit can be found defective as a result of having one or more defects; Defects Per Unit (DPU) = Defects/Units produced Defectives: Completed units of work that are classified as “bad”. A single unit can be found “defective” regardless of the number of defects. Yield = Non-defectives/Total Unit © Max Zornada (2005) Slide 97 FSG Revisited Customer Support Operations Assume customer support receive calls re: 8% of transactions 6% they fix, 2% get passed onto PIT. This means 8% defects must be getting through the CC Team and onto the customer. 2% or transactions Customer Support (Call Centre) Problem Investigation Team 2% internal errors Document Receiving Team Customer Communications Team Transaction Processing Team 8% rework © Max Zornada (2005) Slide 98 Yields The process’s Final Yield is 92% - the is the percentage of good units making it to the customer; However, of the work done by the processing team they find: 2 defectives internally; 8 come back from CCT 6 get fixed by Call Centre 2 Come back from PIT. First Pass Yield = 82% © Max Zornada (2005) Slide 99 Six Sigma and DPMO Defects Per Million Opportunities is the key Six Sigma Metric DPMO = Total Defects Total Opportunities © Max Zornada (2005) X 1,000,000 Slide 100 DPMO An opportunity is any opportunity to produce a defect e.g. a ten step process with one activity at each step, provides 10 opportunities to get something wrong; Allows us to compare products and processes of differing complexity. © Max Zornada (2005) Slide 101 Calculating Sigma Level Select the process, unit and requirements: Identify the process you want to evaluate? (Process) What is the “thing” produced by the process? (Unit) What are the customer requirements for the thing? Define what a “defect” is and the number of opportunities: Possible defects: How many defects could be found on a single unit? Collect Data and Calculate DPMO Collect end-of-process data. No. of Units counted Total defects counted Determine total opportunities in data collected: Units counted X Opportunities/Unit = Total Opportunities Calculate DPMO = (Defects/Total Opportunities) X 106 = DPMO Convert to a Sigma Level using a conversion table. Estimated Sigma Level = © Max Zornada (2005) Slide 102 Sigma Conversion Table © Max Zornada (2005) Slide 103 Statistical Process Control © Max Zornada (2005) Slide 104 Types of Control Charts Measuring Not much data available i and mr Charts Lots of data available X and R Charts Continuous Data Are you counting or measuring ? Constant Sample Size Counting Categorical Data No Can you have more than one count per unit np Chart Sample size not constant p Chart Constant Sample Size Yes Occurrence Data c Chart Sample size not constant © Max Zornada (2005) u Chart Slide 105 X and R Charts Use for continuous data, when degree of normality of process is uncertain; Relies on taking samples and plotting sample averages; We plot 2 charts together: The X chart to monitor variation in location; Plot the average of each sample X. The R chart to monitor variation in spread. Plot the range of each sample R. © Max Zornada (2005) Slide 106 The Central Limit Theorem The Central Limit Theorem says that averages of groups of data will be more normally distributes that the individual data points; The more points you sample, the closer those averages will be to the normal distribution; If you sample up to 30 points, the distribution of those 30 points will effectively be normal; This means that even non-normal data can be plotted using control charts, by taking averages of sub-groups of data. © Max Zornada (2005) Slide 107 X and R Charts The X Chart: Centreline = X = = Average of the sample averages X1 + X2 + …. Xn Total number of samples taken (m) We plot the sample averages - X as the individual data points Upper and Lower Control Limits UCL = X + A2R LCL = X - A2R The average of the Ranges A constant we look up in a table © Max Zornada (2005) Slide 108 X and R Charts The R Chart: Centreline = R = = Average of the sample ranges R1 + R2 + …. Rn Total number of samples taken (m) We plot the sample averages - R’s as the individual data points Upper and Lower Control Limits UCL = D4R LCL = D3R D3 and D4 are constants we look up in a table © Max Zornada (2005) Slide 109 Values of Constants for X & R Charts Sub-group size (n) 2 3 4 5 6 7 8 9 10 X Chart A2 2.66 1.023 0.729 0.577 0.483 0.419 0.373 0.337 0.308 R Chart LCL R Chart UCL D3 0 0 0 0 0 0.076 0.136 0.184 0.223 © Max Zornada (2005) D4 3.267 2.574 2.282 2.114 2.004 1.924 1.864 1.816 1.777 Slide 110 Chebyshev’s Rule Regardless of the shape of the distribution: At least 75% of the distribution will fall within 2 standard deviations of the mean; At least 8/9ths (88.89%) of the distribution will fall within 3 standard deviations of the mean;. Chebyshev’s rule is generally used when the underlying distribution is unknown, except for the mean and the variance. © Max Zornada (2005) Slide 111 Data Collection Sheet: Attributes Data No. 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 np c No. np c 16 17 18 19 20 21 22 23 24 25 Total Average © Max Zornada (2005) Slide 112 The np-chart (Number Defective) Use when we are measuring the number of nonconforming or “defective” items in each sample; The number of defective items in each sample is np; The individual data points we plot on the control chart are the np’s; Sample size must be constant. © Max Zornada (2005) Slide 113 The np-chart (Number Defective) Average proportion defective: np = Total number of defectives Number of samples Average proportion defective: p= Total defective= Total inspected Snp Sn Upper and Lower Control Limits UCL = np + 3 np(1-p) LCL = np - 3 np(1-p) n © Max Zornada (2005) Slide 114 The p-chart (Proportion Defective) If sample size will not be constant, makes more sense to plot % defective; Convert the individual data points (np) to p (%) by dividing each by the sample size (n). Plot the p’s. If sample size changes, we need to recalculate the control limits for each new value of sample size; Therefore, control limits will only be straight lines if the sample size remains constant. © Max Zornada (2005) Slide 115 The p-chart (Proportion Defective) Average proportion defective: p= Total defective= Total inspected Snp Sn Upper and Lower Control Limits UCL = p + 3 p(1-p) n LCL = p - 3 p(1-p) n © Max Zornada (2005) Slide 116 Categorical Data Categorical data tends to follow the Binomial Distribution: Finite - known number of identical samples collected under identical circumstances; Only two possible outcomes e.g. OK/Not OK Probability remains constant Data points are independent; If np > 5 and n(1-p) > 5: The approximately Normal. © Max Zornada (2005) Slide 117 The c-chart Used to count occurrences when we don’t know the nonoccurrences eg. Record customer complaints but don’t know how many customers didn’t complain (ie. we can’t work out a %) When we can get multiple counts per unit of measure; We can count defects but not non-defects; e.g. a form may be counted as defective because someone made one error or several errors. The sample “frame” must be remain constant: e.g. complaints/day - for 8 hour day sample frame. © Max Zornada (2005) Slide 118 The c-chart Average non-conformities: c= Sum of all data points = Number of samples c1 + c2 + …. cn number of samples Upper and Lower Control Limits UCL = c+3 c LCL = c-3 c © Max Zornada (2005) Slide 119 The u-chart The same as a c - chart when the sample “frame” is not constant: e.g. complaints/day - but not all days are 8 hours long, due to late night shopping on some days and shorter opening hours on week ends. Every time the sample frame changes, we have to recalculate the control limits. © Max Zornada (2005) Slide 120 The c-chart Average non-conformities: u= Sum of all data points = Number of units sampled c1 + c2 + …. cn n1 +n2 + ….. nn Upper and Lower Control Limits UCL = u + 3 u/n LCL = u - 3 u/n © Max Zornada (2005) Slide 121 Occurrence Data Occurrence data tends to follow the Poisson Distribution: Average number of events is proportional to the length of the interval; Two events are unlikely to occur together; Events are independent; Mean = Variance If c > 5, approximately Normal © Max Zornada (2005) Slide 122 Sampling © Max Zornada (2005) Slide 123 Types of Sampling Population sampling vs. process sampling; Population sampling: Conducting a customer survey; Quality Control Acceptance sampling; Compiling reasons for calls to call centre Process Sampling: Tracking process performance on a daily, weekly, monthly basis; Recording hourly call volumes to call centre. © Max Zornada (2005) “Snapshots” Ongoing monitoring Slide 124 Sampling Population or Process Population mean = Population std dev = Sample m Statistical Inference Conclusions about the population or process © Max Zornada (2005) Sample Statistics Sample mean = X Sample std dev = s etc., etc., …. Slide 125 Approaches to Sampling Systematic Sampling: Taking data at regular intervals; Random Sampling: Stratified Sampling: Stratify the population into significant subgroups of interest, then use random or systematic sampling. © Max Zornada (2005) Slide 126 Bias Bias is the difference between the data in the sample and the true nature of the population or process; Sampling bias can arise due to: Convenience sampling: sampling the data that is convenient to get, rather than what is needed; Judgement sampling: where sampling is driven by judgements regarding what is important and what is not. © Max Zornada (2005) Slide 127 Daily Processing Quantity Sample Size Selection – Daily Data Minimum Daily Sample Size © Max Zornada (2005) Slide 128 Weekly Processing Quantity Sample Size Selection – Weekly Data Minimum Weekly Sample Size © Max Zornada (2005) Slide 129 Sample Size Formula for populations The general form of the “sample size” equation. Z = the Z-Value corresponding to the 2 2 level of confidence required. Z x 95%: Z= 2 n 2 99.7%: Z = 3 Error Error = +/- accuracy required e.g. if error +/- 10%, error = 10% x2 = variance of the population For categorical data specifically: p = proportion expected n = 36 p (1 - p) Range = 2 X Error (or Total 2 (Range) Absolute Error) © Max Zornada (2005) Slide 130 Tollgate Review: Measure Phase © Max Zornada (2005) Slide 131 Toll Gate Review: Measure Determined what we need to know about the process and where in the process we can get the data. Identified the types of measures we want to collect and have a balance of input, process, output and leading indicators. Developed clear operational definitions of the things we want to measure. Clarified stratification factors we need to identify to facilitate data analysis. Identified what data needs to be collected as new data vs utilising data that is already recorded. Identified appropriate sample sizes, subgroup quantities and sampling frequencies to ensure we get an adequate representation of the process. Used our data to establish the process baselines - stability, capability, sigma level and yield. © Max Zornada (2005) Slide 132 End of Module © Max Zornada (2005) Slide 133 © Max Zornada (2005) Slide 134