Lean Sigma is different to many traditional process improvement initiatives in its reliance on data to make decisions. During the define stage of lean sigma you identify the major metrics, otherwise known as the Key Process Output Variables (KPOVs) or Big Ys, to measure performance of the process.
Thenceforth for the improvement you use the equation Y = f(X1,X2,..., Xn) to identify and narrow down the Xs that drive the Y. Improvement is completely data driven.
Prior to undertaking that road, you need to have a basic understanding of data, measurements and how to deal with them.
This Data Analysis Video teaches you the basic tools for understanding, summarizing, and making future predictions with your collected data. Includes MS Excel templates.
Continuous versus Attribute
Results versus Predictors
Efficient versus Effective
The type of data is important to process improvement. Types of data affects how measurements are defined, how to go about collecting data, and which tools to use in analyzing the data. An example is shown below
Selecting the statistical tool based on the data type.
Continuous measures come from an infinitely divisible scale, such as weight, temperature, height, time, speed, and decibels. They are often called Variable measures.
One prefers continuous data over Attribute type data because it allows the use of more powerful statistical tests. Continuous data facilitates decisions being made with much less data (a sample size of 30 works well). Continuous data allows measurement of performance (i.e., how good is it?) versus conformance (i.e., is it just good or bad?).
If the data used in a process improvement project is attribute type, then we recommend identifying a related continuous type measure and use that instead.
Learn SPC in an hour. Train your employees. Improve your processes and products. Prevent defects and save your company money.
Non continuous measurements are known as Attribute or Discrete measures. Despite the lower statistical power, they are generally easier and faster to use than continuous data. Attribute data are effective for measuring intangible factors such as customer satisfaction or perception.
There are several different types of Attribute data along a spectrum as follows:
As the number of categories increase in Attribute data, it becomes more like continuous data.
The key failings of Attribute data is the need for a large number of observations to get valid information from any tests performed. Also attribute data can hide important discrimination.
The Lean Sigma equation Y = f(X1,X2,..., Xn) shows the relationship between the Result (the Y) and the Predictors (the Xs). Results are often called "output" factors or KPOVs. Results might be short-term (on-time delivery) or longer-term (Customer satisfaction), but they are typically lagging indicators.
Since multiple Xs drive the results in the process improvement, results are reactive rather than proactive variables.
Results tend to be much easier to collect than Predictors. Analysis of a result is typically done against a specification and becomes a Process Capability measure.
8D Manager Software with 8D, 9D, 5Y and 4M report generator. Your corrective action software for managing, measuring, and reporting issues.
Predictors are the upstream factors that, if measured, can forecast events downstream in the process. For example, an increase in raw-material order lead times might predict an increase in late deliveries. Predictors are the Xs in the equation Y = f(X1,X2,..., Xn). We call these input factors or KPIVs.
Predictors are the leading indicators of the process. Generally are more difficult to identify and collect. When starting a process improvement project, it is unlikely that this data is available historically. The Team must set up a measurement system to collect KPIVs.
Due to the proactive nature of Predictors, they form the basis of the strongest process improvement measures.
We use two categories of metrics to describe the Big Ys in the process, namely effectiveness and efficiency. Effectiveness externally measures the process related to the voice of the customer (VOC). An effective process produces the desired results from the Customer's perspective.
To understand effectiveness, ask questions such as...
Effectiveness measures include
TrainingKeeper Software. Keep, organize and plan all your employees' training and activities. Software includes multi-user support with reports, certs, and calendars.
Efficiency internally measures the process. Efficiency relates to the level of resources used in the process to achieve the desired result.
To understand process improvement efficiency, ask these business internal questions...
Efficiency metrics include
During process improvement each process should have at least two efficiency requirements established:
Your comprehensive package for learning and training the TRIZ problem solving methodology. Includes mobile app, books, PowerPoints and real world examples.
After identifying a metric, the work isn't complete. To be a useful metric, you must create a clear, precise description of the metric. The description allows everyone to evaluate and count the same way. There is a common understanding of and agreement on the results.
In transactional processes in particular, Operational Definitions are often the only real way to maintain control. If poorly defined they are commonly the biggest cause of measurement system issues.
To describe a measure fully, create an Operational Definition that includes
Follow this roadmap to identifying, defining, and capturing data during your process improvement:
Identify the process improvement measures that translate into meaningful data. Determine if each measure is a KPOV or KPIV, if it is a measure of effectiveness or efficiency, and if it is continuous or attribute.
Create a clear Operational definition of the characteristic to be measured (who, what, when, where, how, and why).
Comprehensive ISO 9001:2015 Audit Checklist for Internal, Gap and Certification Audits.
Test the Operational Definition with process improvement the Stakeholders (all those that affect and are affected by the process improvement) to ensure consistency in understanding. Revise the operational definition if necessary. Any metric chosen should be both repeatable and reproducible
Lean Sigma projects normally actively capture data so the team has to create a data collection and sampling plan. Historical data can be extremely useful when available; however, it is typically not based on the same operational definitions. Historical data may be hard to use (not stratified the right way, not sortable, and so on,) and generally incomplete.
Before collecting the data, take time to consider how the team will analyze the data after it is collected. Determine the Xs with respect to their effects on the Ys, and so on. To facilitate this, build the Xs need into the data collection plan. KPIV may include things such as time, date, person, shift, operation number, Customer, buyer, machine, subassembly #, defect type, defect location, component and defect impact and criticality, and so on. The Xs should come from the Cause & Effect Matrix.
Be aware though that the number of sub-categories has a big impact on the amount of data required from which to make statistical inferences.
Your easy to edit ISO 9001:2015 Quality Assurance Manual.
Create a sampling plan. Data collected is only a sample from the process; it never represents every entity to be processed (known as the population). Even if you collect 100% of the data points for the process improvement during a certain period of time, it is still only a sample of the whole population. In fact, this issue drives the use of statistics in the roadmap.
We use statistics because only a sample taken from the population. The statistical properties of the sample draws inferences (predictions, guesses) as to the properties of the population. Statistics make it faster, less costly way to gain insight into a process or large population.
The trick is to have the best sample from which to make inferences. Thus the sample needs to be a miniature version of the population, only smaller. To achieve this consider these two elements when sampling : sample quality and sample size.
Sample quality measures how well the sample represents the population and meets the needs of the sampler. Lean Sigma statistical methods generally assume random sampling in which every entity or member of the full population has an equal chance of being selected in the sample. See below figure for different sampling strategies.
When sampling the team must strive to minimize the bias in the sampling procedure. Bias is the difference between the nature of the data in the sample and the true nature of the entire population. Avoid convenience sampling. Don't sample because the item is easy to get to. Difficult customers are hard to capture data from but represent a valuable source of information. Minimize data errors and missing data.
Snap Sampling Plans! software combines all the key industry standard attribute sampling plans into one place.
With over 10,000 sampling possibilities, Snap Sampling Plans! guides you in selecting the correct AQL sampling plan.
Sample Size is important because, generally, precision in sampling results increases as the sample size n increases. Unfortunately, this is not a one-to-one relationship. In fact, mathematically precision's increases in proportion to so doubling sample size doesn't double accuracy.
For Continuous data meaningful Sample Sizes are
For Attribute Data meaningful Sample Sizes are
Based on the sampling plan, create a data collection form and tracking system. This comprises the who, what, where, when, and how the data is captured. See the below Check sheet for an example.
Create procedures for completing the data collection forms. Any instructions should be visual examples and understandable by all. It is useful to include pictures of the form, describe what goes in each box and include examples of the completed forms. Choose the data collectors carefully and train them on the procedures using the instructions to ensure consistent data collection.
This Data Analysis Video teaches you the basic tools for understanding, summarizing, and making future predictions with your collected data. Includes MS Excel templates.
Test the data collection method. Do a short dry run of the data capture (a few data points) to identify problems and make adjustments. After completion, assess the accuracy, repeatability, and reproducibility of the data collection system
Collect the data as per the data collection procedure. The team must follow the sampling plan consistently.
Record any changes in operating conditions not part of the normal or initial operating conditions. Any events out of the ordinary should be immediately written into a logbook.
Audit the data collection procedures and assure they're followed at all times. When the desired sample size is reached, stop the data collection.
Enter the collected data promptly into a database, such as a spreadsheet or statistical software. Make a backup copy of the electronic file. Keep all the paper copies to be archived as part of the project report.
Quality Assurance Solutions Robert Broughton (805) 419-3344 USA |
Software, Videos, Manuals, On-Line Certifications | ||
An Organizational Task Management System. Projects, Meetings, Audits & more | ||
Corrective Action Software | ||
Plan and Track Training | ||
AQL Inspection Software |
450+ Editable Slides with support links | ||
Learn and Train TRIZ | ||
Editable Template | ||
Templates, Guides, QA Manual, Audit Checklists | ||
EMS Manual, Procedures, Forms, Examples, Audits, Videos | ||
On-Line Accredited Certifications Six Sigma, Risk Management, SCRUM | ||
Software, Videos, Manuals, On-Line Certifications |