Assignment Statistics Ankit Agarwal May 2024
Assignment Statistics Ankit Agarwal May 2024
Following the DMAIC methodology, the team completed the Define process by analyzing the business
process map, SIPOC and voice of customers as part of the Define phase as part of first exercise (Day 1).
This information formed the background to define the relevant CTQs. Since the exercise is performed by
Safety Engineers, CTQs that have direct safety impacts are rated higher.
After the cause and effect matrix is performed together with the team, the resulting X’s are plotted on a
simple graph based on the priority number (high to low). Several X’s are found to be close to each other.
This indicated that there are multiple contributing factors in the CTQ. These X's range from human error
(inconsistent sampling, analysis, and data recording) to equipment issues (calibration, breakdowns,
outdated) to personnel training/knowledge requirements.
A. Process Analysis: This approach involves directly examining the process itself to identify potential
bottlenecks and opportunities for improvement. Key tools and techniques under this approach
include:
o Process Mapping: Visualizing the steps in a process using flowcharts can help identify areas
of inefficiency, and delay.
o Value Stream Mapping: This technique (referred to as Makigami) highlights the value-adding
steps within a process and identifies areas of "waste" (non-value-adding activities). Reducing
such waste can minimize errors and improve overall process flow. The rationale behind this
is that complex processes create more opportunities for mistakes. Streamlining the process
simplifies tasks and reduces error risk.
o Focus on Streamlining: The goal of process analysis is to identify and eliminate redundant and
risk-inducing steps, ultimately leading to a more efficient and reliable process.
Process Optimization Through Redundancy Review: This section proposes a review of a specific process
step to identify potential areas for streamlining. The process involves document checks by both security
guard and laboratory. This step was performed during the lecture.
Current Process Analysis:
1. Input (I): Security Guard receives CMR (transport document) and WIP (waste identity passport).
2. Output (O): Security Guard verifies documents and grants entry or initiates a "non-conformity
chain" if discrepancies are found.
3. Input (I): Laboratory receives the same documents (CMR and WIP).
4. Output (O): Laboratory verifies documents and initiates a "non-conformity chain" for
discrepancies.
Proposed Optimization:
• Security Guard: Focus on core security functions.
o Rationale: Security guards lack training in chemical components, making WIP verification
less reliable.
• Laboratory: Maintain document verification due to expertise.
o Rationale: Laboratory personnel are trained to understand both documents and verify.
Eliminating Customer Reaction Check:
Safety considerations don't require knowledge of customer response to discrepancies. Proposal is to
exclude this information from document checks.
B. Data-Driven Analysis: This section outlines a structured approach to process analysis using data. The
approach, known as "Data-Driven Analysis or Data door," emphasizes a sequential execution of three
key stages (if necessary) for rigorous and actionable insights. It's important to follow these stages in
order, as each builds upon the findings of the previous one. If a stage provides a clear picture of the
root cause, it might be unnecessary to proceed further.
Stage 1: Visual Exploration The first stage focuses on visualizing data to identify patterns, trends, and
potential issues. This involves using various graphical tools, such as:
• Histograms: These reveal the distribution of data, highlighting skewness, outliers, or unexpected
patterns.
• Time Series Plots: These charts show data points over time, helping detect trends, shifts, or
cyclical patterns.
• Pareto Charts: These highlight the most significant factors contributing to an issue, following the
80/20 rule (80% of the effect comes from 20% of the causes).
• Control Charts: These monitor process stability and identify deviations from desired control limits.
• Box Plots: These provide a quick overview of data distribution by summarizing quartiles and
identifying outliers.
These techniques give a comprehensive picture of the data and potential areas for improvement.
Depending on the insights gained here, it may or may not be necessary to proceed to further stages.
Stage 2: Statistical Inferences This stage utilizes statistical methods to draw more robust conclusions
about the data. Here are some key tools used:
• Confidence Intervals: These estimate the range within which a population parameter is likely to
lie, based on a sample.
• Hypothesis Testing: This involves testing assumptions about a population parameter (e.g.,
average) to determine if there's evidence to support a specific claim.
• Correlation and Regression Analysis: These assess the strength and direction of relationships
between variables. Regression analysis can even predict the impact of one variable on another.
• Normality Testing: This verifies if a data set follows a normal distribution, which is a prerequisite
for many statistical tests.
In essence, the Data Door and Process Door work together during DMAIC analyses:
• The Process Door helps define the "what" and "how" of the process being analyzed.
• The Data Door helps analyze the "why" behind process performance and identify areas for
improvement.
a. Prepare: Some part of the data is collected using automated system and part of it is manually filled
in. The data was checked for logic and completeness. Some findings are:
i. Truck residence time as significant number of trucks have a low residence time (<5mins). This
seems not reasonable when we consider the practicalities.
ii. Many trucks have similar residence time e.g. 0:00:13; 0:05:43; 0:02:50 etc. which points to some
manipulation on the residence time.
iii. All country codes are in order “Cxx” but one of them is P-L
iv. The manual part of data is filled but it is hard to identify typing errors
b. Grouping: To get more insight on the data table, following groups are added:
i. Day of the week
ii. Residence Time
iii. Country of origin (haulier)
iv. Standard/non-standard of truck
v. Waste type
vi. Kind of trucks
Next, excel based pivot tables are used to calculate the 5 CTQs. The whole exercise is performed as a
group exercise (participated by Femke, James and Ankit)
While faster sampling and unloading times are desirable from an efficiency standpoint, safety
considerations require a balanced approach. Using the in and out data of trucks, a simple residence time
(in minutes) is calculated and plotted on a histogram to understand the underlying trend.
The histogram shows a significant number of trucks (~350) with a residence time of less than 11 minutes.
Majority of them are showing conformity which means they go to unloading area, get a conformity check
from production and unload and then leave. To have all of it in less than 11 minutes does raise question
on the integrity of the data recording.
This histogram does not show a normal distribution behavior indicating that multiple factors are playing
a role in the outcome. A normality check with a QQ plot is performed and the data in residence time is
not normally distributed as it is showing heavy tailed distribution at both ends.
Graphical Analysis (1st step for Data door): The goal is to identify potential root causes of these issues by
applying graphical analysis.
Pareto Analysis: Given the discrete nature of the data (categorized as conforming or non-conforming),
Pareto charts were determined to be a suitable tool for further analysis. Following the Pareto Principle
(80/20 rule), Pareto charts can help identify the most frequent non-conformance (NC) categories, which
are likely to represent the root causes for most of the issues.
While most subsets yielded limited insights, the country analysis revealed a potential area for further
investigation. The Pareto chart for this subset highlighted Country C03 and Country C99 as significant
contributors to NCs, together accounting for a substantial portion of the total.
Drilling Down by Country:
Conclusion:
The analysis clearly shows that specific customers are the primary contributors to NCs. Further statistical
analysis like hypothesis testing or regression might not be necessary at this stage. Instead, the priority
should be to develop targeted actions to address the identified root causes and improve compliance.
Process Improvement Strategy: This section outlines a multi-pronged approach to address the identified
issues and improve overall waste management compliance.
1. Customer engagement, auditing and follow it up with corrective action plan.
2. Collaborative Learning via training, workshops, and role play.
3. Continuous monitoring and real time feedback.
4. Establish six sigma benchmark and define measurable KPI: to elaborate on this → Set specific,
measurable targets for reducing NC rates. For example, the goal might be to achieve less than 5% NCs for
high-volume customers and less than 10% for all others within a defined timeframe (e.g., next quarter).
5. Streamlining documentation and integrity in data collection.
Enhancing fishbone and C&E matrix:
The previous fishbone and Cause-and-Effect (C&E) matrix analysis focused heavily on internal factors
contributing to NCs, such as inadequate training, maintenance issues, and detection by guards, labs, and
operators. While valuable, this approach primarily examined internal company operations and neglected
to consider potential inefficiencies within the company itself, such as redundant guard tasks or the lack
of source tracing for NCs. Addressing the root cause of an NC directly translates to a more efficient
downstream process. Also, the data analysis provided insights on external e.g. certain countries and
customers.
Going forward, it's crucial to incorporate the customer as a potential "X" factor in future fishbone
diagrams. By doing so, the customer is more likely to be identified as a high-ranking factor in the C&E
matrix, enabling quicker and more targeted interventions to improve the overall process. This reinforces
the importance of a comprehensive understanding of the entire process and a willingness to think
creatively when constructing an Ishikawa diagram. This approach ultimately increases the chances of
pinpointing root causes during the C&E matrix or Failure Mode and Effects Analysis (FMEA) stages.