Fishbone Diagram Examples: 5 Worked Cause-and-Effect Models

Fishbone diagram examples for manufacturing, software, churn, project delays, and service quality. Step-by-step Ishikawa method with full category breakdowns.

David · Former BCG consultant with 5 years of experienceFebruary 6, 202613 min read

Fishbone diagram examples appear in every root cause analysis playbook, but most versions online stop at labeling six empty branches and calling it a framework. That is a brainstorming template, not an analysis.

A useful fishbone diagram -- also called an Ishikawa or cause-and-effect diagram, originally developed by Kaoru Ishikawa at the University of Tokyo in the 1960s -- maps specific, testable causes across structured categories, then prioritizes which ones to verify with data. The diagram is the hypothesis generation step, not the conclusion.

After applying fishbone diagrams across 60+ operational improvement, quality, and due diligence engagements, we have found that the difference between productive and wasteful root cause sessions comes down to three things: a tightly scoped problem statement, category adaptation for the context, and disciplined verification after brainstorming. This guide walks through five worked fishbone diagram examples across manufacturing, software, customer churn, project delays, and service quality, then covers how to build one step by step and when to use alternatives. For the broader toolkit, see our Strategic Frameworks Guide.

Fishbone diagram example showing six cause categories and root cause analysis

How to Build a Fishbone Diagram: Step-by-Step#

Before diving into examples, here is the six-step process we use in practice. Each example below follows this method.

Step 1: Define the problem statement. Write it as a specific, measurable effect. "Defect rate increased from 1.2% to 3.8% in Q3" works. "Quality is bad" does not. The problem goes at the head of the fish.

Step 2: Select cause categories. For manufacturing, use the 6Ms (Man, Machine, Method, Material, Measurement, Mother Nature). For service or software contexts, adapt to People, Process, Technology, Policy, Environment, and Measurement. Choose 4-6 categories relevant to the problem.

Step 3: Brainstorm causes within each category. Work one category at a time. Ask "what within [category] could cause [problem]?" for each branch. Aim for 3-5 specific causes per category.

Step 4: Add sub-causes. For each primary cause, ask "why does this happen?" to add one level of depth. This connects the fishbone to 5 Whys thinking without going five levels deep on every branch.

Step 5: Prioritize. Mark the 3-5 most likely root causes using dot voting or impact-likelihood scoring. Not every bone deserves investigation -- focus on the causes with both high probability and high impact.

Step 6: Verify with data. Collect evidence for each prioritized cause. Run a controlled test, pull historical data, or conduct interviews. Remove causes that the data does not support. The remaining verified causes become your action plan.

5 Fishbone Diagram Examples Across Industries#

Example 1: Manufacturing Defect (6M Categories)#

Problem statement: Injection molding defect rate increased from 1.2% to 3.8% over 8 weeks at Plant B.

CategoryPrimary CausesSub-Causes
ManNew operator cohort (4 of 12 hired in last 90 days)Training program compressed from 3 weeks to 1 week due to staffing urgency
MachineMold temperature inconsistency on Lines 3 and 7Thermocouple calibration overdue by 6 weeks; PM schedule missed during shift change
MethodQuality gate at Stage 3 removed during process redesignRedesign team assumed automated vision system covered the gap; system only catches surface defects
MaterialResin batch variability from new supplier (switched in Q2)Incoming material inspection reduced from 100% to sample-based after supplier certification
MeasurementSPC charts not reviewed until end of shiftReal-time alerting disabled after false positive complaints in Q1
Mother NatureAmbient humidity spike during July-AugustPlant B HVAC system operates at 60% capacity in summer; dehumidifiers not installed on production floor

Verified root causes: Thermocouple calibration drift on Lines 3 and 7 correlated with 68% of defects. Resin supplier batch variability accounted for 22%. The compressed training program was not independently causal -- experienced operators on the same lines had similar defect rates during drift periods.

Key insight: The team's instinct was to blame new operators. The fishbone redirected investigation toward equipment and materials, where data confirmed the actual causes.

Example 2: Software Bug (Adapted Categories)#

Problem statement: Payment processing failure rate increased from 0.3% to 2.1% after the v4.2 release, affecting 1,400 transactions per day.

CategoryPrimary CausesSub-Causes
CodeRace condition in concurrent payment session handlingNew async handler introduced in v4.2 lacks mutex on shared transaction state
CodeTimeout threshold reduced from 30s to 10s in payment gateway callChange made to improve UX responsiveness; edge cases with slow bank APIs not tested
InfrastructureDatabase connection pool exhaustion during peak hours (2-4 PM)Pool size unchanged since v3.0 when traffic was 40% lower
ProcessIntegration test suite does not cover concurrent payment scenariosTest coverage focused on single-session flows; no load testing in CI/CD pipeline
DataCurrency conversion rounding errors on JPY and KRW transactionsFloating-point arithmetic applied to zero-decimal currencies after library update
Third PartyPayment gateway provider migrated API version without advance noticeDeprecated response field used for idempotency check returns null in new version

Prioritized root causes (verified): The race condition accounted for 74% of failures, confirmed by reproducing under load test. The currency rounding error caused 18% of failures, isolated to JPY/KRW transactions. The third-party API migration caused 8% -- the null idempotency field triggered retry loops that compounded the race condition.

Key insight: Software fishbone diagrams work best with adapted categories: Code, Infrastructure, Process, Data, Third Party, and Environment. The standard 6Ms do not map well to engineering problems. Notice that two causes interacted -- the API migration amplified the race condition -- which a simple 5 Whys chain would have missed.

Example 3: Customer Churn (Service/SaaS Context)#

Problem statement: Monthly churn rate for mid-market accounts increased from 2.8% to 4.6% over two quarters, representing $1.2M in lost ARR.

CategoryPrimary CausesSub-Causes
PeopleCSM-to-account ratio degraded from 1:25 to 1:45 after hiring freezeQBRs dropped from quarterly to semi-annual; at-risk signals missed for 30+ accounts
ProductCore workflow feature (bulk editing) broken for 6 weeks after v8 releaseEngineering prioritized new feature development over bug backlog
ProcessNo structured onboarding for accounts migrated to new pricing tierMigrated accounts received no training on new features included in tier; perceived as price increase with no added value
Pricing22% price increase applied uniformly without segmentationHigh-usage accounts saw reasonable per-unit cost; low-usage accounts saw 40%+ effective increase
CompetitionTwo competitors launched free tiers targeting mid-market in Q1Sales team not equipped with competitive battle cards; no retention offers authorized
MeasurementHealth score model not updated after product changes in v8Usage metrics weighted toward features deprecated in v8; healthy accounts flagged at-risk and vice versa

Prioritized root causes (verified): Exit surveys and win/loss interviews attributed 41% of churned revenue to the pricing increase without segmentation, 28% to the broken bulk editing feature, and 19% to competitive displacement. The CSM ratio degradation was an amplifier -- accounts with regular QBR contact churned at 2.1% even with the price increase, versus 5.9% without.

Key insight: Churn fishbone diagrams require cross-functional categories because churn is never single-cause. The pricing and product causes were both necessary conditions -- fixing either one alone would have reduced churn by approximately 30%, but addressing both together recovered an estimated 65%. This is the kind of multi-cause interaction that decision trees alone cannot capture.

Example 4: Project Delay#

Problem statement: ERP implementation delivered 14 weeks behind schedule, exceeding budget by $2.4M.

CategoryPrimary CausesSub-Causes
PeopleKey solution architect left at week 8; replacement took 5 weeks to onboardNo knowledge transfer documentation; architecture decisions stored in personal notes
ProcessScope changes approved without timeline adjustment (7 CRs in 12 weeks)Change control board met monthly; approvals treated as additive without trade-off analysis
TechnologyLegacy data migration took 3x estimated timeSource system had 14 years of undocumented schema changes; data mapping assumed clean schemas
StakeholdersBusiness unit leads unavailable for UAT during Q4 closeUAT scheduled during busiest financial period; no backup testers identified
VendorIntegration partner delivered custom modules 4 weeks latePartner staffed project with junior developers after proposal stage; no SLA enforcement
Requirements23% of requirements changed or added after design freezeDesign freeze treated as guideline rather than gate; no formal sign-off process

Prioritized root causes (verified): Scope creep without timeline adjustment (7 CRs with no schedule impact assessment) accounted for 6 of the 14 weeks of delay. Legacy data migration complexity added 5 weeks. The architect departure added 3 weeks of rework. Vendor delays overlapped with migration work and did not independently extend the timeline.

Key insight: Project delay fishbone diagrams almost always reveal that no single cause explains the full delay. The causes overlap temporally. The fishbone's value here was separating independent delays (scope creep) from concurrent ones (vendor delays that overlapped with migration) to accurately attribute weeks and avoid double-counting.

Example 5: Service Quality Issue (Healthcare)#

Problem statement: Patient satisfaction scores dropped from 4.2 to 3.4 (out of 5) in the outpatient clinic over one quarter.

CategoryPrimary CausesSub-Causes
People3 of 8 experienced nurses transferred to inpatient unitReplacements are recent graduates with limited outpatient triage experience
ProcessAverage wait time increased from 18 to 42 minutesScheduling system overbooking rate increased from 5% to 15% after policy change to reduce no-shows
TechnologyNew EHR system requires 3x more clicks per patient encounterPhysicians spending 40% of visit time on screen instead of patient interaction
EnvironmentWaiting area reduced by 30% during renovationPatients standing or sitting in hallways; perceived crowding amplifies wait time frustration
PolicyInsurance verification moved from pre-visit to check-inAdds 8-12 minutes to check-in; creates bottleneck at reception
MeasurementSatisfaction surveys moved from in-clinic to post-visit emailResponse rate dropped from 65% to 22%; respondents skew toward dissatisfied patients

Prioritized root causes (verified): Wait time increase correlated most strongly with satisfaction decline (r = -0.82). The overbooking policy and insurance verification change together explained the wait time spike. The EHR system usability issue independently reduced satisfaction -- patients with experienced physicians (who adapted faster to the new system) reported 0.4 points higher satisfaction despite identical wait times.

Key insight: The measurement category revealed a confound: the survey method change inflated the apparent decline. Actual satisfaction dropped from approximately 4.2 to 3.7, not 3.4. Without the fishbone forcing examination of the Measurement category, the team would have overestimated the problem severity and potentially over-invested in remediation.

Free consulting slide templates

SWOT, competitive analysis, KPI dashboards, and more — ready-made PowerPoint templates built to consulting standards.

Fishbone Diagram Examples: Comparing Root Cause Analysis Methods#

Fishbone diagrams are one of several root cause analysis tools. Choosing the right one depends on the problem structure.

MethodBest ForStrengthsLimitations
Fishbone (Ishikawa)Multi-cause problems across categoriesForces exhaustive, structured brainstorming; visual and collaborativeDoes not quantify cause likelihood; requires separate verification step
5 WhysSingle causal chains; isolated incidentsSimple, fast, no training requiredStops at one chain; misses parallel causes; depends on who is asking
Fault Tree AnalysisSafety-critical and engineering failuresQuantitative; models AND/OR cause combinationsRequires probability data; time-intensive for complex systems
Pareto AnalysisPrioritizing among known causesData-driven; identifies the vital fewRequires existing defect data; does not discover new causes

When to combine methods: As ASQ's root cause analysis body of knowledge recommends, start with a fishbone to generate the full cause landscape. Use 5 Whys to drill into each prioritized branch. Apply Pareto analysis to rank verified causes by frequency or impact. This three-step sequence -- brainstorm, drill, prioritize -- is the most effective root cause analysis workflow we have seen across engagements.

For a ready-to-use starting point, see the Root Cause Analysis Template.

Common Fishbone Diagram Mistakes#

Too many branches. Fishbone diagrams with 40+ causes across 8 categories become unreadable and unprioritizable. If you have more than 30 causes, the problem statement is too broad. Split it into sub-problems and run separate fishbone sessions for each.

Vague causes. "Communication issues" and "lack of training" are symptoms, not root causes. Push for specificity: "New hires receive no documentation on the legacy data schema" is actionable. "Training is insufficient" is not. Every cause should answer: who, what, when, and how much.

Skipping verification. The fishbone produces hypotheses, not conclusions. We have seen teams implement solutions for all 15 brainstormed causes without testing any of them. In every example above, data verification eliminated 50-70% of initial hypotheses. Budget your time as 30% brainstorming, 70% verification.

Using the wrong categories. The 6Ms work for manufacturing. They do not work for software, service, or strategy problems. Adapt categories to your context. A fishbone diagram for customer churn needs People, Product, Process, Pricing, Competition, and Measurement -- not Man, Machine, Material.

Stopping at the first level. A fishbone with only primary causes (one level of branches) rarely reaches root causes. Add at least one sub-cause level by asking "why does this cause exist?" for each primary branch. The sub-cause is usually where the actionable fix lives.

Key Takeaways#

Fishbone diagrams are hypothesis generation tools, not answer generators. Their value is in forcing structured, exhaustive thinking across categories before the team converges on a solution -- which prevents the common failure of fixing the most obvious cause while the actual root cause persists.

  • Scope the problem tightly -- a measurable problem statement prevents the diagram from becoming an everything-is-wrong brainstorm
  • Adapt categories to context -- use 6Ms for manufacturing, adapted categories (People, Process, Technology, Data, Policy) for service and software
  • Prioritize before acting -- mark the 3-5 most likely causes and verify with data before implementing fixes
  • Combine with other methods -- use fishbone for breadth, 5 Whys for depth, and Pareto for prioritization
  • Keep it readable -- 4-6 categories, 3-5 causes per category, one level of sub-causes

For related frameworks, see our SWOT Analysis Examples and Decision Tree Examples. For the full consulting frameworks toolkit, explore the Strategic Frameworks Guide.

Build consulting slides in seconds

Describe what you need. AI generates structured, polished slides — charts and visuals included.

Try Free
Fishbone Diagram Examples: 5 Worked Cause-and-Effect Models | Deckary