

The Occupational Safety Leadership Podcast
Dr. Ayers/Applied Safety and Environmental Management
Interviews along with a Q&A format answering questions about safety. Together we‘ll help answer not just safety compliance but the strategy and tactics to implement injury elimination/severity.
Episodes
Mentioned books

Oct 11, 2023 • 9min
Episode 94 - 5 x 5 Risk Assessment Matrix
Dr. Ayers breaks down the 5×5 Risk Assessment Matrix—a tool that helps leaders evaluate hazards by scoring severity and likelihood on a 1–5 scale. The episode focuses on how to use the matrix correctly, avoid common misapplications, and turn it into a practical decision‑making tool rather than a paperwork exercise.
Key Concepts
1. The Structure of the 5×5 Matrix
The matrix evaluates risk using two dimensions:
Severity (1–5)
1 – Insignificant: No injury or very minor first aid
2 – Minor: Minor injury, short-term discomfort
3 – Moderate: Recordable injury, medical treatment
4 – Major: Serious injury, lost time, hospitalization
5 – Catastrophic: Fatality or life‑altering injury
Likelihood (1–5)
1 – Rare: Highly unlikely
2 – Unlikely: Could happen but not expected
3 – Possible: Happens occasionally
4 – Likely: Happens regularly
5 – Almost Certain: Expected to occur
Risk Score = Severity × Likelihood This produces a range from 1 to 25, which is then categorized (e.g., low, medium, high, critical).
2. The Purpose of the Matrix
Dr. Ayers emphasizes that the matrix is not about creating a perfect numerical score. Its real value is:
Driving conversations about hazards
Prioritizing controls
Documenting risk reduction
Supporting leadership decisions
It’s a thinking tool, not a compliance checkbox.
3. Common Misuses
The episode calls out several pitfalls:
Treating the numbers as precise measurements (They’re estimates, not scientific calculations.)
Using the matrix to justify inaction (“It’s only a 6, so we don’t need to fix it.”)
Failing to reassess after controls (Risk scoring must reflect improvements.)
Ignoring exposure frequency (Likelihood must consider how often workers interact with the hazard.)
4. How to Use the Matrix Effectively
Dr. Ayers offers practical guidance:
A. Score hazards as a team
Different perspectives reduce bias.
B. Focus on credible worst-case severity
Not the most likely outcome—the worst plausible one.
C. Document your reasoning
Why you chose a severity or likelihood score matters more than the number itself.
D. Re-score after controls
This shows whether your interventions actually reduced risk.
E. Use the matrix to prioritize
High‑severity hazards with moderate likelihood often deserve more attention than low‑severity hazards with high likelihood.
5. Leadership Takeaways
The episode reinforces that strong safety leaders:
Use the matrix to guide action, not justify inaction
Encourage open discussion about hazards
Treat risk scoring as a dynamic process
Focus on severity reduction through engineering and administrative controls
Use the matrix to communicate risk clearly to frontline teams and executives
6. Practical Example (from the episode’s style)
A rotating shaft without guarding:
Severity: 5 (catastrophic)
Likelihood: 3 (possible)
Risk Score: 15 (high)
After installing a guard:
Severity: 5 (unchanged—still catastrophic if bypassed)
Likelihood: 1 (rare)
New Score: 5 (low)
This illustrates why controls reduce likelihood, not severity, and why rescoring matters

Oct 10, 2023 • 9min
Episode 93 - 4 x 4 Risk Assessment Matrix
Dr. Ayers explains the 4×4 Risk Assessment Matrix, a simplified version of the more common 5×5 tool. The episode focuses on how reducing the scoring options can actually improve consistency, reduce over‑precision, and make risk conversations more meaningful.
1. Structure of the 4×4 Matrix
The matrix evaluates hazards using Severity and Likelihood, each scored from 1 to 4.
Severity (1–4)
1 – Minor: First aid only
2 – Moderate: Recordable injury
3 – Serious: Lost time or significant medical treatment
4 – Severe/Catastrophic: Permanent disability or fatality
Likelihood (1–4)
1 – Rare: Unlikely to occur
2 – Possible: Could happen occasionally
3 – Likely: Happens regularly
4 – Almost Certain: Expected to occur
Risk Score = Severity × Likelihood Range: 1 to 16, typically grouped into low, medium, high, and critical.
2. Why Use a 4×4 Instead of a 5×5?
Dr. Ayers highlights several advantages:
Less false precision Fewer scoring options reduce the illusion that risk scoring is scientific.
More consistent scoring Teams tend to agree more often when there are fewer choices.
Faster assessments Useful for dynamic or field‑level risk evaluations.
Better focus on discussion The conversation becomes more important than the number.
3. Common Pitfalls
Even with a simpler matrix, leaders can misuse it:
Treating the score as absolute truth It’s still an estimate, not a measurement.
Failing to consider exposure frequency Likelihood must reflect how often workers interact with the hazard.
Not rescoring after controls Controls should reduce likelihood, not severity.
Using the matrix to justify inaction “It’s only an 8, so we’re fine” is not leadership.
4. How to Use the 4×4 Matrix Effectively
A. Score hazards as a group
Reduces bias and improves accuracy.
B. Use credible worst‑case severity
Not the most likely outcome—the worst plausible one.
C. Document the rationale
Why you chose a score matters more than the number.
D. Reassess after controls
Shows whether risk was actually reduced.
E. Prioritize severity first
High‑severity hazards deserve attention even if likelihood is low.
5. Leadership Takeaways
Strong safety leaders:
Use the matrix to drive action, not avoid it
Encourage open hazard discussions
Treat risk scoring as dynamic
Focus on engineering and administrative controls
Communicate risk clearly to frontline teams and executives
6. Example (in the spirit of the episode)
Unprotected elevated work platform:
Severity: 4 (severe)
Likelihood: 2 (possible)
Risk Score: 8 (medium/high depending on scale)
After installing guardrails and requiring fall protection:
Severity: 4 (unchanged)
Likelihood: 1 (rare)
New Score: 4 (low)
This reinforces the principle: controls reduce likelihood, not severity.

Oct 9, 2023 • 10min
Episode 92 - 3 x 3 Risk Assessment Matrix
Dr. Ayers introduces the 3×3 Risk Assessment Matrix, the simplest of the common matrix formats. The episode emphasizes that reducing the scoring options forces teams to focus on meaningful discussion, credible severity, and practical controls, rather than getting lost in numerical precision.
The 3×3 matrix is ideal for quick field-level assessments, dynamic work environments, and frontline decision-making.
1. Structure of the 3×3 Matrix
The matrix evaluates hazards using Severity and Likelihood, each scored from 1 to 3.
Severity (1–3)
1 – Minor: First aid or negligible harm
2 – Moderate: Recordable injury or medical treatment
3 – Severe: Permanent disability or fatality
Likelihood (1–3)
1 – Unlikely: Not expected to occur
2 – Possible: Could occur under the right conditions
3 – Likely: Expected to occur or occurs regularly
Risk Score = Severity × Likelihood Range: 1 to 9, typically grouped into low, medium, and high.
2. Why Use a 3×3 Matrix?
Dr. Ayers highlights several advantages of the simplified format:
Reduces overthinking Fewer choices mean faster, more consistent scoring.
Ideal for dynamic risk assessments Great for pre‑task briefings, JHAs, and field-level hazard checks.
Minimizes false precision You can’t pretend the difference between a “2 vs. 3 likelihood” is scientific.
Improves team agreement Workers tend to align more easily when the scale is simple.
Keeps the focus on controls The conversation becomes: “What can we do about this hazard right now?”
3. Common Pitfalls
Even with a simple matrix, leaders can misuse it:
Treating the score as a justification to proceed A “3” doesn’t mean the hazard is acceptable.
Ignoring credible worst-case severity Severity must reflect what could happen, not what usually happens.
Not considering exposure frequency Likelihood must reflect how often workers interact with the hazard.
Failing to reassess after controls Controls should reduce likelihood, and the matrix should show that.
4. How to Use the 3×3 Matrix Effectively
A. Use it for quick, real-time decisions
Perfect for crews starting a task or adjusting to changing conditions.
B. Score hazards as a group
Frontline workers often see risks leaders miss.
C. Document the reasoning
Even a simple matrix needs context behind the numbers.
D. Re-score after controls
Shows whether risk was actually reduced.
E. Prioritize severity
A severity of 3 always deserves attention, even if likelihood is low.
5. Leadership Takeaways
Strong safety leaders:
Use the matrix to drive action, not to justify continuing work
Encourage open hazard conversations
Treat risk scoring as dynamic and situational
Focus on engineering and administrative controls
Use the matrix as a communication tool, not a compliance form
6. Example (in the spirit of the episode)
Working near a pinch point on a conveyor:
Severity: 3 (severe)
Likelihood: 2 (possible)
Risk Score: 6 (medium/high depending on scale)
After installing a guard and adding a lockout procedure:
Severity: 3 (unchanged)
Likelihood: 1 (unlikely)
New Score: 3 (low)
Again reinforcing the principle: controls reduce likelihood, not severity.

Oct 2, 2023 • 16min
Episode 91 - Matthew Herron of the Southwest Research Insitute
In today's episode, we catch up with Matthew Herron of Southwest Research Institute. Matt is a titan in the field of safety. Today's episode focuses on ergonomics and importance of early reporting.

Sep 20, 2023 • 5min
Episode 90 - Safety Equipment Maintenance Rate
Dr. Ayers explains the concept of the Safety Equipment Maintenance Rate, a metric that helps organizations understand how reliably they are maintaining the equipment that protects workers. The episode emphasizes that safety equipment is only effective if it is functional, inspected, and maintained at a predictable rate—and that many organizations dramatically overestimate how well they are doing.
The Maintenance Rate becomes a leading indicator of system health, not just a compliance statistic.
1. What the Maintenance Rate Measures
The Safety Equipment Maintenance Rate tracks:
How often safety‑critical equipment is inspected
How often it is maintained on schedule
How often it is found in proper working condition
How quickly deficiencies are corrected
Examples of equipment included:
Fall protection gear
Fire extinguishers
Emergency eyewash stations
Machine guards
Ventilation systems
Gas detectors
Lockout/tagout devices
If workers rely on it to prevent injury, it belongs in the metric.
2. Why the Maintenance Rate Matters
Dr. Ayers highlights several reasons this metric is essential:
A. Safety equipment fails silently
Most safety equipment doesn’t show obvious signs of failure until it’s needed—and by then it’s too late.
B. It reveals system weaknesses
Low maintenance rates often point to:
Poor scheduling
Lack of ownership
Inadequate staffing
Weak preventive maintenance programs
Overreliance on reactive repairs
C. It’s a true leading indicator
Unlike injury rates, maintenance rates show future risk, not past outcomes.
D. It builds trust with workers
When workers see broken guards, expired extinguishers, or damaged PPE, they lose confidence in the safety system.
3. How to Calculate the Maintenance Rate
While organizations may tailor the formula, the episode frames it as:
Maintenance Rate = (Number of items maintained on schedule ÷ Total number of items requiring maintenance) × 100
A high rate means:
Inspections are happening
Repairs are timely
Equipment is ready when needed
A low rate means the system is quietly degrading.
4. Common Pitfalls
Dr. Ayers calls out several recurring issues:
Counting inspections but not repairs A checked box doesn’t mean the equipment works.
Ignoring overdue items “We’ll get to it next month” is a system failure.
No clear ownership If everyone owns it, no one owns it.
Not tracking repeat failures Chronic issues signal deeper design or usage problems.
Assuming equipment is fine because it “looks fine” Many failures are internal or hidden.
5. How to Improve the Maintenance Rate
A. Assign clear ownership
Every safety‑critical asset needs a responsible person or team.
B. Use a preventive maintenance schedule
Don’t rely on memory or ad‑hoc checks.
C. Track deficiencies and close‑out times
Speed matters—slow repairs increase exposure.
D. Prioritize high‑risk equipment
Focus on items that protect against severe hazards.
E. Audit the system regularly
Spot‑check equipment to verify the numbers match reality.
6. Leadership Takeaways
Strong safety leaders:
Treat maintenance as a core safety function, not a support task
Use the Maintenance Rate as a leading indicator
Ensure equipment is functional, not just present
Build systems that prevent silent failures
Reinforce that safety equipment is only as good as its maintenance
7. Practical Example (in the spirit of the episode)
A facility has 200 pieces of safety‑critical equipment. During the month:
170 were inspected and maintained on schedule
30 were overdue or missed
Maintenance Rate = 170 ÷ 200 = 85%
If the organization’s target is 95%, this signals a gap that could expose workers to hidden risks.

Sep 19, 2023 • 10min
Episode 89 - Safety Training Completion Rate
Dr. Ayers explains the Safety Training Completion Rate, a leading indicator that measures how reliably an organization ensures workers receive the training they need before they perform hazardous tasks. The episode emphasizes that training is only effective when it is completed on time, tracked accurately, and aligned with real job demands—not when it’s treated as a paperwork exercise.
1. What the Training Completion Rate Measures
The metric evaluates:
Whether required training is completed on schedule
Whether workers are current on refresher requirements
Whether new hires receive training before exposure
Whether training is task‑specific, not generic
Whether the organization can prove completion, not just assume it
Training categories typically included:
OSHA‑required courses
Equipment‑specific training (forklifts, aerial lifts, cranes)
Hazard‑specific training (LOTO, confined space, fall protection)
Annual or periodic refreshers
Site‑specific orientation
If a worker needs it to perform a task safely, it belongs in the metric.
2. Why the Training Completion Rate Matters
A. It predicts future incidents
Workers without proper training are more likely to make errors, misuse equipment, or misunderstand hazards.
B. It exposes system weaknesses
Low completion rates often reveal:
Poor onboarding processes
Inconsistent supervisor follow‑through
Scheduling bottlenecks
Outdated training records
Overreliance on “tribal knowledge”
C. It builds or erodes trust
Workers notice when training is rushed, skipped, or treated as a formality.
D. It’s a true leading indicator
It measures readiness, not outcomes.
3. How the Training Completion Rate Is Calculated
A common formula:
Training Completion Rate = (Number of workers current on required training ÷ Total workers who require the training) × 100
High rate → workforce is prepared Low rate → workers are exposed to preventable risk
4. Common Pitfalls
Dr. Ayers highlights several recurring issues:
Counting scheduled training as completed “They’re signed up” is not the same as “they’re trained.”
Allowing workers to perform tasks before training A major system failure.
Inaccurate or outdated records Many organizations discover their LMS data is wrong.
One‑size‑fits‑all training Generic training doesn’t prepare workers for specific hazards.
No accountability for overdue training If no one owns it, it doesn’t get done.
5. How to Improve the Training Completion Rate
A. Assign clear ownership
Supervisors must ensure workers are trained before exposure.
B. Use a reliable tracking system
LMS or spreadsheet—accuracy matters more than complexity.
C. Prioritize high‑risk tasks
Training for hazardous work must be completed first.
D. Integrate training into onboarding
New hires should not touch equipment until trained.
E. Audit training records regularly
Spot‑check to ensure the data matches reality.
6. Leadership Takeaways
Strong safety leaders:
Treat training as a risk‑control measure, not a compliance checkbox
Use the Completion Rate as a leading indicator
Ensure workers are trained before they face hazards
Hold supervisors accountable for training readiness
Align training with real work, not generic modules
7. Practical Example (in the spirit of the episode)
A facility has 120 workers who must complete annual fall‑protection training. Currently:
102 are current
18 are overdue
Training Completion Rate = 102 ÷ 120 = 85%
If the organization’s target is 95%, the gap signals a readiness problem and potential exposure.

Sep 18, 2023 • 12min
Episode 88 - Hazard Identification and Resolution Rate
Dr. Ayers introduces the Hazard Identification and Resolution Rate, a powerful leading indicator that measures how effectively an organization finds hazards and—more importantly—fixes them. The episode stresses that identifying hazards is only half the job; the real value comes from closing them out quickly and reliably.
This metric reveals the health of a safety culture far more accurately than injury rates.
1. What the Metric Measures
The Hazard Identification and Resolution Rate tracks:
A. Hazard Identification
How many hazards workers and leaders are finding
Whether hazards are being reported consistently
Whether reporting is encouraged or discouraged
Whether the organization is generating enough “eyes on risk”
B. Hazard Resolution
How many identified hazards are actually corrected
How quickly they are resolved
Whether fixes are temporary or permanent
Whether high‑risk hazards are prioritized
The metric captures both volume and follow‑through.
2. Why This Metric Matters
A. It predicts future incidents
Unresolved hazards are direct precursors to injuries.
B. It reveals cultural health
High identification + high resolution = strong safety culture Low identification + low resolution = fear, apathy, or disengagement
C. It exposes system weaknesses
Low resolution rates often point to:
Poor maintenance support
Lack of ownership
Slow approval processes
Understaffed teams
Leaders who don’t follow up
D. It builds trust
When workers see hazards fixed quickly, they believe leadership cares.
3. How the Rate Is Calculated
Organizations may tailor the formula, but the episode frames it as two related metrics:
Hazard Identification Rate
Number of hazards identified ÷ Number of workers (or hours worked)
Hazard Resolution Rate
Number of hazards resolved ÷ Number of hazards identified
High identification + high resolution = a healthy, proactive system.
4. Common Pitfalls
Dr. Ayers highlights several traps:
Focusing only on identification Finding hazards without fixing them creates frustration.
Focusing only on resolution Fixing a few hazards looks good on paper but hides under‑reporting.
Punishing workers for reporting hazards This kills the identification rate instantly.
Treating all hazards equally High‑severity hazards must be resolved first.
Using temporary fixes as “resolution” Tape and zip‑ties don’t count.
5. How to Improve the Metric
A. Encourage reporting
Reward workers for identifying hazards, not for staying quiet.
B. Assign ownership
Every hazard needs a responsible person and a due date.
C. Prioritize by risk
Fix high‑severity hazards first.
D. Track close‑out times
Speed matters—slow fixes increase exposure.
E. Audit the system
Verify that “resolved” hazards are actually resolved.
6. Leadership Takeaways
Strong safety leaders:
Treat hazard identification as a positive behavior
Ensure hazards are fixed quickly, not just logged
Use the metric as a leading indicator of system health
Build trust by closing the loop with workers
Focus on permanent controls, not temporary patches
7. Practical Example (in the spirit of the episode)
A facility identifies 60 hazards in a month. Of those:
48 are resolved
12 remain open
Hazard Resolution Rate = 48 ÷ 60 = 80%
If the organization’s target is 90%, the gap signals slow follow‑through or resource constraints.

Aug 30, 2023 • 7min
Episode 87 - Hazard Identification and Risk Rating Metrics
Dr. Ayers explains two foundational leading indicators—Hazard Identification Metrics and Risk Rating Metrics—and how they work together to show not just how many hazards an organization finds, but how serious those hazards are. The episode emphasizes that strong safety systems don’t just count hazards; they evaluate risk, prioritize, and drive action.
These metrics reveal whether an organization is truly seeing its risk landscape or simply checking boxes.
1. Hazard Identification Metrics
These metrics measure how effectively the organization is finding hazards. They answer questions like:
Are workers and supervisors actively identifying hazards?
Are hazard reports increasing, decreasing, or stagnant?
Are we finding hazards across all departments or only in certain areas?
Are leaders spending enough time in the field to see real conditions?
What They Track
Number of hazards identified
Hazard identification rate per worker or per labor hour
Distribution of hazards (by department, shift, task, etc.)
Who is identifying hazards (frontline workers vs. supervisors vs. safety staff)
Why They Matter
High identification = engaged workforce
Low identification = fear, apathy, or lack of field presence
They reveal whether the organization is truly “looking for risk”
2. Risk Rating Metrics
Once hazards are identified, the next step is to rate their risk so the organization can prioritize.
Risk Rating Metrics evaluate:
Severity of potential harm
Likelihood of occurrence
Overall risk level (using a matrix such as 3×3, 4×4, or 5×5)
Distribution of risk across the organization
What They Reveal
Whether the organization is finding mostly low‑risk hazards
Whether high‑risk hazards are being identified and escalated
Whether risk ratings are consistent across teams
Whether leaders understand credible worst‑case severity
Why They Matter
They prevent “hazard blindness” where all hazards are treated equally
They help leaders allocate resources to the highest‑risk issues
They show whether the organization is improving or degrading over time
3. How the Two Metrics Work Together
Dr. Ayers emphasizes that neither metric is meaningful alone:
High identification + low risk ratings → workers may be finding only minor issues
Low identification + high risk ratings → workers may be afraid to report
High identification + high risk ratings → strong visibility into real risk
Low identification + low risk ratings → dangerous blind spots
Together, these metrics show:
Volume of hazards
Quality of hazard identification
Risk distribution
Prioritization needs
Cultural health
4. Common Pitfalls
Dr. Ayers highlights several traps organizations fall into:
Counting hazards without rating them Leads to poor prioritization.
Rating hazards without finding enough of them Indicates weak field engagement.
Inconsistent risk scoring Teams interpret severity and likelihood differently.
Ignoring credible worst‑case severity Underestimates true risk.
Using the metrics to punish This kills reporting instantly.
5. How to Improve These Metrics
A. Increase field engagement
Leaders must spend time where the work happens.
B. Train teams on consistent risk scoring
Use examples, calibration exercises, and group scoring.
C. Encourage reporting
Reward identification, not silence.
D. Prioritize high‑risk hazards
Fix severe hazards first, even if they are rare.
E. Track trends over time
Look for patterns in both identification and risk levels.
6. Leadership Takeaways
Strong safety leaders:
Treat hazard identification as a positive behavior
Use risk ratings to prioritize action, not justify inaction
Look for patterns, not isolated numbers
Build a culture where workers feel safe reporting hazards
Use these metrics as leading indicators of system health
7. Practical Example (in the spirit of the episode)
A facility identifies 100 hazards in a quarter:
70 are low‑risk
25 are medium‑risk
5 are high‑risk
If the previous quarter had 0 high‑risk hazards identified, this doesn’t mean risk increased—it may mean workers are finally identifying the real hazards that were always there.
This is why identification metrics and risk rating metrics must be interpreted together.

Aug 29, 2023 • 2min
Episode 86 - Safety Metrics
Dr. Ayers introduces the purpose, structure, and limitations of safety metrics, emphasizing that metrics should help leaders understand system performance, predict future risk, and drive action—not simply generate reports. The episode stresses that many organizations misuse metrics by focusing on lagging indicators or treating numbers as goals instead of tools.
This episode sets the stage for the entire safety‑metrics series.
1. What Safety Metrics Are Supposed to Do
Dr. Ayers explains that effective safety metrics should:
Reveal system health, not just outcomes
Predict future risk, not just record past injuries
Guide decision‑making
Highlight weak processes
Support resource allocation
Drive continuous improvement
Metrics are diagnostic tools, not scorecards.
2. The Problem With Traditional Safety Metrics
The episode critiques the overreliance on lagging indicators such as:
Total Recordable Incident Rate (TRIR)
Lost‑Time Injury Rate (LTIR)
Days Away, Restricted, or Transferred (DART)
These metrics:
Reflect past events, not current risk
Are influenced by reporting culture, not actual safety
Can be manipulated through classification decisions
Often drive fear‑based behaviors
Do not help leaders understand why incidents occur
Lagging indicators are necessary but not sufficient.
3. The Shift Toward Leading Indicators
Dr. Ayers emphasizes the need for leading indicators—metrics that measure the inputs to safety, not the outputs.
Examples include:
Hazard identification
Hazard resolution
Training completion
Equipment maintenance
Field engagement
Risk assessments
Quality of controls
Leading indicators help leaders:
See risk before it becomes an incident
Identify weak processes
Strengthen systems proactively
Build trust with workers
4. Characteristics of Good Safety Metrics
According to the episode, strong metrics are:
A. Actionable
They point to a specific behavior or process that can be improved.
B. Understandable
Frontline workers and executives should interpret them the same way.
C. Measurable
Data must be reliable and consistently collected.
D. Relevant
Metrics must reflect real hazards and real work.
E. Leading
They should predict future performance, not just describe the past.
5. Common Pitfalls in Safety Metrics
Dr. Ayers highlights several traps:
Using metrics as goals instead of tools (“We must hit zero injuries” creates fear and underreporting.)
Focusing on quantity instead of quality Counting inspections without evaluating their effectiveness.
Measuring what’s easy, not what matters Convenience often replaces relevance.
Failing to validate data Many organizations discover their numbers are inaccurate.
Ignoring context A high number of hazards found may indicate strong engagement, not poor safety.
6. How Leaders Should Use Safety Metrics
Strong safety leaders:
Look for trends, not isolated numbers
Use metrics to ask better questions, not assign blame
Pair leading and lagging indicators for a full picture
Share metrics transparently with workers
Use metrics to prioritize resources
Treat metrics as conversation starters
Metrics should drive learning, not fear.
7. Practical Example (in the spirit of the episode)
A site reports:
Zero injuries
Low hazard identification
Low training completion
Poor equipment maintenance
On paper, the site looks “safe,” but the leading indicators show a high‑risk environment with weak systems and low engagement.
This is why leading indicators matter.

Aug 28, 2023 • 3min
Episode 85 - Who Should Write Equipment Procedures?
Episode 85 centers on a simple but powerful idea: the people who actually use the equipment should be the ones who write the procedures. Dr. Ayers explains that frontline employees bring practical insight, real‑world experience, and a deep understanding of how work is actually performed—making them the most qualified authors of safe, effective procedures.
Why Frontline Employees Should Write Procedures
Frontline workers understand the equipment in ways that supervisors, engineers, or safety staff often don’t. They know the shortcuts people are tempted to take, the steps that are easy to miss, and the conditions that make tasks harder or riskier. When they write procedures:
The steps reflect actual work, not idealized work.
The instructions are practical and realistic.
The procedure captures tribal knowledge that might otherwise be lost.
Workers feel ownership, which increases compliance and engagement.
This approach also reduces the common gap between “what the procedure says” and “what people really do.”
How Leaders Support the Process
Dr. Ayers emphasizes that leaders still play a critical role. They must:
Provide structure and expectations for the procedure format.
Facilitate collaboration between workers, maintenance, engineering, and safety.
Ensure the final procedure meets regulatory and organizational requirements.
Validate that the steps are correct, complete, and safe.
The goal is not to remove leaders from the process—it’s to shift authorship to the people closest to the work while leaders guide, review, and approve.
Benefits of Employee‑Written Procedures
Organizations that adopt this approach typically see:
Higher buy‑in and fewer workarounds.
More accurate and detailed procedures.
Stronger safety culture through participation.
Better identification of hazards and failure points.
Increased consistency across shifts and teams.
When workers help create the procedures they follow, they are far more likely to trust them and use them.
Leadership Takeaway
The most effective equipment procedures are written with the people who perform the work—not handed down to them. Leaders who empower employees to write procedures build stronger systems, safer operations, and a more engaged workforce.


