AI SkillDesign support opsCustomer Success

When support volume grows, /support-operations designs SLA policies and tier structures, so you can scale without quality drops. — Claude Skill

A Claude Skill for Claude Code by Nick Jensen — run /support-operations in Claude·Updated

Compatible withChatGPT·Claude·Gemini·OpenClaw

Build SLA policies, escalation tiers, and support metric frameworks.

  • SLA policy design with response and resolution targets by priority and tier
  • L1/L2/L3 escalation structures with routing rules and ownership clarity
  • Knowledge base architecture with article templates and search optimization
  • Support metric dashboards tracking CSAT, FRT, TTR, and FCR
  • Ticket categorization taxonomies for root cause analysis and product feedback loops

Who this is for

What it does

SLA framework buildout

Run /support-operations with your current ticket volumes and customer tiers to generate SLA policies with response times, resolution targets, and breach escalation procedures.

Tier structure design

Use /support-operations to define L1/L2/L3 responsibilities, skill requirements, routing logic, and escalation triggers — sized to your team and ticket mix.

Knowledge base launch

Feed /support-operations your top 50 ticket categories to generate a knowledge base architecture with article templates, tagging taxonomy, and self-serve deflection targets.

Support metrics overhaul

Run /support-operations to build a support scorecard tracking FRT, TTR, FCR, and CSAT with benchmarks, alerting thresholds, and team-level breakdowns.

How it works

1

Analyze your current support landscape — ticket volumes, categories, team size, tools, and existing SLAs — to establish the baseline.

2

Design tiered SLA policies matching response and resolution commitments to customer segments, priority levels, and contract terms.

3

Build the escalation framework: L1 scope and scripts, L2 technical depth, L3 engineering handoff criteria, and management escalation triggers.

4

Create the knowledge base plan with article hierarchy, content templates, review cadence, and metrics for deflection tracking.

5

Deliver the complete ops package: SLA matrix, escalation flowcharts, KB architecture, metric dashboards, and staffing model.

Example

Support org context
Team: 8 agents, 1 lead. Monthly tickets: 1,200. Current FRT: 4.2 hours. CSAT: 78%. No formal SLAs. No knowledge base. Tools: Zendesk. Customer tiers: Enterprise (15%), Mid-Market (45%), SMB (40%). Top categories: login issues (22%), API errors (18%), billing questions (15%).
Support operations blueprint
SLA Matrix
Enterprise P1: 15min response / 4hr resolution
Enterprise P2: 1hr response / 8hr resolution
Mid-Market P1: 30min response / 8hr resolution
Mid-Market P2: 2hr response / 24hr resolution
SMB P1: 1hr response / 24hr resolution
SMB P2: 4hr response / 48hr resolution
Tier Structure
L1 (5 agents): Login resets, billing questions, known-issue triage. Target: resolve 65% of tickets.
L2 (2 agents): API errors, configuration issues, integration debugging. Target: resolve 30% of tickets.
L3 (1 agent + eng rotation): Product bugs, data issues, custom implementations. Target: 5% of tickets.
Target Metrics
FRT: 4.2hr -> 1.5hr (within 90 days)
CSAT: 78% -> 88% (within 6 months)
FCR: Establish baseline, target 72%
KB deflection: 0% -> 25% within 6 months (focus on login + billing articles first)

Metrics this improves

Activation Rate
+10-20%
Customer Success
Engagement
+15-25%
Customer Success

Works with

Support Operations

Strategic support operations expertise for customer-facing teams — from ticket management and SLA design to escalation workflows and self-service optimization.

Philosophy

Great support isn't about closing tickets fast. It's about solving customer problems permanently while building scalable systems.

The best support operations teams:

  1. Prevent before they support — Self-service and proactive help reduce ticket volume
  2. Measure what drives loyalty — Resolution quality beats response speed
  3. Escalate with context — Every handoff preserves customer history
  4. Feed insights upstream — Support data drives product and success improvements

How This Skill Works

When invoked, apply the guidelines in rules/ organized by:

  • ticket-* — Ticket management, prioritization, queue optimization
  • sla-* — SLA design, compliance monitoring, escalation triggers
  • tier-* — Support tier structure, skill-based routing, specialization
  • knowledge-* — Knowledge base strategy, self-service, deflection
  • metrics-* — CSAT, FRT, TTR, FCR, quality scoring
  • escalation-* — Severity definitions, escalation paths, incident management
  • tooling-* — Support stack optimization, integrations, automation
  • feedback-* — Support-to-CS handoffs, product feedback loops, voice of customer

Core Frameworks

The Support Operations Hierarchy

LevelFocusMetricsOwner
TicketsIndividual resolutionHandle time, CSATAgents
QueueFlow optimizationWait time, backlogTeam leads
ChannelChannel effectivenessDeflection, containmentManagers
OperationsSystem performanceCost per ticket, NPSDirectors
StrategyBusiness impactRetention, expansionVP/C-level

The Support Tier Model

┌─────────────────────────────────────────────────────────────────┐
│                         TIER 3 (L3)                              │
│  Engineering escalation, code-level issues, custom development  │
│  Target: <5% of tickets | SLA: Best effort                      │
├─────────────────────────────────────────────────────────────────┤
│                         TIER 2 (L2)                              │
│  Technical specialists, complex troubleshooting, integrations   │
│  Target: 15-25% of tickets | SLA: 4-8 hours                     │
├─────────────────────────────────────────────────────────────────┤
│                         TIER 1 (L1)                              │
│  First response, common issues, documentation guidance          │
│  Target: 60-80% resolution | SLA: 15-60 minutes                 │
├─────────────────────────────────────────────────────────────────┤
│                      SELF-SERVICE (L0)                           │
│  Knowledge base, chatbots, community forums, in-app help        │
│  Target: 30-50% deflection | SLA: Instant                       │
└─────────────────────────────────────────────────────────────────┘

Ticket Priority Matrix

PriorityBusiness ImpactResponse SLAResolution SLAExamples
P1 CriticalComplete outage, data loss15 min4 hoursSystem down, security breach
P2 HighMajor feature broken1 hour8 hoursKey workflow blocked
P3 MediumFeature impaired4 hours24 hoursPartial functionality
P4 LowMinor issue, cosmetic8 hours72 hoursUI bug, minor inconvenience
P5 RequestFeature request, how-to24 hours5 daysEnhancement, training

Support Metrics Framework

MetricDefinitionTargetWarning
CSATCustomer satisfaction score90%+<85%
FRTFirst response time<1 hour>4 hours
TTRTime to resolution<24 hours>72 hours
FCRFirst contact resolution70%+<50%
NPSNet promoter score30+<10
Ticket VolumeTickets per 100 customers5-15>25
Deflection RateSelf-service success30-50%<20%
Escalation RateTickets escalated10-20%>30%
Reopen RateTickets reopened<5%>10%
Agent UtilizationProductive time70-80%<60% or >90%

The Ticket Lifecycle

┌─────────────────────────────────────────────────────────────────┐
│                                                                  │
│  NEW → TRIAGED → ASSIGNED → IN PROGRESS → PENDING → RESOLVED   │
│                                    │          │                  │
│                                    ▼          ▼                  │
│                              ESCALATED    WAITING                │
│                                    │     (Customer)              │
│                                    ▼                             │
│                              ENGINEERING                         │
│                                                                  │
└─────────────────────────────────────────────────────────────────┘

Channel Strategy Matrix

ChannelBest ForCostScalabilityPersonal
Self-serviceCommon issuesLowestHighestLowest
ChatbotQuick questionsLowHighLow
Live chatReal-time helpMediumMediumMedium
Email/TicketComplex issuesMediumMediumMedium
PhoneUrgent/sensitiveHighLowHigh
VideoTechnical demosHighLowHighest

Severity Levels

SeverityDefinitionEscalation PathCommunication
SEV1System-wide outageImmediate to engineering + execStatus page, proactive email
SEV2Major feature broken1 hour to L3Affected users notified
SEV3Feature degraded4 hours to L2Standard ticket updates
SEV4Minor impactNormal queueStandard ticket updates

Key Formulas

Cost Per Ticket

Cost Per Ticket = (Total Support Cost) / (Total Tickets Handled)
Target: $5-25 depending on complexity

Support Capacity Planning

Required Agents = (Ticket Volume × Handle Time) / (Available Hours × Utilization Rate)

Example:
(500 tickets × 20 min) / (8 hours × 60 min × 0.75) = 28 agents

Self-Service ROI

Savings = (Deflected Tickets × Cost Per Ticket) - Self-Service Investment

Anti-Patterns

  • Speed over quality — Fast wrong answers create repeat contacts
  • Ticket tennis — Multiple handoffs without resolution
  • Knowledge hoarding — Solutions in heads, not documentation
  • Metric gaming — Closing tickets prematurely to hit targets
  • Escalation avoidance — L1 struggling when L2 is needed
  • Channel forcing — Making customers switch channels unnecessarily
  • Copy-paste responses — Generic answers that don't address the issue
  • Invisible backlog — Tickets aging without visibility
  • No feedback loop — Support insights never reach product
  • Over-automation — Bots handling issues that need humans

Reference documents


title: Section Organization

1. Ticket Management (ticket)

Impact: CRITICAL Description: Ticket prioritization, queue management, triage workflows, and assignment rules. The foundation of efficient support operations.

2. SLA Design (sla)

Impact: CRITICAL Description: Service level agreement creation, compliance monitoring, business hours configuration, and breach prevention strategies.

3. Support Tier Structure (tier)

Impact: CRITICAL Description: L1/L2/L3 tier design, skill-based routing, specialization paths, and escalation criteria between tiers.

4. Knowledge Base Strategy (knowledge)

Impact: HIGH Description: Self-service optimization, article structure, deflection measurement, and knowledge maintenance workflows.

5. Support Metrics (metrics)

Impact: HIGH Description: CSAT, FRT, TTR, FCR measurement, quality scoring, agent performance, and operational dashboards.

6. Escalation Procedures (escalation)

Impact: CRITICAL Description: Severity definitions, escalation paths, incident management, on-call rotations, and war room protocols.

7. Support Tool Stack (tooling)

Impact: MEDIUM-HIGH Description: Help desk configuration, integrations, automation rules, and multi-channel orchestration.

8. Support-to-CS Feedback Loops (feedback)

Impact: HIGH Description: Voice of customer programs, product feedback channels, at-risk customer handoffs, and cross-functional insights sharing.

9. Quality Assurance (quality)

Impact: HIGH Description: Ticket review processes, agent coaching, quality scoring rubrics, and continuous improvement programs.

10. Workforce Management (workforce)

Impact: MEDIUM-HIGH Description: Capacity planning, scheduling, forecasting ticket volume, and optimizing agent utilization.


title: Agent Onboarding & Training Programs impact: HIGH tags: onboarding, training, development, certification, enablement

Agent Onboarding & Training Programs

Impact: HIGH

Well-trained agents resolve issues faster, create better customer experiences, and stay longer. Poor onboarding creates knowledge gaps that persist for months. Invest in training early and continuously for long-term support excellence.

The Agent Development Journey

┌─────────────────────────────────────────────────────────────────┐
│                     ONBOARDING (Week 1-8)                        │
│  Company culture, product training, tools, shadowing, ramp      │
├─────────────────────────────────────────────────────────────────┤
│                     PROFICIENCY (Month 2-6)                      │
│  Independent work, specialty development, quality improvement   │
├─────────────────────────────────────────────────────────────────┤
│                     MASTERY (Month 6-18)                         │
│  L2 skills, mentoring, process improvement, specialization      │
├─────────────────────────────────────────────────────────────────┤
│                     LEADERSHIP (Year 2+)                         │
│  Team lead, trainer, specialist, or management track            │
└─────────────────────────────────────────────────────────────────┘

Onboarding Program Structure

WeekFocusActivitiesSuccess Measure
1Company & CultureOrientation, values, meet teamsQuiz, culture buddy assigned
2Product Deep DiveFeatures, use cases, hands-onProduct certification
3Tools & SystemsHelp desk, CRM, integrationsTool proficiency check
4ShadowingObserve senior agents, discussShadow log completed
5-6Supervised TicketsHandle tickets with mentor reviewQA score 70+
7-8Increasing LoadFull ticket queue, decreasing oversightQA score 80+, FCR 60%+

Day 1 Checklist

Pre-Arrival:
□ Workstation/laptop ready
□ System access provisioned
□ Email and Slack accounts created
□ Training schedule shared
□ Buddy/mentor assigned
□ Welcome kit prepared

Day 1 Schedule:
9:00  - Welcome from manager
9:30  - HR orientation
10:30 - IT setup and access
11:30 - Lunch with team
1:00  - Company overview presentation
2:00  - Support team overview
3:00  - Product intro (high level)
4:00  - Tour and introductions
4:30  - Wrap up with manager, questions

Good Onboarding Practices

✓ Structured curriculum
  → Clear learning path
  → Defined milestones
  → Assessment at each stage

✓ Blended learning
  → Mix of self-paced and instructor-led
  → Videos, docs, hands-on practice
  → Live Q&A sessions

✓ Gradual ramp
  → Don't throw into queue day 1
  → Shadowing before handling
  → Supervised before independent

✓ Assigned mentor
  → Go-to person for questions
  → Regular check-ins
  → Safe space to learn

✓ Early feedback
  → Weekly 1:1s during onboarding
  → QA on all tickets initially
  → Course corrections early

Bad Onboarding Practices

✗ Drink from firehose
  → Too much info in week 1
  → Retention is minimal

✗ Sink or swim
  → "Here's the queue, good luck"
  → Mistakes damage customer experience

✗ Tribal knowledge only
  → Learn from whoever's around
  → Inconsistent information

✗ No documentation
  → "Ask Sarah about that"
  → Single points of failure

✗ Same for everyone
  → Experienced hires treated like new grads
  → New grads expected to know basics

✗ Training then forgotten
  → No ongoing development
  → Skills stagnate

Product Training Curriculum

ModuleContentDurationAssessment
Core ProductMain features, workflows4 hoursQuiz + demo
IntegrationsCommon integrations, setup2 hoursQuiz
Admin/SettingsConfiguration, user management2 hoursQuiz
BillingPlans, upgrades, invoices1 hourQuiz
API BasicsWhen to use, common endpoints2 hoursQuiz
MobileMobile app features, differences1 hourQuiz
Common IssuesTop 20 tickets, solutions4 hoursCase studies

Soft Skills Training

SkillWhy It MattersTraining Method
EmpathyCustomers feel heardRole play, customer stories
Clear writingFirst-time understandingWriting workshops, reviews
Active listeningCorrect problem solvingRole play, call shadowing
De-escalationAngry customer handlingScenarios, real examples
Time managementEfficiency without rushingTips, workflow optimization
Saying no nicelyDecline without damageScripts, practice

Training Content Types

TypeBest ForExample
VideosVisual processes, demosProduct feature walkthroughs
Written docsReference, detailed stepsTroubleshooting guides
InteractivePractice, explorationSandbox environment
Live sessionsQ&A, nuance, cultureWeekly new hire sessions
ShadowingReal-world contextSitting with experienced agent
Case studiesComplex scenarios"Here's what happened, what would you do?"

Certification Program

Certification Levels:

LEVEL 1: Foundational (Week 8)
├── Product fundamentals
├── Core tools proficiency
├── Basic troubleshooting
└── Assessment: 80% quiz + 75% QA

LEVEL 2: Proficient (Month 4)
├── Advanced features
├── Integration knowledge
├── Complex troubleshooting
└── Assessment: 85% quiz + 85% QA + case study

LEVEL 3: Specialist (Month 8+)
├── Deep specialty area
├── Escalation handling
├── Mentoring capability
└── Assessment: 90% quiz + 90% QA + presentation

LEVEL 4: Expert (Year 2+)
├── Cross-functional knowledge
├── Process improvement
├── Training delivery
└── Assessment: Peer review + manager recommendation

Ongoing Training Program

CadenceTraining TypeContent
WeeklyTeam meeting learningOne topic deep dive
MonthlySkills workshopCommunication, efficiency, etc.
QuarterlyProduct updatesNew features, changes
Bi-annualCertification renewalRefresh assessments
As neededIssue-specificTrending problems

Knowledge Assessment Methods

MethodWhat It TestsWhen to Use
QuizKnowledge recallAfter content modules
Case studyApplicationEnd of training sections
Live demoHands-on abilityProduct knowledge
Role playSoft skillsCommunication training
QA reviewReal-world applicationOngoing
Peer teachingDeep understandingMastery validation

Training Metrics

MetricDefinitionTarget
Time to proficiencyDays to hit FCR target<30 days
Certification pass rate% passing on first try85%+
Training completion% completing curriculum100%
Post-training QAQA score after training80+
Knowledge retentionScore on refresh quiz85%+ at 6 months
Training satisfactionNew hire feedback4.5/5

Mentor Program

Mentor Responsibilities:
├── Daily check-in during onboarding
├── Answer questions without judgment
├── Review first 20 tickets together
├── Provide feedback on approach
├── Escalate concerns to manager
└── Be available for ad-hoc support

Mentor Selection:
├── 6+ months tenure
├── QA score 85+
├── Interest in developing others
├── Patient communication style
└── Time allocated for mentoring

Training Resources Checklist

Documentation:
□ Product documentation (internal and customer-facing)
□ Troubleshooting guides
□ Process documentation
□ FAQ and common issues
□ Escalation procedures

Tools:
□ LMS or training platform
□ Product sandbox/test environment
□ Recording/video tools
□ Quiz/assessment platform
□ Knowledge base

People:
□ Assigned mentor
□ Training facilitators
□ Subject matter experts for sessions
□ Manager for 1:1s

Career Development Paths

                    ┌─────────────────┐
                    │  Support Agent  │
                    └────────┬────────┘
                             │
         ┌───────────────────┼───────────────────┐
         ▼                   ▼                   ▼
  ┌─────────────┐    ┌─────────────────┐   ┌──────────────┐
  │ Specialist  │    │ Team Lead/Mgr   │   │ Trainer      │
  │ Track       │    │ Track           │   │ Track        │
  ├─────────────┤    ├─────────────────┤   ├──────────────┤
  │ L2 Specialist│   │ Team Lead       │   │ Training Spec│
  │ L3 Engineer │    │ Support Manager │   │ Enablement   │
  │ Solutions   │    │ Support Director│   │ Trainer Lead │
  │ Architect   │    │ VP Support      │   │ Content Lead │
  └─────────────┘    └─────────────────┘   └──────────────┘

Common Onboarding Mistakes

MistakeImpactSolution
Information overloadLow retentionSpace learning over weeks
No practice timeTheory onlyInclude sandbox exercises
Skipping soft skillsPoor communicationIntegrate throughout
Generic trainingIrrelevant contentRole-specific paths
No feedback loopIssues persistRegular check-ins
Ending at week 4Knowledge gapsOngoing program

Anti-Patterns

  • Training as checkbox — Completed but not learned
  • Firehose approach — All information in week 1
  • No hands-on practice — Reading without doing
  • Tribal knowledge only — "Ask Bob"
  • One-time training — No ongoing development
  • Ignoring soft skills — Technical only
  • No assessment — Assume learning happened
  • Senior agent = trainer — Good agent ≠ good teacher

title: Escalation Procedures & Severity Definitions impact: CRITICAL tags: escalation, severity, incident, on-call, war-room, crisis

Escalation Procedures & Severity Definitions

Impact: CRITICAL

Clear escalation procedures ensure critical issues reach the right people fast. Vague severity definitions lead to under-escalation (customers suffer) or over-escalation (alert fatigue). Get this right and crises become manageable incidents.

Severity Level Definitions

SeverityNameDefinitionExamples
SEV1CriticalComplete system outage, data loss, security breachProduction down, data breach, complete loss of service
SEV2HighMajor feature broken for many usersCore workflow blocked, payment processing down
SEV3MediumFeature degraded or broken for some usersSlow performance, partial functionality
SEV4LowMinor issue, workaround availableUI bug, cosmetic issue, edge case
SEV5MinimalEnhancement request, how-to questionFeature request, documentation question

Severity Decision Matrix

                    ┌─────────────────────────────────────┐
                    │            USERS AFFECTED           │
                    ├──────────┬──────────┬───────────────┤
                    │ One User │ Multiple │ All Users     │
┌───────────────────┼──────────┼──────────┼───────────────┤
│ Complete Block    │   SEV3   │   SEV2   │     SEV1      │
├───────────────────┼──────────┼──────────┼───────────────┤
│ Major Impairment  │   SEV4   │   SEV3   │     SEV2      │
├───────────────────┼──────────┼──────────┼───────────────┤
│ Minor Impairment  │   SEV5   │   SEV4   │     SEV3      │
└───────────────────┴──────────┴──────────┴───────────────┘

Escalation Paths by Severity

SeverityFirst Response30 min1 hour2 hours4 hours
SEV1On-call engineerVP EngCTOCEOBoard update
SEV2L3 engineerEng ManagerVP EngCTO-
SEV3L2 specialistL3 engineerEng Manager--
SEV4L1 agentL2 specialist---
SEV5L1 agent----

Escalation Response Requirements

SeverityResponse SLAUpdate FrequencyCommunication
SEV115 minutesEvery 15 minutesStatus page, proactive email, exec bridge
SEV230 minutesEvery 30 minutesAffected users, internal Slack
SEV31 hourEvery 2 hoursTicket updates
SEV44 hoursEvery 8 hoursTicket updates
SEV58 hoursAt resolutionTicket updates

Good Escalation Practices

✓ Clear severity criteria
  → Written definitions with examples
  → Decision tree for ambiguous cases
  → Regular calibration across team

✓ No-blame escalation culture
  → Better to escalate than miss
  → Learning from near-misses
  → Celebrate good catches

✓ Context travels with escalation
  → Full ticket history
  → Troubleshooting already done
  → Customer impact documented

✓ Defined communication templates
  → Internal update format
  → Customer communication format
  → Status page update process

✓ Post-incident review
  → Every SEV1/SEV2 gets review
  → Blameless root cause analysis
  → Action items tracked to completion

Bad Escalation Practices

✗ Escalation as punishment
  → "You should have handled this"
  → Agents afraid to escalate

✗ No severity definitions
  → Everything is SEV1 or nothing is
  → Inconsistent customer experience

✗ Escalation black holes
  → Tickets go up, never come back
  → Engineering ignores queue

✗ Context-free handoffs
  → "It's broken" with no details
  → Customer re-explains 3 times

✗ No update cadence
  → Customer in the dark
  → Internal teams unaware

✗ Skip-level escalations
  → Everyone emails the CEO
  → Process bypassed

On-Call Structure

TierCoverageResponsibilities
Primary On-Call24/7First responder for all alerts
Secondary On-Call24/7Backup if primary unreachable
Manager On-Call24/7Escalation point, decision maker
Executive On-Call24/7Major incident communication

On-Call Rotation Best Practices

Rotation Schedule:
├── 1-week rotations (not longer)
├── Minimum 2 people per rotation
├── Time zone coverage consideration
├── Holidays covered and compensated
└── Handoff at consistent time (e.g., Monday 9am)

On-Call Expectations:
├── Respond to page within 5 minutes
├── Acknowledge or escalate within 15 minutes
├── Laptop and internet access required
├── Not impaired (alcohol, etc.)
└── Backup coverage for personal conflicts

Incident Command Structure (SEV1/SEV2)

RoleResponsibilityExample Actions
Incident CommanderOverall coordinationRuns bridge, makes decisions
Technical LeadTechnical investigationDebugging, deploying fixes
Communications LeadUpdates stakeholdersStatus page, customer emails
ScribeDocumentationTimeline, actions, decisions
Subject Matter ExpertsSpecific expertiseCalled in as needed

War Room Protocol (SEV1)

Initiation (0-15 minutes):
□ Page on-call and backup
□ Open incident bridge (Zoom/Slack)
□ Assign Incident Commander
□ Initial impact assessment
□ Post to status page

Active Incident:
□ Update status page every 15 minutes
□ Scribe maintains timeline
□ IC coordinates investigation
□ Parallel workstreams if needed
□ All changes announced before made

Resolution:
□ Confirm customer impact resolved
□ Update status page to resolved
□ Send all-clear internally
□ Draft customer communication
□ Schedule post-incident review

Customer Communication Templates

SEV1 - Initial Notification:

Subject: [Service Name] - Service Disruption

We are currently experiencing an issue affecting [description].

Impact: [What customers are experiencing]
Status: Our team is actively investigating
Next Update: Within 30 minutes

We apologize for the inconvenience and will keep you updated.

[Link to status page]

SEV1 - Resolution:

Subject: [Service Name] - Issue Resolved

The issue affecting [description] has been resolved.

Duration: [Start time] to [End time]
Impact: [What was affected]
Resolution: [Brief explanation]

We apologize for any inconvenience caused. A detailed
post-incident report will be shared within 48 hours.

[Link to status page]

Internal Escalation Template

## Escalation Summary

**Severity:** [SEV1/2/3/4]
**Time Detected:** [Timestamp]
**Escalated By:** [Name]
**Escalated To:** [Name/Team]

### Issue Description
[2-3 sentences describing the problem]

### Customer Impact
- Affected customers: [Number or percentage]
- Impact type: [Outage/Degradation/Error]
- Customer-facing symptoms: [What they're seeing]

### Investigation So Far
- [Step 1] → [Result]
- [Step 2] → [Result]
- [Current hypothesis]

### Immediate Actions Needed
1. [Action 1]
2. [Action 2]

### Communication Status
- Status page: [Updated/Pending]
- Customer notification: [Sent/Pending/Not needed]
- Internal channel: [Posted in #channel]

Post-Incident Review Template

## Post-Incident Review: [Incident Title]

**Date:** [Date]
**Duration:** [Start to End]
**Severity:** [SEV level]
**Incident Commander:** [Name]

### Summary
[1-2 paragraph summary of what happened]

### Timeline
- [HH:MM] First customer report
- [HH:MM] On-call paged
- [HH:MM] Incident declared
- [HH:MM] Root cause identified
- [HH:MM] Fix deployed
- [HH:MM] Customer impact resolved

### Impact
- Customers affected: [Number]
- Duration of impact: [Time]
- Revenue impact: [If applicable]
- SLA impact: [Breaches if any]

### Root Cause
[Technical explanation of what went wrong]

### What Went Well
- [Positive 1]
- [Positive 2]

### What Could Be Improved
- [Improvement 1]
- [Improvement 2]

### Action Items
| Action | Owner | Due Date | Status |
|--------|-------|----------|--------|
| [Action 1] | [Name] | [Date] | Open |
| [Action 2] | [Name] | [Date] | Open |

Anti-Patterns

  • Severity inflation — Everything is SEV1, real emergencies ignored
  • Escalation avoidance — Agents sit on critical issues too long
  • Hero culture — One person handles all incidents alone
  • No post-mortems — Same issues repeat without learning
  • Customer last — Internal comms before customer updates
  • War room theater — 50 people on a call, 3 doing work
  • Blame-focused reviews — "Who did this?" instead of "How do we prevent?"
  • On-call burnout — Same people always on call, no rotation

title: Support-to-CS Feedback Loops impact: HIGH tags: feedback, voice-of-customer, product-feedback, handoffs, insights

Support-to-CS Feedback Loops

Impact: HIGH

Support is on the front lines of customer experience. Every ticket is a data point. When feedback flows to CS, Product, and Engineering, the entire organization improves. When it doesn't, support becomes a cost center instead of a strategic asset.

The Feedback Loop Ecosystem

┌─────────────────────────────────────────────────────────────────┐
│                        SUPPORT TEAM                              │
│  Tickets, chats, calls, patterns, sentiment                     │
└──────────────────────────┬──────────────────────────────────────┘
                           │
           ┌───────────────┼───────────────┐
           ▼               ▼               ▼
    ┌─────────────┐ ┌─────────────┐ ┌─────────────┐
    │  CUSTOMER   │ │   PRODUCT   │ │ ENGINEERING │
    │   SUCCESS   │ │    TEAM     │ │    TEAM     │
    │             │ │             │ │             │
    │ At-risk     │ │ Feature     │ │ Bug reports │
    │ accounts    │ │ requests    │ │ Priority    │
    │ Expansion   │ │ Pain points │ │ Patterns    │
    │ signals     │ │ Use cases   │ │ Root cause  │
    └─────────────┘ └─────────────┘ └─────────────┘
           │               │               │
           └───────────────┼───────────────┘
                           ▼
              ┌───────────────────────┐
              │   IMPROVED PRODUCT    │
              │   BETTER EXPERIENCE   │
              │   REDUCED TICKETS     │
              └───────────────────────┘

Feedback Types

TypeDefinitionDestinationAction
Bug ReportProduct not working as expectedEngineeringFix and close
Feature RequestNew capability wantedProductEvaluate and roadmap
UX FeedbackConfusion or frictionProduct/DesignImprove usability
Documentation GapMissing or unclear docsContent/DocsUpdate content
At-Risk SignalCustomer frustration patternCustomer SuccessIntervention
Praise/DelightPositive feedbackCS/MarketingCase study, retention
Competitive IntelMentions of competitorsProduct/SalesPositioning update

Good Feedback Loop Practices

✓ Structured capture
  → Standard fields for feedback type
  → Linked to customer and ticket
  → Searchable and reportable

✓ Regular synthesis
  → Weekly top themes shared
  → Monthly trend analysis
  → Quarterly deep dives

✓ Closed loop
  → Product responds to feedback
  → Support informed of outcomes
  → Customer notified when fixed

✓ Cross-functional visibility
  → Product sees real customer words
  → Engineering understands impact
  → CS sees patterns before escalation

✓ Metrics on feedback impact
  → Tickets reduced by product fixes
  → NPS improved by changes
  → Feature adoption from requests

Bad Feedback Loop Practices

✗ Feedback graveyard
  → Collected but never reviewed
  → No one reads the reports

✗ Lost in translation
  → Support paraphrases poorly
  → Customer voice lost

✗ One-way communication
  → Product never responds
  → "We'll pass it along" forever

✗ No aggregation
  → Individual tickets, no patterns
  → Same issue reported 100 times separately

✗ Credit hoarding
  → "This was my idea"
  → Collaboration suffers

Support-to-CS Handoff Triggers

SignalHandoff TypeUrgency
Multiple escalations in 30 daysAt-risk accountHigh
Executive complaintVIP escalationImmediate
Churn mentionSave opportunityHigh
Expansion interestUpsell leadMedium
Champion departureRelationship riskHigh
Negative CSAT trendProactive interventionMedium
Major incident impactExecutive touchHigh
Feature blocker identifiedProduct alignmentMedium

At-Risk Account Alert Template

## At-Risk Alert: [Account Name]

**Date:** [Date]
**Flagged By:** [Support Agent]
**CSM:** [CSM Name]
**Account Tier:** [Tier]
**MRR:** [Amount]

### Risk Indicators
- [Indicator 1: e.g., "3 escalations in past 2 weeks"]
- [Indicator 2: e.g., "Champion expressing frustration"]
- [Indicator 3: e.g., "Mentioned evaluating competitors"]

### Recent Ticket Summary
| Ticket | Date | Issue | Status |
|--------|------|-------|--------|
| #1234 | [Date] | [Issue] | [Status] |
| #1235 | [Date] | [Issue] | [Status] |

### Customer Sentiment
[Direct quotes from recent tickets showing frustration]

### Recommended Actions
1. [Suggested CS intervention]
2. [Suggested exec outreach if needed]
3. [Product team involvement if needed]

### Supporting Links
- [Link to customer account]
- [Link to tickets]
- [Link to health score]

Product Feedback Submission

## Product Feedback: [Brief Title]

**Feedback Type:** [Feature Request / Bug / UX Issue / Other]
**Submitted By:** [Support Agent]
**Date:** [Date]

### Customer Context
- Customers requesting: [Number]
- Segments: [Enterprise/SMB/Startup]
- Revenue represented: [$ or %]
- Notable accounts: [Names if relevant]

### The Request
[Clear description of what customers are asking for]

### Customer Voice
> [Direct quote from customer 1]

> [Direct quote from customer 2]

### Use Case
[Why customers need this, what problem it solves]

### Current Workaround
[How customers are working around this today, if any]

### Business Impact
- Churn risk: [Low/Medium/High]
- Expansion blocker: [Yes/No]
- Competitive gap: [Yes/No]

### Related Tickets
- [Link to ticket 1]
- [Link to ticket 2]

Weekly Voice of Customer Report

## Voice of Customer: Week of [Date]

### Top Themes This Week

**1. [Theme 1]** — [X] tickets
   Impact: [High/Medium/Low]
   Sample: "[Customer quote]"
   Action: [Assigned to Product/Eng/None yet]

**2. [Theme 2]** — [X] tickets
   Impact: [High/Medium/Low]
   Sample: "[Customer quote]"
   Action: [Assigned to Product/Eng/None yet]

**3. [Theme 3]** — [X] tickets
   Impact: [High/Medium/Low]
   Sample: "[Customer quote]"
   Action: [Assigned to Product/Eng/None yet]

### Bug Impact
| Bug | Tickets | Customers | Est. Resolution |
|-----|---------|-----------|-----------------|
| [Bug 1] | [X] | [Y] | [Date/Unknown] |
| [Bug 2] | [X] | [Y] | [Date/Unknown] |

### Feature Requests Trending
1. [Feature] — [X] requests
2. [Feature] — [X] requests

### Wins This Week
- [Positive feedback or success story]

### At-Risk Accounts Flagged
- [Account 1] — [CSM notified]
- [Account 2] — [CSM notified]

### Ticket Trends
- Total tickets: [X] ([+/-]% vs last week)
- Avg CSAT: [X] ([+/-] vs last week)
- Top category: [Category]

Cross-Functional Meetings

MeetingFrequencyAttendeesAgenda
Support-Product SyncWeeklySupport Lead, PMTop issues, feature requests, bug priority
Support-CS SyncWeeklySupport Lead, CS LeadAt-risk accounts, handoffs, patterns
Support-Eng SyncBi-weeklySupport Lead, Eng LeadBug prioritization, upcoming releases
Voice of CustomerMonthlySupport, Product, CS, ExecTrends, themes, strategic insights

Measuring Feedback Loop Effectiveness

MetricDefinitionTarget
Feedback SubmittedTickets tagged for feedback5-10% of tickets
Feedback Response RateProduct/Eng response to feedback80%+
Time to AcknowledgmentDays until feedback acknowledged<5 days
Tickets ReducedTickets avoided by product fixesTrack quarter-over-quarter
Feature AdoptionUsage of requested featuresHigh adoption = good feedback

Technology for Feedback Loops

Tool TypePurposeExamples
Product feedbackCollect and prioritize requestsProductboard, Canny, Pendo
Bug trackingEngineering integrationJira, Linear, GitHub
Customer successAccount health, handoffsGainsight, Totango, ChurnZero
Knowledge managementPattern documentationNotion, Guru, Confluence
CommunicationCross-team alertsSlack, Teams channels

Anti-Patterns

  • Feedback theater — Collected but ignored
  • Lost customer voice — Paraphrased beyond recognition
  • Siloed insights — Support knows things no one else does
  • Feature factory — Building what's loudest, not what's important
  • Blame forwarding — "Product needs to fix this" without partnership
  • No closed loop — Customer never hears outcome
  • Anecdote over data — One customer drives roadmap
  • Handoff and forget — CS gets alert, support never follows up

title: Knowledge Base & Self-Service Strategy impact: HIGH tags: knowledge-base, self-service, deflection, documentation, help-center

Knowledge Base & Self-Service Strategy

Impact: HIGH

A well-maintained knowledge base deflects 30-50% of support tickets while improving customer satisfaction. Customers prefer instant answers over waiting for human responses. Self-service is a win-win when done right.

The Self-Service Hierarchy

┌─────────────────────────────────────────────────────────────────┐
│                      IN-APP CONTEXTUAL HELP                      │
│  Tooltips, inline guidance, smart suggestions at point of need  │
│  Deflection: Highest | Effort: Lowest                           │
├─────────────────────────────────────────────────────────────────┤
│                          CHATBOT / AI                            │
│  Instant answers, guided troubleshooting, intelligent routing   │
│  Deflection: High | Personalization: Medium                     │
├─────────────────────────────────────────────────────────────────┤
│                       KNOWLEDGE BASE                             │
│  Searchable articles, how-to guides, troubleshooting docs       │
│  Deflection: Medium-High | Depth: High                          │
├─────────────────────────────────────────────────────────────────┤
│                      COMMUNITY FORUMS                            │
│  Peer-to-peer help, discussions, shared solutions               │
│  Deflection: Medium | Engagement: High                          │
├─────────────────────────────────────────────────────────────────┤
│                        VIDEO TUTORIALS                           │
│  Visual guides, product walkthroughs, feature deep-dives        │
│  Deflection: Medium | Learning: High                            │
└─────────────────────────────────────────────────────────────────┘

Knowledge Base Metrics

MetricDefinitionTargetWarning
Deflection Rate% users who didn't file ticket after KB30-50%<20%
Article Helpfulness% yes on "Was this helpful?"80%+<60%
Search Success% searches with article click60%+<40%
Zero Results Rate% searches with no results<10%>20%
Article Coverage% ticket types with KB article80%+<50%
Content Freshness% articles reviewed in 6 months90%+<70%
Time to AnswerAvg time to find answer<2 min>5 min

Good Knowledge Base Structure

✓ Clear information architecture
  → Logical categories that match user mental models
  → Max 3 levels deep
  → Featured/popular articles visible

✓ Search-first design
  → Prominent search bar
  → Synonym support
  → Suggested articles while typing

✓ Scannable content
  → Clear headings and subheadings
  → Bullet points and numbered lists
  → TL;DR at the top

✓ Visual aids
  → Screenshots with annotations
  → Video embeds for complex processes
  → GIFs for quick demonstrations

✓ Actionable troubleshooting
  → Problem → Cause → Solution format
  → Step-by-step instructions
  → "If this doesn't work, try..." fallbacks

✓ Cross-linking
  → Related articles linked
  → Prerequisites linked
  → Next steps suggested

Bad Knowledge Base Design

✗ Buried search
  → Categories-first navigation
  → Users can't find search

✗ Technical jargon
  → Internal terminology
  → No plain language

✗ Wall of text
  → No formatting or structure
  → Users give up reading

✗ Outdated content
  → Screenshots don't match UI
  → Steps no longer work
  → Confusion and distrust

✗ No feedback mechanism
  → Can't report problems
  → Can't request missing content

✗ Desktop-only
  → Not mobile responsive
  → Bad mobile experience

Article Template

# [Task User Wants to Accomplish]

[1-2 sentence summary of what this article covers]

## Quick Answer
[TL;DR - the most common solution in 1-2 sentences]

## Before You Start
- [Prerequisite 1]
- [Prerequisite 2]

## Step-by-Step Instructions

### Step 1: [Action]
[Description of what to do]

[Screenshot with annotation]

### Step 2: [Action]
[Description of what to do]

> **Note:** [Important callout if needed]

### Step 3: [Action]
[Description of what to do]

## Troubleshooting

**Problem:** [Common issue]
**Solution:** [How to fix]

**Problem:** [Another common issue]
**Solution:** [How to fix]

## Related Articles
- [Link to related article 1]
- [Link to related article 2]

## Still Need Help?
[Contact support CTA]

---
*Last updated: [Date] | Was this article helpful? [Yes] [No]*

Content Categories

CategoryPurposeExamples
Getting StartedOnboarding, quick winsSetup, first steps, basics
How-To GuidesTask completionSpecific feature usage
TroubleshootingProblem resolutionError fixes, common issues
ReferenceTechnical detailsAPI docs, configurations
FAQQuick answersCommon questions
Release NotesWhat's newUpdates, changes
Best PracticesOptimal usageTips, recommendations

Deflection Strategies

StrategyImplementationImpact
Pre-ticket searchShow articles before ticket formHigh
Suggested articlesAI-powered recommendationsHigh
In-app helpContext-aware tooltipsHigh
Chatbot firstBot deflection before humanMedium-High
Community redirectPoint to community for how-toMedium
Video tutorialsVisual learners servedMedium

Content Maintenance Workflow

TaskFrequencyOwner
Review helpfulness scoresWeeklyKB Manager
Update flagged articlesWeeklyContent team
Audit for outdated contentMonthlyProduct + Support
Add missing contentOngoingSupport agents flag
Review zero-result searchesWeeklyKB Manager
Archive obsolete articlesQuarterlyProduct + Content

Measuring Self-Service ROI

Ticket Deflection Value:
─────────────────────────
Deflected Tickets = KB Visits × Deflection Rate
Savings = Deflected Tickets × Cost Per Ticket

Example:
10,000 monthly KB visits × 35% deflection = 3,500 deflected tickets
3,500 × $15 per ticket = $52,500/month saved

Knowledge Creation from Support

Ticket PatternActionPriority
10+ tickets on same topic/weekCreate new articleHigh
Repeated escalationsAdd troubleshooting sectionHigh
Complex explanations in ticketsConvert to how-toMedium
Feature questions after releaseAdd feature documentationHigh
Agent finds workaroundDocument in KBMedium

Search Optimization

Good Search Practices:
├── Use customer language, not internal terms
├── Add synonyms and alternate phrasings
├── Include common misspellings
├── Tag with related topics
├── Optimize titles for search
└── Include question-format titles (How do I...?)

Search Analytics to Monitor:
├── Top searches (are we covering these?)
├── Zero-result searches (content gaps)
├── Search refinements (poor initial results)
├── Exit after search (didn't find answer)
└── Ticket creation after search (deflection failure)

Chatbot Integration

ScenarioBot BehaviorFallback
FAQ matchProvide answer"Was this helpful?"
Article matchShare article linkOffer live chat
Low confidence"Let me connect you..."Transfer to agent
After hoursProvide resources"We'll respond tomorrow"
VIP customerSkip bot (optional)Direct to human

Anti-Patterns

  • Write once, never update — Stale content worse than no content
  • Internal voice — Writing for support team, not customers
  • No search analytics — Flying blind on content gaps
  • Support as gatekeepers — Hiding KB behind login
  • Article graveyards — Thousands of articles, none maintained
  • All or nothing chatbot — Bot handles everything or nothing
  • Duplicate content — Same info in multiple places, different versions
  • No escalation path — Self-service dead end, can't reach human

title: Support Metrics (CSAT, FRT, TTR, FCR) impact: HIGH tags: metrics, CSAT, FRT, TTR, FCR, NPS, analytics, performance

Support Metrics (CSAT, FRT, TTR, FCR)

Impact: HIGH

What gets measured gets managed. The right metrics drive the right behaviors. Wrong metrics create perverse incentives. Support metrics should measure customer outcomes, not just agent activity.

Core Metrics Overview

MetricFull NameWhat It MeasuresTarget
CSATCustomer SatisfactionCustomer happiness90%+
FRTFirst Response TimeSpeed to acknowledge<1 hour
TTRTime to ResolutionSpeed to solve<24 hours
FCRFirst Contact ResolutionSolved without escalation70%+
NPSNet Promoter ScoreLoyalty/advocacy30+
CESCustomer Effort ScoreEase of resolution<2 (low effort)

Metric Formulas

CSAT = (Satisfied Responses / Total Responses) × 100
Target: 90%+ | Warning: <85%

FRT = Sum(First Response Times) / Total Tickets
Target: <1 hour | Warning: >4 hours

TTR = Sum(Resolution Times) / Total Resolved Tickets
Target: <24 hours | Warning: >72 hours

FCR = (Tickets Resolved on First Contact / Total Tickets) × 100
Target: 70%+ | Warning: <50%

NPS = % Promoters (9-10) - % Detractors (0-6)
Target: 30+ | Warning: <10

Reopen Rate = (Reopened Tickets / Resolved Tickets) × 100
Target: <5% | Warning: >10%

The Metrics Hierarchy

┌─────────────────────────────────────────────────────────────────┐
│                     BUSINESS OUTCOMES                            │
│  Customer retention, expansion, advocacy, cost efficiency        │
├─────────────────────────────────────────────────────────────────┤
│                     CUSTOMER EXPERIENCE                          │
│  CSAT, NPS, CES, Customer Lifetime Value                         │
├─────────────────────────────────────────────────────────────────┤
│                     OPERATIONAL METRICS                          │
│  FRT, TTR, FCR, SLA compliance, backlog, escalation rate        │
├─────────────────────────────────────────────────────────────────┤
│                     ACTIVITY METRICS                             │
│  Tickets handled, handle time, utilization, replies per ticket  │
└─────────────────────────────────────────────────────────────────┘

Good Metric Practices

✓ Measure outcomes, not just activity
  → CSAT over tickets closed
  → FCR over speed alone
  → Quality over quantity

✓ Context with metrics
  → Segment by ticket type
  → Compare similar complexity
  → Account for team differences

✓ Balanced scorecard
  → Speed + Quality + Volume
  → No single metric obsession
  → Trade-offs acknowledged

✓ Agent-level + Team-level
  → Individual accountability
  → Team collaboration visible
  → Neither in isolation

✓ Regular calibration
  → Define what "good" looks like
  → Review outliers
  → Adjust targets based on data

Bad Metric Practices

✗ Vanity metrics
  → "We closed 1000 tickets!"
  → (But 400 were reopened)

✗ Speed at all costs
  → Racing to respond/close
  → Quality sacrificed

✗ Gaming the system
  → Closing to hit targets
  → Solving easy tickets only
  → Marking pending to pause clock

✗ Metrics without context
  → Comparing unlike tickets
  → Ignoring complexity factors

✗ Punishment-focused
  → Metrics used to punish
  → Fear-based culture
  → Hide problems instead of solve

✗ Single metric obsession
  → Optimize FRT, ignore quality
  → Hit FCR, don't escalate (badly)

CSAT Deep Dive

ScoreMeaningAction
5Very SatisfiedIdentify what worked, replicate
4SatisfiedGood, look for improvement areas
3NeutralInvestigate, follow up
2DissatisfiedImmediate follow-up required
1Very DissatisfiedManager review, service recovery

CSAT Survey Best Practices:

✓ Send within 24 hours of resolution
✓ Keep survey short (1-3 questions)
✓ Include open-ended comment field
✓ Follow up on low scores within 24 hours
✓ Track CSAT by agent, category, complexity

Sample Survey:
1. How satisfied are you with your support experience? (1-5)
2. [Optional] What could we have done better?
3. [Optional] Anything else you'd like to share?

Agent Performance Dashboard

MetricWhat It ShowsReview Frequency
CSAT (individual)Agent qualityWeekly
Tickets handledProductivityDaily
Avg handle timeEfficiencyWeekly
FCR rateResolution skillWeekly
Escalation rateSelf-sufficiencyWeekly
Reopen rateResolution qualityWeekly
QA scoreProcess adherenceWeekly
UtilizationTime on ticketsDaily

Team Performance Dashboard

MetricWhat It ShowsReview Frequency
CSAT (team)Overall qualityDaily
SLA compliancePromise keepingReal-time
BacklogCapacity healthReal-time
Wait timeCustomer experienceReal-time
Escalation volumeL2/L3 loadDaily
Trending issuesProduct problemsDaily
Channel mixResource allocationWeekly

Quality Assurance Scoring

DimensionWeightCriteria
Solution Quality30%Correct, complete, first-time
Communication25%Clear, empathetic, professional
Process Adherence20%Followed procedures, documentation
Efficiency15%Appropriate handle time, no waste
Customer Experience10%Tone, personalization, going extra mile

Sample QA Rubric:

□ Greeting was professional and personalized
□ Issue was correctly understood and restated
□ Troubleshooting was logical and complete
□ Solution was correct and clearly explained
□ Next steps were provided
□ Ticket was properly categorized and documented
□ Tone was empathetic and appropriate
□ Handle time was reasonable for complexity

Reporting Cadence

ReportMetrics IncludedAudienceFrequency
Agent DashboardPersonal statsAgentsReal-time
Team StandupDaily snapshotTeam leadsDaily
Weekly ReviewTrends, outliersManagersWeekly
Monthly Business ReviewStrategic metricsDirectorsMonthly
Quarterly Board ReportCost, CSAT, NPSExecutivesQuarterly

Benchmarks by Company Stage

StageCSATFRTTTRFCR
Startup85%+<4 hours<48 hours60%+
Growth88%+<2 hours<24 hours65%+
Scale90%+<1 hour<12 hours70%+
Enterprise92%+<30 min<8 hours75%+

Diagnosing Metric Problems

SymptomPossible CausesInvestigation
Low CSATWrong solutions, bad communicationQA reviews, comment analysis
High FRTUnderstaffed, poor routingCapacity analysis, queue review
Low FCRKnowledge gaps, over-escalationTraining needs, KB gaps
High reopensPremature closure, wrong solutionsQA reviews, root cause analysis
High escalationL1 undertrained, complexity increaseSkill assessment, ticket analysis

Metric Improvement Playbook

CSAT Improvement:
1. Analyze low-score tickets for patterns
2. Implement service recovery for 1-2 scores
3. Train on communication and empathy
4. Add quality checkpoints before closure
5. Recognize and share high-CSAT behaviors

FCR Improvement:
1. Identify common escalation reasons
2. Create KB articles for top escalations
3. Train L1 on frequently escalated topics
4. Empower L1 with more resolution authority
5. Add decision trees for complex issues

Anti-Patterns

  • CSAT gaming — Only surveying easy tickets
  • FRT racing — Copy-paste first response to hit metric
  • Premature closure — Closing to hit TTR, causing reopens
  • FCR inflation — Not escalating when you should
  • Metric of the month — Constantly changing focus
  • Comparison without context — Ignoring ticket complexity
  • Activity over outcomes — Celebrating tickets closed, not problems solved
  • Individual over team — Missing collaborative behaviors

title: Quality Assurance & Agent Coaching impact: HIGH tags: quality, QA, coaching, training, performance, reviews

Quality Assurance & Agent Coaching

Impact: HIGH

Quality assurance ensures consistent, excellent customer experiences. Without QA, agent performance varies wildly and issues go undetected. Done right, QA is a coaching tool that elevates the entire team.

The QA Framework

┌─────────────────────────────────────────────────────────────────┐
│                    QUALITY MEASUREMENT                           │
│  Ticket reviews, score tracking, trend analysis                 │
└──────────────────────────┬──────────────────────────────────────┘
                           │
           ┌───────────────┼───────────────┐
           ▼               ▼               ▼
    ┌─────────────┐ ┌─────────────┐ ┌─────────────┐
    │  INDIVIDUAL │ │    TEAM     │ │   PROCESS   │
    │   COACHING  │ │   TRENDS    │ │ IMPROVEMENT │
    │             │ │             │ │             │
    │ 1:1 feedback│ │ Common gaps │ │ KB updates  │
    │ Training    │ │ Best sharing│ │ Tool fixes  │
    │ Development │ │ Calibration │ │ Workflow    │
    └─────────────┘ └─────────────┘ └─────────────┘
                           │
                           ▼
              ┌───────────────────────┐
              │  IMPROVED CUSTOMER    │
              │     EXPERIENCE        │
              └───────────────────────┘

QA Scorecard Dimensions

DimensionWeightWhat It Measures
Solution Quality30%Was the answer correct and complete?
Communication25%Was the message clear, professional, empathetic?
Process Adherence20%Were procedures and documentation followed?
Efficiency15%Was handle time appropriate for complexity?
Customer Experience10%Did the interaction feel personalized and helpful?

Detailed Scoring Rubric

Solution Quality (30 points)

30: Perfect - Correct, complete, first-time resolution
25: Good - Correct, minor details missing
20: Adequate - Mostly correct, needed follow-up
15: Needs Work - Partially correct, customer confused
10: Poor - Wrong solution, customer frustrated
0:  Fail - Critical error, escalation required

Communication (25 points)

25: Excellent - Clear, warm, professional, personalized
20: Good - Clear and professional, minimal personalization
15: Adequate - Understandable, somewhat impersonal
10: Needs Work - Confusing or too casual/formal
5:  Poor - Unclear, rude, or unprofessional
0:  Fail - Inappropriate, offensive, or damaging

Process Adherence (20 points)

20: Perfect - All procedures followed, complete documentation
15: Good - Minor documentation gaps
10: Adequate - Some steps skipped, functional documentation
5:  Needs Work - Multiple steps skipped
0:  Fail - Critical process violation

Efficiency (15 points)

15: Excellent - Optimal time for complexity level
12: Good - Slightly over/under optimal time
9:  Adequate - Noticeably inefficient but reasonable
6:  Needs Work - Significant time waste
3:  Poor - Extreme inefficiency or rushing
0:  Fail - SLA breach due to handling

Customer Experience (10 points)

10: Delightful - Went above and beyond
8:  Good - Met expectations, professional
6:  Adequate - Functional, impersonal
4:  Needs Work - Customer had to work hard
2:  Poor - Frustrating experience
0:  Fail - Customer explicitly complained

Good QA Practices

✓ Random sampling
  → Representative ticket mix
  → Multiple categories reviewed
  → Both good and problematic tickets

✓ Calibrated reviewers
  → Regular calibration sessions
  → Same ticket, compare scores
  → Align on standards

✓ Timely feedback
  → Reviews within 48 hours
  → Coaching within a week
  → Real-time flags for critical issues

✓ Growth-focused
  → Celebrate improvements
  → Actionable feedback
  → Development plans

✓ Data-driven
  → Track trends over time
  → Identify systemic issues
  → Measure improvement

Bad QA Practices

✗ Gotcha culture
  → Looking for mistakes to punish
  → Agents fear QA

✗ Inconsistent standards
  → Different reviewers, different scores
  → No calibration

✗ Delayed feedback
  → Reviews from months ago
  → Issues already repeated

✗ Score obsession
  → Number without context
  → No coaching conversation

✗ Cherry-picking
  → Only reviewing easy tickets
  → Missing complex cases

✗ No positive feedback
  → Only hear about problems
  → Demotivating

QA Review Workflow

Weekly QA Process:
┌─────────────────────────────────────────────────────────────────┐
│ Day 1-2: Ticket Selection                                        │
│  □ Pull random sample (3-5 tickets per agent)                   │
│  □ Include mix of categories and complexity                     │
│  □ Add any flagged tickets from previous week                   │
├─────────────────────────────────────────────────────────────────┤
│ Day 3-4: Review and Score                                        │
│  □ Apply scorecard to each ticket                               │
│  □ Document specific feedback points                            │
│  □ Flag exceptional (good or bad) examples                      │
├─────────────────────────────────────────────────────────────────┤
│ Day 5: Calibration and Coaching                                  │
│  □ Calibration meeting with reviewers                           │
│  □ Share scores and feedback with agents                        │
│  □ Schedule 1:1 coaching for low scores                         │
└─────────────────────────────────────────────────────────────────┘

Sample Size Guidelines

Team SizeTickets/Agent/WeekTotal Monthly
1-5 agents5 tickets100+ tickets
6-15 agents3-4 tickets200-300 tickets
16-30 agents2-3 tickets300-400 tickets
30+ agents2 ticketsVaries

Calibration Session Format

Monthly Calibration (1 hour):
1. Pre-work: All reviewers score same 3 tickets
2. Compare scores and discuss differences
3. Align on correct interpretation of rubric
4. Update guidelines if needed
5. Document decisions for reference

Calibration Metrics:
- Score variance should be <5 points
- Major discrepancies discussed until aligned
- Track calibration quality over time

Coaching Conversation Template

## 1:1 Coaching: [Agent Name]

**Date:** [Date]
**Review Period:** [Dates]
**Average QA Score:** [X/100]

### Strengths (Lead with positives)
- [Specific strength with example]
- [Another strength with example]

### Ticket Review
Let's look at [Ticket #X] together...
- What went well: [Specific feedback]
- Opportunity: [Specific feedback]

### Development Focus
This week, let's focus on: [One specific area]
- Why it matters: [Impact explanation]
- How to improve: [Specific actions]
- Resources: [Training, shadowing, documentation]

### Goals for Next Week
1. [Specific, measurable goal]
2. [Specific, measurable goal]

### Follow-up
- Next 1:1: [Date]
- Check-in on: [Specific item]

Common Coaching Scenarios

IssueCoaching Approach
Technical gapsPair with specialist, assign training
Communication issuesReview examples, practice rewrites
Process non-complianceExplain why, shadow high performers
Efficiency problemsTime management tips, tool training
Empathy lackingRole-play exercises, customer stories
Over-escalationDecision trees, confidence building
Under-escalationClear escalation criteria, safety net

Performance Improvement Plan

## Performance Improvement Plan: [Agent Name]

**Start Date:** [Date]
**Review Date:** [30/60/90 days out]
**Manager:** [Name]

### Current Performance
- QA Score: [X/100] (vs team avg [Y/100])
- Specific concerns:
  - [Issue 1]
  - [Issue 2]

### Performance Standards Required
- QA Score: [Target, e.g., 80+]
- CSAT: [Target]
- Handle time: [Target]

### Support Provided
- Weekly 1:1 coaching
- Buddy system with [Name]
- Additional training: [Specific courses]
- Reduced ticket load for [X weeks]

### Milestones
- Week 2: [Milestone and measure]
- Week 4: [Milestone and measure]
- Week 6: [Milestone and measure]
- Week 8: [Final review]

### Success Criteria
[Clear definition of what success looks like]

### Signatures
Agent: ________________  Date: _______
Manager: _______________  Date: _______

QA Metrics Dashboard

MetricDefinitionTarget
Avg QA ScoreTeam average score85+
Score Distribution% in each bandNormal curve
Calibration VarianceReviewer agreement<5 points
Review Coverage% agents reviewed100% weekly
Critical FailsFailed tickets<2%
Improvement RateScore change over timePositive trend
QA-CSAT CorrelationDo high QA scores = high CSAT?Strong positive

Recognition Program

Excellence Recognition:
├── Weekly: "Ticket of the Week" - shared in team meeting
├── Monthly: Top performer award - tangible recognition
├── Quarterly: Quality champion - career development opportunity

Peer Recognition:
├── Shout-outs in Slack
├── Best practice sharing
├── Mentorship opportunities

Anti-Patterns

  • Punitive QA — Used for discipline, not development
  • Score inflation — Everyone gets 90+, no differentiation
  • Random rubric — Different scores for same quality
  • Feedback void — Scores given, never discussed
  • Perfectionism — Only 100% acceptable
  • Recency bias — Only recent tickets reviewed
  • Volume over quality — Too many reviews, shallow feedback
  • Ignoring top performers — Only coaching struggling agents

title: SLA Design & Compliance impact: CRITICAL tags: sla, service-level, compliance, response-time, resolution

SLA Design & Compliance

Impact: CRITICAL

Service Level Agreements define customer expectations and team accountability. Well-designed SLAs balance customer needs with operational capacity. Poorly designed SLAs create false promises, stressed agents, and dissatisfied customers.

SLA Components

ComponentDefinitionExample
First Response TimeTime to initial acknowledgment"We'll respond within 1 hour"
Resolution TimeTime to issue closure"We'll resolve within 24 hours"
Update FrequencyTime between status updates"Updates every 4 hours until resolved"
Business HoursWhen SLA clock runs"9am-6pm PST, Mon-Fri"
Escalation TriggerWhen to escalate automatically"If unresolved after 80% of SLA"

SLA by Priority Level

PriorityFirst ResponseResolutionUpdate Frequency
P1 Critical15 minutes4 hoursEvery 30 minutes
P2 High1 hour8 hoursEvery 2 hours
P3 Medium4 hours24 hoursEvery 8 hours
P4 Low8 hours72 hoursEvery 24 hours
P5 Request24 hours5 business daysAt status change

SLA by Customer Tier

TierMultiplierP1 ResponseP1 Resolution
Enterprise0.5x (faster)7 minutes2 hours
Business1x (standard)15 minutes4 hours
Professional1.5x30 minutes6 hours
Starter2x1 hour8 hours
Free4x4 hoursNext business day

Good SLA Design

✓ Realistic targets
  → Based on historical data and capacity
  → Buffer for complexity variation
  → Achievable 95% of the time

✓ Clear definitions
  → What counts as "response"?
  → What counts as "resolution"?
  → When does clock start/stop?

✓ Business hours clarity
  → Defined time zones
  → Holiday calendars
  → 24/7 for critical only

✓ Escalation automation
  → Warning at 80% of SLA
  → Auto-escalate at breach
  → Visibility to managers

✓ Regular review
  → Monthly SLA performance analysis
  → Adjust based on capacity
  → Customer feedback incorporation

Bad SLA Design

✗ Promise more than you can deliver
  → 15-minute response with 2 agents
  → Creates breach patterns

✗ Same SLA for all priorities
  → P4 tickets block P1 attention
  → Resources misallocated

✗ Vague definitions
  → "We'll get back to you soon"
  → No measurable commitment

✗ 24/7 claims without 24/7 staff
  → SLA breaches overnight/weekends
  → Angry customers Monday morning

✗ No pause mechanism
  → Clock runs while waiting on customer
  → Unfair agent metrics

✗ SLA as punishment
  → Metrics used to penalize, not improve
  → Gaming behavior emerges

SLA Clock Rules

EventClock Behavior
Ticket createdClock starts
Agent respondsResponse SLA met
Status → Pending CustomerClock pauses
Customer respondsClock resumes
Status → ResolvedResolution SLA met
Ticket reopenedNew SLA period begins
Business hours endClock pauses (if applicable)

SLA Breach Prevention

Warning LevelTriggerAction
Yellow50% of SLA elapsedVisible in queue view
Orange80% of SLA elapsedAlert to assignee
Red90% of SLA elapsedAlert to team lead
Breach100% elapsedManager notification, escalation
Severe Breach2x SLA elapsedDirector notification

SLA Compliance Dashboard

MetricCalculationTarget
Response SLA %Met / Total tickets95%+
Resolution SLA %Met / Total tickets90%+
Avg Time to First ResponseSum(response times) / count<50% of SLA
Avg Resolution TimeSum(resolution times) / count<75% of SLA
Breach CountTickets past SLATrending down
Severe Breach CountTickets 2x+ past SLAZero

SLA Reporting Cadence

ReportFrequencyAudience
Real-time dashboardLiveAgents, leads
Daily summaryDailyManagers
Weekly analysisWeeklyDirectors
Monthly reviewMonthlyVP, execs
Quarterly trendsQuarterlyC-level, board

SLA Exception Handling

ExceptionHow to Handle
Customer unresponsivePause clock after 2 follow-ups
Third-party dependencyDocument, may pause clock
Product bug requires fixTrack separately, link to bug
Out of scope requestClose with explanation, no breach
Duplicate ticketMerge, use original SLA
Customer requests pausePause with documentation

SLA Template

## [Company Name] Support SLA

### Coverage Hours
Support is available Monday-Friday, 9:00 AM - 6:00 PM [Timezone].
Critical (P1) issues receive 24/7 coverage.

### Response Time Commitments

| Priority | Description | First Response | Resolution Target |
|----------|-------------|----------------|-------------------|
| Critical | System outage | 15 minutes | 4 hours |
| High | Major feature broken | 1 hour | 8 hours |
| Medium | Feature impaired | 4 hours | 24 hours |
| Low | Minor issue | 8 hours | 72 hours |

### Definitions
- **First Response**: Initial acknowledgment from support agent
- **Resolution**: Issue resolved or workaround provided
- **Business Hours**: Clock pauses outside coverage hours (except P1)

### Escalation
Tickets approaching SLA are automatically escalated to team leads.
SLA breaches are reported to management for review.

### Exclusions
SLA does not apply to: feature requests, out-of-scope issues,
or delays caused by customer unresponsiveness.

SLA Implementation Checklist

Design Phase:
□ Analyze historical ticket data
□ Calculate current response/resolution times
□ Set targets based on capacity + 20% buffer
□ Define clear priority criteria
□ Document business hours and holidays
□ Create customer tier matrix

Configuration Phase:
□ Set up SLA policies in help desk
□ Configure clock pause rules
□ Create escalation workflows
□ Build breach notifications
□ Set up reporting dashboards

Launch Phase:
□ Train agents on SLA expectations
□ Communicate SLA to customers
□ Monitor closely for first 2 weeks
□ Adjust unrealistic targets
□ Celebrate compliance wins

Anti-Patterns

  • Vanity SLAs — Impressive numbers no one can meet
  • Clock manipulation — Pausing for illegitimate reasons
  • Resolution without resolution — Closing tickets to hit SLA
  • One SLA fits all — Same targets for all ticket types
  • SLA without capacity — Promises without staffing to deliver
  • Breach normalization — Accepting regular breaches as normal
  • Customer tier opacity — Customers don't know their SLA level

title: Ticket Management & Prioritization impact: CRITICAL tags: tickets, prioritization, queue, triage, workflow

Ticket Management & Prioritization

Impact: CRITICAL

Effective ticket management is the foundation of support operations. Without proper prioritization and queue management, critical issues get buried, customers wait too long, and agents waste time on low-impact work.

The Ticket Prioritization Matrix

PriorityImpactUrgencyAction
P1Many users or revenueComplete blockImmediate all-hands
P2Multiple usersSignificant blockNext available agent
P3Single userPartial blockQueue position
P4Single userMinor inconvenienceLow priority queue
P5EnhancementFuture improvementBacklog

Triage Decision Tree

New Ticket Received
        │
        ▼
Is it a system-wide outage?
    YES → P1 Critical → Immediate escalation
    NO  ↓
        ▼
Is revenue or data at risk?
    YES → P2 High → Priority queue
    NO  ↓
        ▼
Is core functionality blocked?
    YES → P3 Medium → Standard queue
    NO  ↓
        ▼
Is it a how-to or feature request?
    YES → P5 Request → Knowledge base + backlog
    NO  → P4 Low → Low priority queue

Good Ticket Prioritization

✓ Customer impact determines priority
  → Multiple users affected = higher priority
  → Revenue impact = escalated path
  → Severity based on business impact, not loudness

✓ Clear priority definitions
  → Written criteria everyone understands
  → Examples for each level
  → Edge case guidance

✓ Dynamic reprioritization
  → Tickets age into higher priority
  → Customer tier adjusts SLA
  → Pattern detection for trending issues

✓ Fair queue management
  → First-in-first-out within priority
  → No cherry-picking easy tickets
  → Visibility into all queues

✓ Smart auto-assignment
  → Skill-based routing
  → Workload balancing
  → Customer-agent continuity when helpful

Bad Ticket Prioritization

✗ Loudest customer wins
  → Squeaky wheel gets attention
  → Strategic customers ignored

✗ All tickets are P1
  → When everything's urgent, nothing is
  → True emergencies get buried

✗ Manual triage bottleneck
  → Single person assigns all tickets
  → Delays before work begins

✗ Cherry-picking
  → Agents grab easy tickets first
  → Complex issues age indefinitely

✗ No aging escalation
  → Old tickets stay low priority
  → SLA breaches pile up

✗ Priority by customer emotion
  → Angry ≠ urgent
  → Business impact should drive priority

Queue Health Metrics

MetricDefinitionTargetWarning
Backlog AgeAvg days tickets in queue<1 day>3 days
Queue DepthTotal tickets waitingManageable3x normal
Priority Distribution% in each priorityP1<5%, P2<15%P1>10%
Wait TimeTime until first responsePer SLAApproaching SLA
Agent UtilizationTime on tickets vs available70-80%<60% or >90%

Ticket Assignment Models

ModelHow It WorksBest For
Round-robinNext agent in rotationEqual skill levels
Skill-basedMatch ticket type to expertiseSpecialized teams
Load-balancedRoute to lowest backlogUneven workloads
Customer-ownedSame agent for accountRelationship-based
SwarmingTeam collaborates on ticketComplex issues

Ticket Field Requirements

FieldRequired?Purpose
PriorityYesSLA and queue position
CategoryYesRouting and reporting
CustomerYesContext and SLA tier
Product/FeatureYesSpecialist routing
ChannelAutoResponse expectations
AssigneeAuto/YesAccountability
StatusYesWorkflow tracking

Automation Opportunities

TriggerActionBenefit
Keyword in subjectAuto-categorizeFaster triage
Customer tier lookupSet priority modifierSLA compliance
Similar recent ticketSuggest solutionFaster resolution
Idle for X hoursAlert or reassignPrevent aging
Sentiment detected angryFlag for supervisorProactive save
Known issue matchLink to parent ticketAvoid duplicates

Ticket Tagging Strategy

Good Tags:
├── Product area (auth, billing, integrations)
├── Issue type (bug, how-to, feature-request)
├── Root cause (user-error, product-bug, documentation)
├── Outcome (resolved, known-issue, feature-logged)
└── Customer segment (enterprise, startup, trial)

Bad Tags:
├── Agent names
├── One-time-use tags
├── Overlapping meanings
├── Free-form text
└── Too granular (hundreds of rarely-used tags)

Queue Management Checklist

Daily Queue Review:
□ Check P1/P2 tickets first
□ Review aged tickets (>24 hours)
□ Identify trending issues
□ Reassign from OOO agents
□ Balance workload across team

Weekly Queue Review:
□ Analyze backlog trends
□ Review priority accuracy
□ Check for stuck tickets
□ Identify training needs
□ Update automation rules

Anti-Patterns

  • Priority inflation — Everything marked P1 to get attention faster
  • Queue hiding — Tickets moved to personal views, invisible to team
  • Status limbo — Tickets sit "in progress" without updates
  • Duplicate sprawl — Same issue creates multiple tickets
  • Tag chaos — Inconsistent or excessive tagging
  • Assignment hoarding — Agents claim more than they can handle
  • Triage delay — New tickets sit unassigned too long

title: Support Tier Structure (L1/L2/L3) impact: CRITICAL tags: tier, escalation, L1, L2, L3, skill-routing, specialization

Support Tier Structure (L1/L2/L3)

Impact: CRITICAL

A well-designed tier structure ensures tickets reach the right expertise level efficiently. Too flat and specialists drown in simple questions. Too layered and customers wait through unnecessary handoffs.

The Three-Tier Model

┌─────────────────────────────────────────────────────────────────┐
│                          TIER 3 (L3)                             │
│  ┌─────────────────────────────────────────────────────────────┐ │
│  │ Engineering, product, security specialists                  │ │
│  │ Code-level debugging, custom development, architecture     │ │
│  │ Target: <5% of tickets                                     │ │
│  └─────────────────────────────────────────────────────────────┘ │
├─────────────────────────────────────────────────────────────────┤
│                          TIER 2 (L2)                             │
│  ┌─────────────────────────────────────────────────────────────┐ │
│  │ Technical specialists, senior support engineers            │ │
│  │ Complex troubleshooting, integrations, edge cases          │ │
│  │ Target: 15-25% of tickets                                  │ │
│  └─────────────────────────────────────────────────────────────┘ │
├─────────────────────────────────────────────────────────────────┤
│                          TIER 1 (L1)                             │
│  ┌─────────────────────────────────────────────────────────────┐ │
│  │ Frontline support agents, customer service                 │ │
│  │ Common issues, how-to, documentation, basic troubleshooting│ │
│  │ Target: 60-80% first-contact resolution                    │ │
│  └─────────────────────────────────────────────────────────────┘ │
├─────────────────────────────────────────────────────────────────┤
│                        TIER 0 (Self-Service)                     │
│  ┌─────────────────────────────────────────────────────────────┐ │
│  │ Knowledge base, chatbots, community, in-app help           │ │
│  │ Instant answers, guided troubleshooting, FAQs              │ │
│  │ Target: 30-50% deflection rate                             │ │
│  └─────────────────────────────────────────────────────────────┘ │
└─────────────────────────────────────────────────────────────────┘

Tier Responsibilities

TierHandlesSkills RequiredEscalates When
L0FAQ, how-to, guided troubleshootingContent, UXNeeds human judgment
L1Account issues, common bugs, setup helpProduct knowledge, communicationRequires technical investigation
L2Complex bugs, integrations, configurationsTechnical depth, debuggingRequires code changes
L3Product bugs, architecture, securityEngineering, developmentCannot be fixed (limitation)

Tier Metrics by Level

MetricL1 TargetL2 TargetL3 Target
First Contact Resolution70-80%N/AN/A
Avg Handle Time10-15 min30-60 minHours-days
Tickets/Agent/Day30-5010-203-8
Escalation Rate15-25%5-10%N/A
CSAT90%+90%+85%+
Technical Accuracy95%+98%+99%+

Good Tier Design

✓ Clear escalation criteria
  → Documented triggers for each tier transition
  → No ambiguity about who handles what
  → Decision trees for common scenarios

✓ Skill-based routing
  → Tickets auto-route to right tier
  → Specialists handle specialty topics
  → Generalists handle breadth

✓ Information flows with ticket
  → Full context passes on escalation
  → Customer doesn't re-explain
  → Prior troubleshooting documented

✓ Feedback loops
  → L2/L3 trains L1 on patterns
  → Common escalations become L1 capable
  → Knowledge base grows from escalations

✓ Career progression
  → L1 → L2 → L3 path defined
  → Skills assessment for promotion
  → Training programs at each level

Bad Tier Design

✗ L1 as gatekeepers only
  → No resolution authority
  → Customers forced through hoops

✗ Escalation = handoff
  → Original agent disappears
  → Context lost, customer frustrated

✗ All-or-nothing expertise
  → No middle ground specialists
  → L1 overwhelmed or L3 drowning

✗ Tier hopping
  → L2 bounces back to L1
  → L1 escalates without trying
  → Tickets ping-pong

✗ No escalation documentation
  → Agent intuition only
  → Inconsistent customer experience

✗ L3 as ticket graveyard
  → Engineering never responds
  → Tickets age indefinitely

Escalation Triggers

FromToTrigger Conditions
L0L1Customer requests human, confidence low, sensitive issue
L1L2Requires logs/debugging, integration issue, 2+ failed attempts
L2L3Confirmed bug, requires code fix, security concern
AnyManagerCustomer escalation request, VIP customer, legal/compliance

Escalation Handoff Template

## Escalation Summary

**Ticket ID:** [ID]
**Customer:** [Name] | [Company] | [Tier]
**Original Agent:** [Name]
**Escalation Time:** [Timestamp]

### Issue Summary
[2-3 sentences describing the customer's problem]

### Troubleshooting Completed
- [Step 1] - [Result]
- [Step 2] - [Result]
- [Step 3] - [Result]

### Relevant Information
- Product version: [X.X.X]
- Environment: [Production/Staging]
- Error logs: [Link or attached]
- Screenshots: [Link or attached]

### Why Escalating
[Clear explanation of why L1/L2 cannot resolve]

### Customer Expectation
[What customer expects and timeline discussed]

### Suggested Next Steps
[Agent's hypothesis or recommendation for L2/L3]

Skill-Based Routing Matrix

Ticket CategoryPrimary SkillTierFallback
Billing/AccountAccount ManagementL1Finance
Password/LoginAuthenticationL1L2 Security
API ErrorsAPI IntegrationL2L3 Engineering
PerformanceInfrastructureL2L3 DevOps
Data IssuesData EngineeringL2L3 Database
SecuritySecurityL2L3 Security
Mobile AppMobileL2L3 Mobile Dev
Enterprise SSOIdentityL2L3 Enterprise

Tier Staffing Ratios

Company StageL1 : L2 : L3
Early Stage5 : 2 : 1
Growth8 : 3 : 1
Scale10 : 3 : 1
Enterprise Focus6 : 4 : 2

L3 Engineering Integration

ModelDescriptionBest For
Dedicated Support EngEngineers only do supportHigh-touch, complex products
RotationEngineers rotate into supportShared ownership, empathy
On-callEngineers available for escalationLow volume L3 needs
Bug QueueSupport files bugs, eng triagesStandard products

Career Progression Framework

L1 Support Agent (0-12 months)
├── Product knowledge certification
├── Communication skills
├── Basic troubleshooting
└── Ready for L2: 70% FCR, consistent CSAT, technical curiosity

L2 Support Specialist (1-3 years)
├── Advanced technical training
├── Debugging and log analysis
├── Integration expertise
└── Ready for L3: Complex case leadership, mentoring L1

L3 Support Engineer (3+ years)
├── Engineering collaboration
├── Code-level debugging
├── Product feedback influence
└── Paths: Engineering, Support Management, Solutions Architect

Anti-Patterns

  • Escalation as avoidance — L1 escalates instead of learning
  • L3 bottleneck — Engineering never prioritizes support tickets
  • Context-free handoffs — "Here's a ticket" without details
  • Tier rigidity — Cannot skip tiers even when obvious
  • No cross-training — Tiers operate in silos
  • Specialist overload — One person handles all of a specialty
  • Escalation shame — Agents afraid to escalate, take too long

title: Support Tool Stack Optimization impact: MEDIUM-HIGH tags: tooling, help-desk, zendesk, intercom, freshdesk, integrations, automation

Support Tool Stack Optimization

Impact: MEDIUM-HIGH

The right tool stack amplifies team effectiveness. The wrong stack creates friction, workarounds, and data silos. Optimize for agent efficiency, customer experience, and operational insights.

The Support Technology Stack

┌─────────────────────────────────────────────────────────────────┐
│                     ANALYTICS & BI LAYER                         │
│  Looker, Tableau, Mode, Metabase, native reporting               │
├─────────────────────────────────────────────────────────────────┤
│                      HELP DESK CORE                              │
│  Zendesk, Intercom, Freshdesk, HubSpot Service, Salesforce      │
├─────────┬─────────────┬──────────────┬─────────────┬────────────┤
│ CHAT    │ PHONE       │ KNOWLEDGE    │ AUTOMATION  │ QUALITY    │
│ Intercom│ Aircall     │ Guru         │ Zapier      │ MaestroQA  │
│ Drift   │ Dialpad     │ Notion       │ Workato     │ Klaus      │
│ Crisp   │ Talkdesk    │ Confluence   │ Tray.io     │ Playvox    │
├─────────┴─────────────┴──────────────┴─────────────┴────────────┤
│                    CUSTOMER DATA LAYER                           │
│  CRM (Salesforce, HubSpot), CDP, Product Analytics              │
├─────────────────────────────────────────────────────────────────┤
│                    COMMUNICATION LAYER                           │
│  Email (SendGrid, Postmark), SMS (Twilio), In-app               │
└─────────────────────────────────────────────────────────────────┘

Tool Selection Criteria

CriteriaWeightQuestions to Ask
Agent Experience25%How many clicks to resolve common issues?
Customer Experience25%Seamless channel switching? Context preserved?
Scalability20%Handle 10x volume without breaking?
Integration15%Works with existing stack? Open API?
Reporting10%Native analytics? Custom reports?
Total Cost5%Price per agent/ticket? Hidden costs?

Help Desk Platform Comparison

FeatureZendeskIntercomFreshdeskHubSpot Service
Ticket ManagementExcellentGoodExcellentGood
Live ChatGoodExcellentGoodGood
Knowledge BaseExcellentGoodGoodGood
AutomationExcellentExcellentGoodGood
ReportingExcellentGoodGoodExcellent
IntegrationsExcellentGoodGoodExcellent (HubSpot)
Pricing$$$$$$$$$$
Best ForLarge teamsProduct-ledSMBHubSpot users

Essential Integrations

IntegrationPurposePriority
CRMCustomer context, account infoCritical
Product/AppUser actions, account statusCritical
BillingPayment status, subscription infoHigh
IdentitySSO, user verificationHigh
SlackInternal escalations, notificationsHigh
Engineering toolsBug tracking (Jira, Linear)High
Knowledge baseAnswer suggestions, deflectionMedium
PhoneCall handling, recordingMedium
CalendarMeeting schedulingLow

Good Tool Stack Design

✓ Single pane of glass
  → All customer context visible in one view
  → No tab-switching to find information
  → CRM + Product + Billing integrated

✓ Automation-first
  → Repetitive tasks automated
  → Routing rules, not manual assignment
  → Macros for common responses

✓ Channel unification
  → Email, chat, phone in one queue
  → Customer history across channels
  → Seamless channel switching

✓ Scalable foundation
  → Can handle 10x volume
  → Easy to add new agents
  → Performance doesn't degrade

✓ Data accessibility
  → Easy reporting and exports
  → API access for custom needs
  → Real-time dashboards

Bad Tool Stack Design

✗ Tool sprawl
  → 10 tools for 5 functions
  → Constant context switching
  → Data in silos

✗ Manual everything
  → Copy-paste between systems
  → No automation rules
  → Human routing all tickets

✗ Channel silos
  → Separate tools for email vs chat
  → No unified customer history
  → Agent retraining per channel

✗ Integration nightmares
  → Custom code breaking
  → No native connectors
  → IT bottleneck for changes

✗ Reporting gaps
  → Can't answer basic questions
  → Excel exports for everything
  → No real-time visibility

Automation Opportunities

AutomationTriggerActionImpact
Auto-routingTicket createdAssign based on skill/categoryFRT reduction
Auto-responseTicket createdAcknowledgment + KB suggestionCustomer experience
SLA alertsApproaching breachNotify assignee + managerSLA compliance
EscalationTime elapsedAuto-escalate to next tierResolution speed
CSAT surveyTicket resolvedSend satisfaction surveyFeedback collection
Duplicate detectionSimilar ticket existsLink and notifyEfficiency
VIP detectionHigh-value customerPriority flag + routingCustomer retention

Workflow Automation Examples

Auto-Categorization:

IF subject contains "password" OR "login" OR "can't access"
THEN set category = "Authentication"
AND route to "Authentication Team"
AND suggest KB article "Password Reset Guide"

Escalation Automation:

IF priority = "High"
AND status = "Open"
AND hours since created > 2
THEN send Slack alert to #support-escalations
AND add internal note "Approaching SLA breach"
AND assign to on-call manager

CSAT Trigger:

IF status changed to "Resolved"
AND channel != "Internal"
AND not tagged "no-survey"
THEN wait 2 hours
AND send CSAT survey

Macro/Template Best Practices

CategoryExample MacrosUsage Tips
GreetingsInitial response templatesPersonalize opening
TroubleshootingStep-by-step guidesAdjust for context
EscalationHandoff templatesInclude full context
ResolutionClosing templatesSummarize solution
DelaysStatus update templatesSet expectations
RefundsBilling response templatesFollow policy

Agent Workspace Optimization

Ideal Agent View:
┌─────────────────────────────────────────────────────────────────┐
│  Customer Info          │  Ticket Details                       │
│  ─────────────────────  │  ───────────────────────────────────  │
│  Name: John Smith       │  Subject: Cannot export data          │
│  Company: Acme Inc      │  Priority: Medium                     │
│  Plan: Enterprise       │  Status: Open                         │
│  MRR: $5,000           │  Assigned: You                        │
│  Health: At Risk        │  Channel: Email                       │
│  CSM: Jane Doe          │                                       │
├─────────────────────────┤                                       │
│  Recent Tickets (3)     │  Conversation                         │
│  - Export bug (open)    │  ───────────────────────────────────  │
│  - SSO setup (closed)   │  [Customer message]                   │
│  - Pricing Q (closed)   │                                       │
├─────────────────────────┤  [Reply box]                          │
│  Account Activity       │                                       │
│  - Last login: Today    │  Suggested Articles:                  │
│  - Feature use: High    │  - Data Export Guide                  │
│  - NPS: 8 (Passive)     │  - Troubleshooting Exports            │
└─────────────────────────┴───────────────────────────────────────┘

Tool Implementation Checklist

Pre-Implementation:
□ Define requirements and success metrics
□ Evaluate 3+ options with demos
□ Check integration compatibility
□ Calculate total cost of ownership
□ Plan data migration strategy

Implementation:
□ Configure workflows and automations
□ Set up integrations
□ Import historical data
□ Create macros and templates
□ Build reporting dashboards

Training & Launch:
□ Train agents on new workflows
□ Document processes and edge cases
□ Soft launch with subset of tickets
□ Monitor for issues
□ Iterate based on feedback

Post-Launch:
□ Review automation effectiveness
□ Optimize based on usage data
□ Gather agent feedback
□ Regular workflow audits
□ Plan for scale

Cost Optimization

StrategySavings PotentialTrade-offs
Right-size licenses10-20%Audit required
Negotiate annually15-25%Cash flow impact
Automate deflection20-40%Upfront investment
Consolidate tools10-30%Migration effort
Usage-based pricingVariableUnpredictable

Anti-Patterns

  • Shiny object syndrome — Switching tools for new features
  • Over-customization — Too complex to maintain
  • Under-utilization — Paying for features not used
  • Integration debt — Custom code no one understands
  • Vendor lock-in — Can't migrate data out
  • Agent workarounds — Using Excel because tool doesn't work
  • No governance — Everyone creates their own automations
  • Metric blindness — Can't measure what matters

title: Workforce Management & Capacity Planning impact: MEDIUM-HIGH tags: workforce, capacity, scheduling, forecasting, staffing, utilization

Workforce Management & Capacity Planning

Impact: MEDIUM-HIGH

Right-sizing your support team is a constant balancing act. Too few agents and customers wait too long. Too many and you're burning budget. Workforce management turns ticket volume into staffing decisions.

The Capacity Planning Equation

                    Ticket Volume × Handle Time
Required Agents = ─────────────────────────────────────
                   Available Hours × Utilization Rate

Example:
    500 tickets/day × 20 min/ticket = 10,000 minutes
    8 hours × 60 min × 75% utilization = 360 min/agent/day
    10,000 ÷ 360 = 28 agents needed

Capacity Variables

VariableDefinitionHow to Get
Ticket VolumeTickets per time periodHistorical data, forecasting
Handle TimeAvg time to resolveTicket analytics
Available HoursHours worked per agentSchedules, less meetings
Utilization Rate% time on ticketsTarget 70-80%
ShrinkageNon-productive timeTraining, breaks, meetings

Utilization Breakdown

ActivityTarget %Notes
Ticket handling70-80%Core productive time
Training5-10%Ongoing development
Meetings5-8%Team, 1:1s, calibration
Admin3-5%Documentation, projects
Breaks/Buffer5-10%Bathroom, mental breaks
Shrinkage total20-30%Not available for tickets

Good Workforce Planning

✓ Data-driven forecasting
  → Historical patterns analyzed
  → Seasonality accounted for
  → Growth factored in

✓ Flexible staffing
  → Part-time for peak hours
  → Cross-trained backups
  → Contractor buffer available

✓ Real-time monitoring
  → Queue visibility
  → Quick response to spikes
  → Adjust on the fly

✓ Forward-looking
  → 3-month rolling forecast
  → Hiring lead time included
  → Training time planned

✓ Agent wellbeing
  → Sustainable utilization
  → Predictable schedules
  → Burnout prevention

Bad Workforce Planning

✗ Reactive staffing
  → Hire after crisis
  → Always behind

✗ One-size schedules
  → Same coverage all hours
  → Peaks understaffed

✗ 100% utilization target
  → No buffer for spikes
  → Agent burnout

✗ Ignoring seasonality
  → January staff for December volume
  → Surprised every year

✗ Training as afterthought
  → New hires unproductive for months
  → Existing team can't grow

Forecasting Methods

MethodDescriptionBest For
Historical averageLast N periods averageStable volume
Trend analysisApply growth rateGrowing companies
Seasonal adjustmentLast year same periodClear seasonality
Moving averageRolling X-week averageSmoothing noise
RegressionVolume vs driversComplex patterns

Seasonal Patterns to Watch

PatternWhenPlanning Action
Holiday dipsDec 24-Jan 1Reduced staffing OK
End of monthLast 3 daysBilling questions spike
Monday surgeMonday AMFull staffing
Release spikesAfter deploymentsExtra staffing
Renewal periodsQuarterlyCS/billing tickets up
Tax seasonApril (US)Financial software surge

Staffing Model by Channel

ChannelAgent:Tickets/HourConcurrencyNotes
Email4-81Deep work, can batch
Live Chat2-42-3Multiple simultaneous
Phone3-51No multitasking
Social5-101-2Often quick responses

Scheduling Best Practices

Schedule Design:
├── Cover peak hours with full staff
├── Stagger start times for coverage
├── Allow schedule preferences within limits
├── Plan for timezone coverage
└── Build in overlap for handoffs

Shift Examples:
├── Early: 6am-2pm (catch Asia-Pacific overlap)
├── Core: 9am-5pm (highest volume)
├── Late: 12pm-8pm (West Coast coverage)
└── Split: 9-1pm + 4-8pm (two peaks)

Schedule Template

DayEarly (6a-2p)Core (9a-5p)Late (12p-8p)Total
Monday4 agents12 agents6 agents22
Tuesday3 agents10 agents5 agents18
Wednesday3 agents10 agents5 agents18
Thursday3 agents10 agents5 agents18
Friday3 agents8 agents4 agents15
Saturday1 agent2 agents1 agent4
Sunday1 agent2 agents1 agent4

Real-Time Management

MetricGreenYellowRedAction
Wait Time<5 min5-15 min>15 minPull from training, escalate
Queue Depth<1010-25>25All hands, defer non-critical
Utilization70-80%80-90%>90%Monitor burnout, get help
Available AgentsPer plan-2-4+Call in backups

Hiring Planning

ScenarioLead TimeConsiderations
Replacement2-3 monthsNotice, interview, training
Growth hire3-4 monthsSame + slower pipeline
New tier/specialty4-6 monthsHarder to find, more training
Seasonal1-2 monthsTemp/contract can be faster

New Hire Ramp Schedule

WeekFocusTicket LoadSupport
1Orientation, product training0%Classroom
2Shadowing0%Paired with mentor
3Supervised tickets25%QA every ticket
4Supervised tickets50%QA every ticket
5-6Increasing independence75%Daily check-ins
7-8Full load100%Standard support

Capacity Dashboard

MetricHow to CalculateReview
Tickets/Agent/DayTotal tickets ÷ FTEDaily
Handle Time TrendAvg over timeWeekly
UtilizationTicket time ÷ Available timeDaily
BacklogTickets >24h oldReal-time
Coverage GapsHours with <50% target staffingWeekly
Cost Per TicketSupport cost ÷ Total ticketsMonthly

Cost Per Ticket Analysis

Full Cost Calculation:
─────────────────────
Agent salary/benefits:     $60,000/year
Tools/systems (per seat):   $5,000/year
Training/development:       $2,000/year
Management overhead:        $8,000/year
Total per agent:          $75,000/year

Tickets per agent/year: 6,000 (25/day × 240 days)

Cost per ticket: $12.50

Reducing handle time by 10% = $1.25/ticket savings
Improving FCR by 5% = 300 tickets avoided = $3,750/agent/year

Capacity Planning Calendar

Monthly:
□ Review previous month actuals vs forecast
□ Update 3-month rolling forecast
□ Adjust schedules for known events
□ Review hiring pipeline status

Quarterly:
□ Deep analysis of volume trends
□ Headcount planning for next 2 quarters
□ Training calendar planning
□ Budget review

Annually:
□ Annual volume forecast
□ Headcount budget request
□ Tool/vendor contract reviews
□ Strategic capacity planning

Scenario Planning

ScenarioVolume ChangeResponse
Product outage+300%All hands, defer non-critical
Major release+50%Pre-scheduled extra coverage
Competitor migration+100%Temp staff, extended hours
Holiday-50%Reduced staffing, skeleton crew
PR crisis+500%Emergency protocol, exec support

Anti-Patterns

  • Heroic unsustainable effort — Team working overtime constantly
  • Understaffed normal — Always behind, never catching up
  • Over-hiring for peak — Paying for capacity used 10% of time
  • Ignoring ramp time — New hires counted as full capacity
  • Static schedules — Same coverage regardless of patterns
  • Utilization obsession — 95% target burns out team
  • No forecast review — Never learning from actuals
  • Last-minute scramble — Emergency scheduling constantly