Help center articles look simple until they are wrong. A single outdated step or an overconfident claim can create tickets, refunds, and frustrated customers. At the same time, documentation teams are often small, and they are expected to keep up with product changes, support feedback, and SEO needs.
AI drafting can help, but only if publishing stays disciplined. The fastest way to get disciplined is to replace subjective review comments like “feels off” with a shared rubric that makes quality measurable, repeatable, and easy to train.
This post gives you a practical rubric and a lightweight workflow you can adopt even if you are a team of one. It is designed for customer-facing help center articles, but the pattern also works for internal runbooks and onboarding docs.
Why a rubric beats gut feel
Without a rubric, reviewers tend to focus on what they personally notice: grammar, tone, or a single missing step. That leaves bigger risks unreviewed, like whether the steps are actually executable, whether the article matches current product behavior, or whether it safely handles edge cases.
A rubric helps because it:
- Creates shared standards: two reviewers can disagree on style but still agree on “does this contain a reversible procedure and a clear warning.”
- Lets you triage: a 6/10 draft should not go live, but it might be good enough for internal use.
- Improves prompts and sources: consistent failures point to missing inputs (policy docs, UI labels, supported plans) instead of vague “the model is bad.”
- Makes sampling possible: you can spot-check 10 percent of published articles each month and track whether quality drifts.
Think of the rubric as a small “contract” between the draft, the reviewer, and your readers.
Define the article contract
Before scoring anything, define what “good” means for your help center. This prevents a common failure mode: the AI produces a well-written article that is about the wrong problem, for the wrong audience, at the wrong level of detail.
Minimum inputs for every draft
Require these inputs before an AI draft is allowed to be generated:
- Single user goal: “Change billing address,” not “Billing settings overview.”
- Audience and prerequisites: role, plan tier, required permissions, and any needed access.
- Source of truth: internal policy notes, product spec, or a verified checklist from a subject matter expert.
- UI vocabulary: current menu names and button labels (even if you keep it high level, it reduces hallucinated UI).
- Support boundary: what to do when the user cannot proceed (contact support, admin-only action, expected error message).
Define your “hard no” rules
Write down a few non-negotiables. Examples: never instruct users to share passwords, never describe unsupported workarounds, and never imply guarantees (“will be resolved in 24 hours”). If your rubric includes these as explicit checks, reviewers stop having to remember them.
The 10-point scoring rubric
This rubric is intentionally small. It is easier to use consistently, and it is easier to train new reviewers on. Score each category 0, 1, or 2. A strong publish threshold is typically 16/20 or higher, with some “must pass” checks noted below.
{
"goalFit": 0-2,
"factualAccuracy": 0-2,
"stepExecutability": 0-2,
"edgeCasesAndLimits": 0-2,
"safetyAndPrivacy": 0-2,
"clarity": 0-2,
"toneAndEmpathy": 0-2,
"structureAndScanability": 0-2,
"supportDeflection": 0-2,
"maintenanceReadiness": 0-2
}
- Goal fit: Does the article solve the stated user goal without drifting? (0 = wrong goal, 1 = partially, 2 = tight)
- Factual accuracy (must pass): Are claims, policy statements, and UI names correct? (0 = unverified/incorrect, 1 = mostly, 2 = verified)
- Step executability (must pass): Can a user follow the steps in order without missing prerequisites? (0 = broken, 1 = some gaps, 2 = runnable)
- Edge cases and limits: Does it mention common blockers like permissions, plan tier, or region? (0 = none, 1 = a little, 2 = key ones)
- Safety and privacy (must pass): No unsafe guidance, no sensitive data requests, correct warnings. (0 = unsafe, 1 = unclear, 2 = safe)
- Clarity: Short sentences, concrete verbs, avoids jargon. (0 = confusing, 1 = ok, 2 = clear)
- Tone and empathy: Neutral, helpful, not blaming the user, avoids sarcasm. (0 = harmful, 1 = mixed, 2 = good)
- Structure and scanability: Good headings, bullets, and “if you see X” guidance. (0 = wall of text, 1 = some structure, 2 = easy to scan)
- Support deflection: Includes troubleshooting and “what to do next,” so the user does not immediately open a ticket. (0 = none, 1 = minimal, 2 = solid)
- Maintenance readiness: Avoids brittle UI details, uses stable concepts, and includes a clear owner or review trigger. (0 = brittle, 1 = mixed, 2 = maintainable)
Key Takeaways
- Score drafts on a small set of risk-focused categories, not just writing quality.
- Make “accuracy,” “executability,” and “safety” explicit must-pass checks.
- Use the rubric results to improve your inputs and sources, not only the wording.
- Keep the workflow lightweight so it actually gets used during busy weeks.
A lightweight review workflow
The goal is speed with guardrails. You want a process that is easy to run repeatedly, and that fails safe when it has to.
Suggested states
- Drafted (AI): the article exists, but has not earned trust.
- Reviewed: rubric completed, required fixes applied.
- Verified: critical steps were executed or confirmed by a subject matter expert.
- Published: live and discoverable.
- Watchlisted: product area is changing; re-check on a schedule.
Copyable reviewer checklist
- Confirm the user goal and audience match the title and first paragraph.
- Check all UI labels against the current app, or rewrite to be less label-dependent.
- Run the procedure (or simulate it) for at least one “happy path.”
- Scan for unsafe guidance: credentials, personal data, irreversible actions without warnings.
- Add at least one troubleshooting section: “If you do not see X, then Y.”
- Add a support handoff line that includes what information to provide (non-sensitive).
- Assign an owner and a review trigger (release note tag, feature flag removal, policy change).
- Fill in the rubric scores and record one sentence on the biggest remaining risk.
If your team is very small, combine “Reviewed” and “Verified,” but keep the idea of verification. Someone should be accountable for whether the steps reflect reality.
A concrete example: billing address changes
Imagine a SaaS company where customers can update their billing address, but only account owners can do it, and only when there are no overdue invoices. Support sees a spike in tickets because users follow an old article that never mentions these constraints.
You generate an AI draft using the “minimum inputs” contract: user goal (“change billing address”), prerequisites (owner role), policy note (overdue invoices block edits), and the correct settings navigation.
During review, the article scores well on clarity and structure, but it gets a 1/2 on edge cases and a 1/2 on support deflection. The reviewer adds:
- A callout: “If you have overdue invoices, address changes are temporarily locked.”
- A troubleshooting step: “If you do not see Billing, you are not the account owner.”
- A support handoff: “If you believe the lock is in error, contact support with your account ID and invoice numbers (do not send card details).”
Now the article does not just explain the happy path. It prevents avoidable tickets by addressing the two most common blockers.
Common mistakes (and how to prevent them)
- Scoring the writing, not the risk: A friendly tone cannot compensate for incorrect steps. Make accuracy and executability must-pass checks.
- Letting the AI invent UI labels: If you cannot provide exact labels, instruct the draft to use stable descriptions like “Open your account settings” and verify during review.
- Over-specifying brittle steps: “Click the blue button in the top right” will age badly. Prefer “Select Update Billing Address” and keep screenshots out of the critical path.
- No ownership: Docs without owners become archaeological sites. Add an owner field and review triggers.
- Publishing without a rollback plan: If an article is wrong, you should be able to unpublish quickly. Treat docs like product: changes can be reverted.
When not to use AI drafting
AI drafting is most effective when the problem is well-specified and the answer is stable. Consider avoiding or limiting AI-generated drafts when:
- The procedure is high risk: actions that can delete data, change security settings, or impact billing in irreversible ways.
- The policy is ambiguous or under debate: if humans do not agree on the rule, the model will not either.
- You cannot provide a source of truth: asking the model to “figure it out” often produces confident fiction.
- The UI is changing weekly: focus on short-lived internal notes, or publish only high-level concepts until the UI stabilizes.
A reasonable compromise is to let AI produce a structured outline and a first draft, while requiring human-authored steps and warnings for the sensitive sections.
FAQ
What publish threshold should we use?
Start with 16/20 plus must-pass checks for factual accuracy, step executability, and safety/privacy. After a month, compare rubric scores to real outcomes like ticket volume or article feedback and adjust.
How do we “verify” an article if we cannot reproduce the issue?
Verification can be a mix of methods: running the steps in a test account, confirming with a subject matter expert, or checking logs and admin tools. If none are possible, mark the article as watchlisted and avoid strong claims.
How do we keep the rubric from slowing everything down?
Time-box reviews. For routine articles, aim for a 10 to 15 minute pass with the checklist. Reserve deeper verification for high-impact topics like login, billing, and account security.
Can we use this rubric for internal runbooks too?
Yes. Internal docs benefit even more from executability and edge-case scoring. You may also add a category for “operational safety” if steps interact with production systems.
Conclusion
AI can make help centers faster to write, but a rubric makes them safer to publish. By defining a simple contract, scoring drafts against risk-focused categories, and using a lightweight review flow, you can scale documentation without scaling avoidable support tickets.
If you want more posts like this, browse the Archive or learn how this site works on the About page.