Recovery-Content Carve-Outs
A practical guide to recovery-content carve-outs for AI content-moderation practitioners.
What This Lesson Covers
Recovery-Content Carve-Outs is a key lesson within Self-Harm & Suicide Content Policy. In this lesson you will learn the underlying content-moderation discipline, the practical artefacts and rituals that operationalise it inside a working team, how to apply the pattern to a live platform, and the failure modes that undermine it in practice.
This lesson belongs to the Harmful Content Categories category. The category covers the specific high-stakes content types that mature platforms write deep policies for — child safety, violent extremism, hate speech, harassment, mis & disinformation, NSFW, self-harm and suicide — each with its own legal regime, evidentiary standard, and operational discipline.
Why It Matters
Engineer for self-harm and suicide content policy. Learn the WHO and Reporting on Suicide safe-messaging guidelines, crisis-response interstitials with hotlines per region, recovery-content carve-outs (lived-experience speech is not the same as glorification), the link to youth-safety policy, and the reviewer-wellness specifics for this high-trauma category.
The reason this lesson deserves dedicated attention is that AI content moderation is now operationally load-bearing: the EU DSA imposes direct duties with material fines, the UK Online Safety Act adds duty-of-care obligations, NetzDG and India IT Rules add jurisdiction-specific requirements, customers and advertisers ask for transparency, civil society and oversight boards inspect decisions, and journalists publish individual cases. Practitioners who reason from first principles will navigate the next obligation, the next incident, and the next stakeholder concern far more effectively than those working from a stale checklist.
How It Works in Practice
Below is a practical content-moderation pattern for recovery-content carve-outs. Read through it once, then think about how you would apply it inside your own platform.
# Content-moderation pattern
MODERATION_STEPS = [
'Anchor the work in a specific policy provision and the harm it addresses',
'Pick the detection layer (hash, classifier, user report, regulator referral)',
'Run the decision with policy-aware reviewer or model, with rationale capture',
'Apply a proportionate action from the action ladder',
'Notify the affected user with a statement of reasons and appeal route',
'Run appeals, oversight, and post-incident review with closure to action items',
'Disclose aggregate enforcement in the transparency report',
]
Step-by-Step Operating Approach
- Anchor in a specific policy provision — Which provision in the policy taxonomy does this work serve, and what harm does it address? Skip this step and you build activity without legal or policy grounding.
- Pick the detection layer — Hash for known-bad, classifier for at-scale recall, user report for context, regulator referral for legal mandates. The right layer matters; the wrong one wastes capacity.
- Run the decision with policy-aware tooling — Reviewers and models need the policy in front of them, with rationale capture for audit. Ad-hoc decisions do not survive contact with appeals or regulators.
- Apply a proportionate action — The action ladder runs from label and demote through remove, suspend, and ban. Proportionality — severity meets action — is what keeps trust intact.
- Notify the affected user — Statement of reasons, policy reference, appeal route. DSA Article 17 makes this a direct duty in the EU; user trust makes it a good idea everywhere else.
- Run appeals, oversight, and PIR — Appeals must be timely, accessible, and effective. Oversight bodies look at patterns. PIRs feed action items back into policy and detection.
- Disclose in the transparency report — Aggregate enforcement, with methodology and comparability, is the public proof the system is real.
When This Topic Applies (and When It Does Not)
Recovery-Content Carve-Outs applies when:
- You are designing, shipping, or operating a platform that hosts user-generated content or AI-generated content
- You are standing up or operating a Trust & Safety function
- You are integrating AI / generative AI features into a regulated platform (EU DSA, UK Online Safety Act, NetzDG, India IT Rules, COPPA / AADC for child-facing surfaces)
- You are responding to a customer, regulator, oversight board, journalist, or board question about moderation practice
- You are running a transparency report, statement-of-reasons program, or third-party moderation audit
- You are defining or honouring moderation commitments in a policy, RSP, or system card
It does not apply (or applies lightly) when:
- The work is pure research with no path to a deployed platform
- The system genuinely hosts no user-generated or AI-generated content visible to others
- The activity is one-shot procurement of a closed-context tool with no public surface
Practitioner Checklist
- Is the policy provision this lesson addresses written in SMART form, with examples and counter-examples, in the public taxonomy?
- Is the detection layer (hash, classifier, user report, referral) appropriate for the harm and validated with slice eval?
- Is the decision captured with rationale linked to the specific policy provision?
- Is the action chosen proportionately from the action ladder, with audit?
- Is the user notified with a statement of reasons and a working appeal route within the SLA?
- Are appeals reviewed independently and tracked for reversal-rate and policy-impact?
- Are aggregate outcomes disclosed in the transparency report with reproducible methodology?
Disclaimer
This educational content is provided for general informational purposes only. It does not constitute legal, regulatory, trust-and-safety, or professional advice; it does not create a professional engagement; and it should not be relied on for any specific moderation decision. Intermediary-liability law, content-moderation duties, and platform-policy norms vary by jurisdiction and sector and change rapidly. Consult qualified platform / media counsel, trust-and-safety practitioners, and risk professionals for advice on your specific situation.
Next Steps
The other lessons in Self-Harm & Suicide Content Policy build directly on this one. Once you are comfortable with recovery-content carve-outs, the natural next step is to combine it with the patterns in the surrounding lessons — that is where doctrinal mastery turns into a working content-moderation capability. Content moderation is most useful as an integrated discipline covering policy, detection, decision, action, notification, appeal, and transparency.
Lilly Tech Systems