Skip to content
M
Mediyn
Back to BlogCompliance & Privacy

De-Identification in Therapy Technology: What You Need to Know

8 min readJanuary 30, 2026

As AI tools become standard in therapy practices, a critical question emerges: how do you use powerful machine learning on sensitive clinical data without compromising patient privacy? The answer is de-identification — the process of removing or obscuring information that could link clinical content to a specific individual.

De-identification isn't new. Researchers have used it for decades to study clinical data without compromising patient confidentiality. What's new is its application in real-time clinical workflows, where AI systems need to process session content immediately while ensuring that identifying information never reaches external servers.

What De-Identification Means Under HIPAA

HIPAA defines two methods for de-identification:

Expert Determination (Section 164.514(b)(1))

A qualified statistical expert certifies that the risk of identifying an individual from the data is "very small." This method is flexible but requires expert analysis.

Safe Harbor (Section 164.514(b)(2))

All 18 HIPAA identifiers are removed, and the covered entity has no actual knowledge that the remaining information could identify an individual. The 18 identifiers include names, dates, geographic data, phone numbers, email addresses, Social Security numbers, medical record numbers, and others.

Once data is properly de-identified under either method, it is no longer considered PHI and is not subject to HIPAA regulations. This is a powerful concept: de-identified data can be processed, transmitted, and stored without the restrictions that apply to PHI.

De-Identification in AI Documentation

In the context of AI therapy documentation, de-identification works like this:

  • A therapy session is transcribed on-device.
  • Named entity recognition algorithms identify all HIPAA identifiers in the transcript.
  • Each identifier is replaced with a category token (e.g., [NAME], [DATE], [LOCATION]).
  • The de-identified transcript is sent to the AI for note generation.
  • The AI generates a clinical note using only de-identified content.
  • Back on the device, tokens are replaced with original values for the final note.

The AI service only ever processes de-identified data. It knows a patient discussed a conflict with [FAMILY_MEMBER] at [LOCATION] on [DATE], but it has no way to determine who the patient is, who the family member is, or where or when the event occurred.

The Technical Challenge: Context Preservation

The art of de-identification in therapy technology is removing identifying information while preserving enough clinical context for the AI to generate an accurate note. Simply stripping all identifying information can leave the AI without enough context to understand relationships and timelines.

Good de-identification systems preserve:

  • Relationship categories: [FAMILY_MEMBER] vs. [COWORKER] vs. [FRIEND] — the AI needs to know the type of relationship to generate clinically relevant notes.
  • Temporal relationships: Even though specific dates are removed, the relative timing ("two weeks ago," "last month") is preserved so the AI can accurately describe the sequence of events.
  • Consistency: The same entity gets the same token throughout the transcript. If the patient mentions "Jen" ten times, it's [FAMILY_MEMBER_1] every time, so the AI can track narrative threads.

De-Identification vs. Anonymization

These terms are often confused but have different meanings:

De-identification is reversible. The original data exists somewhere (in this case, on the local device) and can be re-linked to the de-identified data using the token mapping. This is what allows the final note to contain actual names and dates after the AI draft returns.

Anonymization is irreversible. The link between the data and the individual is permanently destroyed. Anonymized data is more private but less useful for clinical documentation, where you ultimately need the note to contain the patient's actual information.

In therapy AI, de-identification (not anonymization) is the right approach because you need the reversibility to produce a usable clinical note.

Risks and Limitations

No de-identification system is perfect. Known challenges include:

  • Unusual identifiers: A name that's also a common word ("Grace," "Joy") may be missed by NER models.
  • Indirect identifiers: A combination of non-PHI details (profession + city + unusual medical condition) could theoretically identify someone even without any of the 18 HIPAA identifiers.
  • Context leakage: In small populations, even de-identified clinical details might narrow identification (e.g., the only person in a small town who experienced a specific event).

These risks are real but manageable. The human review step — where you check the AI-generated note before signing — serves as the final quality control for any de-identification gaps. If you see a name that should have been redacted, you correct it before the note is finalized.

What to Look for in a Platform

When evaluating AI documentation tools, ask about their de-identification approach:

  • Where does de-identification occur? On-device is the gold standard — PHI never leaves the device in identifiable form.
  • What method is used? NER-based token replacement is state of the art for clinical text.
  • What's the accuracy rate? Look for 97%+ PHI detection accuracy, validated on clinical conversation data.
  • Is the process auditable? Can you review what was redacted and verify completeness?
  • Is there a human-in-the-loop? Does the system allow clinician review before notes are finalized?

The Bigger Picture

De-identification is what makes it possible to harness AI's capabilities for therapy documentation while maintaining the privacy standards your patients deserve. It's the technical foundation that turns "AI processing patient data" — which sounds alarming — into "AI processing clinical content that can't be traced to any individual" — which is genuinely protective.

As a therapist, you don't need to become a de-identification expert. But you should understand the concept well enough to evaluate the tools you use and explain the privacy protections to your patients confidently.

Learn how Mediyn's AI documentation uses on-device de-identification to protect patient privacy at every step.

Ready to see Mediyn in action?

7 days free. Full access. Credit card required. Cancel anytime.