Artificial intelligence tools are rapidly entering healthcare workflows. Clinicians are experimenting with AI to summarize patient encounters, generate chart notes, and answer billing questions. While these tools can improve efficiency, there is an important compliance issue many providers overlook:
Entering patient transcripts into an AI system may constitute a disclosure of Protected Health Information (PHI) under HIPAA.
Even when a patient’s name or date of birth is removed, the information may still be considered identifiable if other data elements remain.
Understanding HIPAA de-identification standards is critical before using AI tools with patient information.
Protected Health Information (PHI) refers to individually identifiable health information held or transmitted by a covered entity or its business associate in any form, including electronic, written, or verbal data.
PHI includes information that:
Examples include patient names, medical records, lab results, diagnoses, medications, and billing data.
Under HIPAA, PHI must be protected from unauthorized disclosure.
Many clinicians are testing AI systems by pasting patient transcripts into tools like ChatGPT or other large language models (LLMs). However, doing so can create compliance risks.
When patient data is transmitted to an external service:
If the AI system does not provide a BAA, transmitting PHI to that system could be considered an unauthorized disclosure under HIPAA.
HIPAA allows healthcare data to be used or shared without restrictions if it has been properly de-identified.
There are two methods for de-identification:
The most widely referenced method is Safe Harbor, which requires removing 18 specific identifiers.
If any of these identifiers remain and the patient could still be recognized, the data may still qualify as PHI.
Under the Safe Harbor method, the following identifiers must be removed to consider information de-identified:
Even after these identifiers are removed, the data must not contain information that could reasonably identify the individual.
For example:
These factors can still make a patient identifiable.
Clinical transcripts often contain detailed contextual information, including:
Even without direct identifiers, these details may allow someone to recognize a patient.
For example:
A transcript describing “a pediatric neurologist in a small town treated for narcolepsy and bipolar disorder after a specific car accident last year” could easily identify an individual.
Because of this, healthcare compliance officers often recommend not placing patient transcripts into public AI systems.
Healthcare providers considering AI documentation tools should follow several compliance practices:
Ensure the AI vendor provides a signed BAA if the system will process PHI.
Consumer AI platforms are typically not designed for HIPAA compliance.
Clinical AI tools should include:
Limit the amount of patient data shared with any external system.
Healthcare organizations should review AI workflows with HIPAA compliance teams or legal counsel.
Mental health documentation presents additional challenges because psychiatric sessions often include:
These elements increase the risk that a transcript may contain identifiable contextual information, even when obvious identifiers are removed.
As a result, AI documentation systems for psychiatry must prioritize privacy-focused design and clinical safeguards.
It depends. If the AI vendor does not provide a Business Associate Agreement (BAA) and the information contains identifiable patient data, it may constitute an unauthorized disclosure of PHI.
No. HIPAA requires the removal of 18 specific identifiers under the Safe Harbor standard. Names are only one of these identifiers.
Yes. Clinical transcripts often contain contextual details that can identify a patient and therefore qualify as Protected Health Information.
Data is considered de-identified if it meets either:
Yes. AI can be used safely when systems are designed with:
Healthcare-specific AI platforms are increasingly being developed to meet these requirements.
Artificial intelligence will continue transforming healthcare documentation, coding assistance, and clinical decision support.
However, privacy protections must remain central to these innovations.
Understanding HIPAA de-identification standards helps clinicians adopt AI responsibly while protecting patient confidentiality.
Sources
U.S. Department of Health and Human Services – HIPAA Privacy Rule
https://www.hhs.gov/hipaa/for-professionals/privacy/index.html
HIPAA De-Identification Guidance
https://www.hhs.gov/hipaa/for-professionals/special-topics/de-identification/index.html
HIPAA Business Associate Guidance
https://www.hhs.gov/hipaa/for-professionals/privacy/guidance/business-associates/index.html