AI Governance

What Counts as PHI in AI Tools? The Mosaic Effect

| | 16 min read | Updated March 1, 2026

Bottom Line Up Front

PHI in AI tools extends beyond the 18 HIPAA Safe Harbor identifiers. The Mosaic Effect allows AI to re-identify patients by combining ZIP codes, dates, and diagnosis codes with public datasets, reaching 87-95% accuracy. Any data entered into an AI tool without a signed BAA constitutes a potential HIPAA violation under 164.308(b)(1), regardless of de-identification efforts.

In 2000, Latanya Sweeney at Carnegie Mellon demonstrated that 87% of the U.S. population becomes uniquely identifiable from three data points: five-digit ZIP code, gender, and date of birth [Sweeney 2000]. She proved it by re-identifying the Massachusetts governor’s medical records from a publicly available “de-identified” state employee health insurance database. The privacy community published the finding. HIPAA regulators published the Safe Harbor method. Organizations removed the 18 listed identifiers and treated the problem as solved.

Twenty-five years later, the same problem returned with exponentially greater reach. Machine learning algorithms now re-identify individuals from de-identified datasets at 95% accuracy [JAMA 2025]. AI tools do not review data the way human analysts do. They cross-reference every query against training data sourced from the indexed internet: obituaries, property records, voter rolls, social media profiles. The 18 identifiers your compliance team removes satisfy the regulatory checklist. The data points remaining satisfy the AI.

PHI in AI tools extends beyond the HIPAA Safe Harbor list. The Mosaic Effect, where individually harmless data points combine to re-identify patients, transforms every clinical dataset entering an AI tool into a potential breach notification. The compliance question is no longer whether you removed the 18 identifiers. The question is whether the remaining data points give a machine enough information to reconstruct the patient’s identity.

At A Glance: PHI in AI tools extends beyond the 18 HIPAA Safe Harbor identifiers. The Mosaic Effect allows AI to re-identify patients by combining ZIP codes, dates, and diagnosis codes with public datasets, reaching 87-95% accuracy. Any data entered into an AI tool without a signed BAA constitutes a potential HIPAA violation under 164.308(b)(1), regardless of de-identification efforts.

The Mosaic Effect: How AI Breaks De-Identification

Three Data Points, 87% Re-Identification

Latanya Sweeney’s landmark research at Carnegie Mellon established the statistical foundation for re-identification risk. By cross-referencing hospital discharge records with public voter registration data, Sweeney demonstrated 87% of the U.S. population (216 million of 248 million) becomes uniquely identifiable with three demographic fields: five-digit ZIP code, gender, and date of birth [Sweeney 2000]. The study famously identified the Massachusetts governor’s medical records from a “de-identified” state employee health insurance database.

HIPAA’s Safe Harbor method addresses this by requiring removal of all 18 specified identifiers [HIPAA 164.514(b)(2)]. The method also restricts ZIP codes to three-digit prefixes for geographic areas with populations under 20,000 [HIPAA 164.514(b)(2)(i)(B)]. These protections assumed human analysts reviewing static spreadsheets.

AI does not work with static spreadsheets. Large language models process queries against training data sourced from the indexed internet: obituaries, GoFundMe campaigns, property records, social media profiles, and voter rolls.

From Three to Ninety-Five: Machine Learning Amplifies the Risk

Sweeney’s 87% figure reflected manual cross-referencing in 2000. Modern machine learning reaches higher. A JAMA-published study tested random forest algorithms against the National Health and Nutrition Examination Survey (NHANES) dataset of 14,451 individuals. The algorithm matched physical activity data and demographic information to 95% of adults [JAMA 2025]. Harvard’s Data Privacy Lab extended this work: combining genomic profiles with public voter lists achieved 84-97% re-identification accuracy [Data Privacy Lab 2024].

A separate study on HIPAA-compliant hospital discharge data found 28.3% of individuals in Maine and 34% in Vermont re-identified from datasets meeting Safe Harbor requirements [AI and Ethics 2025]. The re-identification succeeded because Safe Harbor addresses the 18 listed identifiers but does not account for the combinatorial power of remaining data points. AI exploits precisely this gap.

The HIPAA Safe Harbor Gap

HHS acknowledged neither de-identification method eliminates all re-identification risk [HHS OCR De-Identification Guidance]. The Safe Harbor identifier list was published over two decades ago. It predates social media handles, emotional support animal registries, wearable health device data, and AI-accessible public records databases. Four data points (date of birth, ZIP code, gender, and occupation) now re-identify up to 95% of individuals [aboutmyinfo.org].

Organizations relying exclusively on Safe Harbor compliance before entering data into AI tools operate under a false assumption. The identifiers removed satisfy the regulatory checklist. The identifiers remaining satisfy the AI.

The Audit Fix: Stop treating Safe Harbor as a license to paste data into AI tools.

  1. Run a formal re-identification risk assessment before any de-identified dataset enters an AI tool [HIPAA 164.308(a)(1)(ii)(A)].
  2. Apply the Expert Determination method (164.514(b)(1)) for any dataset containing three or more quasi-identifiers (ZIP prefix, age range, diagnosis code, admission date).
  3. Document the statistical basis for your re-identification risk determination. “We removed the 18 identifiers” is insufficient when AI-enabled cross-referencing is reasonably anticipated.
  4. Treat any AI tool processing clinical data as a business associate. Sign a BAA before the first query [HIPAA 164.308(b)(1)].

Which HIPAA Identifiers Do AI Tools Capture Without Your Knowledge?

The Four Identifiers Developers Overlook

HIPAA’s Safe Harbor method requires removal of 18 categories of identifiers [HIPAA 164.514(b)(2)]. Developers remove names, Social Security numbers, and medical record numbers. They routinely miss four categories AI tools capture automatically.

Identifier HIPAA Reference AI Capture Method and Risk
IP Address #15 [164.514(b)(2)(i)(O)] Logged on every API call. Geolocates to building level.
Internal Patient IDs #7 [164.514(b)(2)(i)(H)] Pasted in JSON during debugging. Links to EHR designated record set.
Voice Recordings #16 [164.514(b)(2)(i)(P)] Captured by AI scribes. Voiceprint accuracy matches fingerprints.
Dates #3 [164.514(b)(2)(i)(C)] Embedded in AI prompts. Triangulates identity with diagnosis and age.

Voice and Biometrics: The Invisible PHI

Voice data qualifies as PHI under two HIPAA provisions. First, voiceprints fall under biometric identifiers (Safe Harbor identifier #16) [HIPAA 164.514(b)(2)(i)(P)]. Second, voice recordings containing clinical content constitute ePHI requiring Security Rule protections [HIPAA 164.312]. A person’s voiceprint remains unique even when they never state their name during a recorded session.

AI scribe adoption surged in healthcare through 2025 and 2026. Consumer-tier tools (Otter.ai free plan, Fireflies basic) lack BAAs, lack encryption at rest, and train on uploaded audio. Voice AI crossed a commercial threshold in 2025, moving from experimental customer support into core healthcare documentation workflows [Speechmatics 2026]. Every unprotected AI scribe recording of a clinical encounter constitutes a potential breach.

Metadata Counts: Timestamps, Device IDs, and Session Logs

HIPAA identifier #14 covers device identifiers and serial numbers [HIPAA 164.514(b)(2)(i)(N)]. Identifier #15 covers web URLs and IP addresses [HIPAA 164.514(b)(2)(i)(O)]. AI tools log both automatically. A ChatGPT API call records the timestamp, the originating IP, the device fingerprint, and the session identifier. None of these appear in the developer’s prompt. All of them appear in the AI vendor’s server logs.

Identifier #18 covers “any other unique identifying number, characteristic, or code” [HIPAA 164.514(b)(2)(i)(R)]. This catch-all provision applies to session tokens, API keys linked to specific clinical systems, and custom internal codes developers embed in prompts. If the code maps back to an individual through your internal systems, it qualifies as PHI.

The Audit Fix: Audit every data field your AI tools capture, not the data fields your developers paste.

  1. Request the full data collection disclosure from each AI vendor. Compare their logged fields against the 18 HIPAA identifiers [HIPAA 164.514(b)(2)].
  2. Block consumer-tier AI scribes on your network. Require HIPAA-compliant alternatives (Zoom for Healthcare, DAX Copilot, or BAA-covered APIs) for clinical voice recording.
  3. Strip metadata before API calls: sanitize IP addresses, device identifiers, and session tokens from AI tool requests using a middleware proxy.
  4. Add AI tool data capture to your annual HIPAA risk assessment scope [HIPAA 164.308(a)(1)(ii)(A)].

The “Random ID” Myth: Why Pseudonymization Fails with AI

What HHS Guidance Says About Re-Identification Codes

Developers frequently replace patient names with pseudonyms: Patient_X99, Case_4471, Subject_Alpha. They assume the substitution satisfies de-identification. HHS guidance states otherwise. Under 164.514(c), a covered entity assigns a code to de-identified data for re-identification purposes only if the code does not derive from the individual’s information and the entity does not use or disclose the code for any other purpose [HIPAA 164.514(c)].

The critical test: if your organization maintains a mapping table linking Patient_X99 to the patient’s actual identity, then Patient_X99 is PHI within the meaning of the Designated Record Set [HIPAA 164.501]. Sending Patient_X99 to an AI tool without a BAA violates 164.308(b)(1). The pseudonym carries the full regulatory weight of the patient’s name because your organization holds the key.

The Designated Record Set Problem

HIPAA defines the Designated Record Set as medical records, billing records, and other records used to make decisions about individuals [HIPAA 164.501]. A mapping table linking pseudonymized IDs to patient identities falls within this definition. The pseudonym does not exist independently. It exists as one column in a relational database where the adjacent column contains the patient’s name, date of birth, and medical record number.

Deleting the mapping table does not solve the problem if backups, EHR audit logs, or version-controlled repositories retain the mapping. The re-identification path persists as long as any copy of the key exists within the covered entity or its business associates.

AI Reverse-Engineers Pseudonyms

Even without the mapping table, AI models cross-reference pseudonymized data against contextual clues. A prompt containing “Patient_X99, 43-year-old male, ZIP 02138, admitted January 15 for stage 3 glioblastoma” gives the AI enough quasi-identifiers to narrow the population to a handful of individuals. Cross-referencing against public cancer registries, obituary databases, or hospital press releases completes the identification. The pseudonym provided a false sense of security. The surrounding data performed the re-identification.

The Audit Fix: Stop relying on pseudonymization as a de-identification method for AI tool inputs.

  1. Treat all pseudonymized data as PHI if your organization retains the re-identification key [HIPAA 164.514(c)].
  2. Apply generalization before AI submission: replace specific ages with decade ranges, replace specific dates with quarter/year, remove diagnosis codes or replace with ICD category-level groupings.
  3. Implement automated scrubbing pipelines between your EHR and AI endpoints. Do not rely on developers to manually redact before pasting.
  4. Document your pseudonymization controls in your HIPAA policies and procedures [HIPAA 164.530(j)]. Retain documentation for six years.

The Minimum Necessary Standard Applies to Every AI Prompt

HIPAA 164.502(b): The Overlooked AI Requirement

HIPAA’s Minimum Necessary Standard requires covered entities to limit PHI disclosures to the minimum amount needed to accomplish the intended purpose [HIPAA 164.502(b)]. This standard applies every time a clinician, developer, or analyst pastes data into an AI tool. Pasting an entire patient chart into ChatGPT to generate a discharge summary violates 164.502(b) when the summary requires only the diagnosis, treatment plan, and follow-up instructions.

The HHS proposed HIPAA Security Rule update (January 2025) specifically requires entities using AI tools to include those tools in risk analysis and risk management activities [HHS OCR Proposed Rule 2025]. Regulators will examine training data provenance, PHI access within model workflows, and role-based controls. The Minimum Necessary Standard becomes the first compliance checkpoint for every AI integration.

How the 2025 Proposed Rules Change AI Risk Analysis

The proposed rule eliminates the distinction between “required” and “addressable” safeguards [HHS OCR Proposed Rule 2025]. Every security control becomes mandatory. For AI tools accessing ePHI, this means encryption in transit and at rest becomes non-negotiable, audit logging of AI interactions becomes mandatory, and access controls must restrict AI tool access to the specific ePHI categories needed for the task.

Organizations deploying AI before the rule finalizes face a retroactive compliance gap. Building these controls now prevents the remediation cost multiplier. McKinsey research shows retrofitting governance costs three times more than building it in from the start [McKinsey 2025].

The “Paste the Whole Chart” Problem

Field observations reveal a consistent pattern. Clinicians paste entire patient charts into AI tools because selective extraction takes longer. Developers paste complete JSON objects including patient_id, date_of_birth, and insurance_member_id because filtering the payload adds development time. Compliance teams discover these practices during audit, not during workflow design.

Technical controls prevent what policies alone do not. Data Loss Prevention (DLP) rules blocking PHI patterns at the network layer stop the leak before it reaches the AI vendor. API middleware stripping identifier fields from outbound requests enforces Minimum Necessary without relying on user behavior.

The Audit Fix: Enforce Minimum Necessary at the technical layer, not the policy layer.

  1. Deploy DLP rules on all AI tool endpoints (ChatGPT, Copilot, Claude, Gemini) to detect and block PHI patterns: SSNs, MRNs, dates of birth, insurance IDs [HIPAA 164.312(e)(1)].
  2. Build API middleware between your clinical systems and AI tools. Strip identifier fields from outbound payloads automatically.
  3. Restrict AI tool access to the minimum ePHI categories required for each use case. Document the access scope in your BAA [HIPAA 164.308(b)(1)].
  4. Log every AI interaction containing clinical data. Retain logs for six years per 164.530(j). Include AI tools in your quarterly access review cycle.

How Do You Build a PHI Containment Framework for AI Tools?

Step 1: BAA Before Access

No AI tool touches PHI without a signed Business Associate Agreement [HIPAA 164.308(b)(1)]. This applies to every tier: enterprise, team, and API. Free-tier AI tools (ChatGPT Free, Gemini Free, Claude Free) do not offer BAAs. Data entered into free tiers feeds model training. Using a free-tier AI tool with PHI constitutes a HIPAA violation regardless of de-identification efforts.

AI Vendor BAA Tier Required Training Opt-Out Status
OpenAI Enterprise / Team Yes (Enterprise) Active
Anthropic (Claude) Enterprise Yes (Enterprise) Active (Jan 2026)
Google (Gemini) Workspace Enterprise Yes Active
Microsoft (Copilot) Enterprise / M365 Yes Active
Otter.ai Business only Yes (Business) No BAA on free/pro

Step 2: DLP Controls on AI Endpoints

Data Loss Prevention rules enforce PHI containment at the network layer. Configure your DLP system to detect HIPAA identifier patterns in outbound traffic to AI tool domains. Block requests containing Social Security number formats, medical record number patterns, insurance member IDs, and date-of-birth fields. Major DLP platforms (Microsoft Purview, Palo Alto, Zscaler) include pre-built HIPAA classifiers.

Shadow AI compounds the DLP challenge. Over 80% of workers use unapproved AI tools [UpGuard 2025]. ECRI named AI chatbot misuse the number-one health technology hazard for 2026 [ECRI 2026]. DLP rules must cover not only sanctioned AI tools but also block access to unsanctioned AI endpoints entirely.

Step 3: The AI Tool Inventory Audit

Build and maintain a complete inventory of every AI tool accessing clinical data, clinical workflows, or clinical communications. Include sanctioned tools (enterprise AI platforms with BAAs) and discovered tools (shadow AI identified through network log analysis). Gartner reports 68% of healthcare organizations lack an AI tool inventory [Gartner 2025]. This inventory gap becomes the first finding in every HIPAA AI audit.

Map each tool against the HIPAA Security Rule requirements: BAA status, encryption controls, access restrictions, audit logging capability, and incident response provisions. Tools failing any requirement lose network access until remediation completes. Tools passing all requirements enter the approved AI governance registry with annual re-certification.

The Audit Fix: Build a three-layer PHI containment framework before your next audit cycle.

  1. Complete an AI tool inventory within 30 days. Cross-reference network logs against known AI tool domains. Flag every tool without a signed BAA [HIPAA 164.308(b)(1)].
  2. Deploy DLP rules on AI endpoints within 60 days. Start with the five highest-volume AI tools. Expand to full coverage within 90 days.
  3. Implement API middleware for clinical-to-AI data flows. Strip all 18 HIPAA identifiers plus quasi-identifiers (age ranges narrower than decade, ZIP codes below three-digit prefix) from outbound payloads.
  4. Schedule quarterly AI access reviews aligned with your existing user access review cycle. Document findings per 164.308(a)(1)(ii)(D).

The HIPAA Safe Harbor method was designed for paper records reviewed by human analysts. AI renders its protections insufficient. Organizations treating Safe Harbor compliance as permission to feed clinical data into AI tools face re-identification rates between 28% and 95%, depending on the dataset and the model. The containment framework requires three controls: BAAs on every AI tool, DLP enforcement at the network layer, and automated scrubbing of quasi-identifiers before any data reaches an AI endpoint. Sign the BAA first. Build the technical controls second. The order matters because the BAA creates the legal obligation for the vendor to protect what your controls miss.

Frequently Asked Questions

Does removing all 18 Safe Harbor identifiers make data safe for AI tools?

Removing the 18 identifiers satisfies HIPAA Safe Harbor requirements [HIPAA 164.514(b)(2)]. It does not eliminate re-identification risk. Machine learning algorithms re-identify individuals from Safe Harbor-compliant datasets at rates between 28% and 95% depending on remaining quasi-identifiers. Treat any dataset entering an AI tool as requiring both Safe Harbor compliance and a signed BAA.

Does OpenAI sign a BAA for the free version of ChatGPT?

OpenAI does not offer BAAs for the free or Plus tiers of ChatGPT, and data entered into these tiers feeds model training, making any PHI entry a HIPAA violation under 164.308(b)(1). Free and Plus tier data feeds model training. Entering PHI into the free version constitutes a HIPAA violation under 164.308(b)(1). The Enterprise tier disables training on customer data and includes BAA provisions.

Is a patient’s voiceprint PHI even if they never state their name?

Voiceprints qualify as biometric identifiers under Safe Harbor identifier #16, meaning any voice recording containing clinical content constitutes ePHI regardless of whether the patient identifies themselves verbally [HIPAA 164.514(b)(2)(i)(P)]. A voice recording containing clinical content constitutes ePHI regardless of whether the patient identifies themselves verbally. AI scribes recording clinical sessions capture PHI through the voiceprint alone.

Does pseudonymization (replacing names with codes) count as de-identification?

Only if your organization does not retain the mapping key. If a lookup table linking Patient_X99 to the patient’s identity exists anywhere in your systems, Patient_X99 remains PHI [HIPAA 164.514(c)]. The pseudonym carries the full regulatory weight of the original identifier because re-identification remains possible through the key.

What is the Mosaic Effect in healthcare AI?

The Mosaic Effect occurs when individually non-identifying data points combine to uniquely identify a person. ZIP code + admission date + diagnosis code + age range produces a unique combination for most individuals. AI amplifies this risk by cross-referencing clinical data against billions of public records in its training data, achieving re-identification without accessing any single identifying field [IAPP 2025].

Does the Minimum Necessary Standard apply to AI prompts?

HIPAA 164.502(b) requires limiting PHI disclosures to the minimum necessary for the intended purpose. Pasting an entire patient chart into an AI tool to generate a two-paragraph summary violates this standard. Restrict AI inputs to the specific data fields required for the output.

Are IP addresses captured by AI chatbots considered PHI?

IP addresses are Safe Harbor identifier #15 under HIPAA, and every web-based AI tool logs the originating IP address on each API call [HIPAA 164.514(b)(2)(i)(O)]. Every web-based AI tool logs the originating IP address on each API call. If the IP address links to a healthcare facility or patient portal session, it becomes PHI associated with the clinical data transmitted in the same request.

Should AI tools be included in HIPAA risk assessments?

The HHS proposed HIPAA Security Rule update (January 2025) explicitly requires entities using AI tools to include them in risk analysis activities [HHS OCR Proposed Rule 2025]. This applies to commercial AI platforms, custom-built models, and AI-powered features embedded in existing software. Add every AI tool to your technology asset inventory and risk assessment scope.

Get The Authority Brief

Weekly compliance intelligence for security leaders and technology executives. Frameworks decoded. Audit strategies explained. Regulatory updates analyzed.

Discipline in preparation. Confidence in the room.

Josef Kamara, CPA, CISSP, CISA, Security+
Josef Kamara
Josef Kamara
CPA · CISSP · CISA · Security+

Former KPMG and BDO. Senior manager over third-party risk attestations and IT audits at a top-five global firm, and former technology risk leader directing the IT audit function at a Fortune 500 medical technology company. Advises growth-stage SaaS companies on SOC 2, HIPAA, and AI governance certifications.