Handling Sensitive Data Safely
One of the highest-risk mistakes nonprofit leaders make with AI is sharing sensitive data—health information, financial data, personally identifiable information—with consumer-grade AI tools. A program director innocently uses ChatGPT to de-identify a client case study but doesn't fully remove identifying information. Accidentally, she's disclosed confidential client data to an AI system she doesn't control. Similarly, a grant writer uses ChatGPT to summarize confidential program data, again unknowingly exposing protected information. These scenarios happen regularly.
This lesson provides practical guidance for protecting sensitive data when using AI tools. We'll explore data sensitivity classifications, safe and unsafe uses of different tool types, de-identification strategies, compliance requirements, and incident response procedures.
Not all data requires the same level of protection. Understanding data sensitivity helps you make appropriate decisions about which tools can safely be used with different types of information.
This data requires maximum protection and strict limitations on AI tool use.
This data requires protection but can be used with some AI tools, particularly with de-identification or appropriate security measures.
This data can generally be safely shared with consumer AI tools without special restrictions.
Different categories of AI tools have different security and privacy characteristics. Understanding these differences helps you make appropriate choices about which tools to use with different data types.
Consumer-grade AI tools transmit data to third-party servers, train on provided data, and may retain that data indefinitely. They're not appropriate for sensitive data. However, using them with lower-sensitivity data and appropriately de-identified moderately sensitive data is fine.
Safe: "Help me brainstorm grant writing strategies for health nonprofits" or "Summarize trends in youth homelessness based on published statistics"
Unsafe: Sharing actual client stories that contain identifying information, sharing program participant lists, sharing health histories of actual clients
Enterprise AI tools typically offer stronger data protection, security certifications (SOC 2, ISO 27001), and commitments not to train on your data. They're appropriate for moderately sensitive data and most use cases with highly sensitive data, provided you have data processing agreements in place. However, some highly sensitive uses may still require additional protections.
AI tools built specifically for healthcare or education and certified as HIPAA or FERPA compliant are appropriate for the data they're designed for. These tools have built-in protections, security certifications, and contractual obligations appropriate for highly sensitive data.
De-identification—removing or altering information that could identify individuals—allows you to use lower-sensitivity versions of sensitive data with consumer AI tools. However, de-identification is complex. Removing obvious identifying information isn't always sufficient if other data could still reveal identity.
Always remove names, addresses, phone numbers, email addresses, dates of birth, social security numbers, and other directly identifying information before sharing data with AI tools.
Even without direct identifiers, combinations of demographic information can identify individuals. A case study mentioning "a 34-year-old transgender woman with HIV in rural Montana" might re-identify the individual within that community. Removing or generalizing quasi-identifiers prevents re-identification.
Instead of specific ages, use age ranges. Instead of specific locations, use regions. Instead of specific diagnoses, use diagnosis categories. This reduces re-identification risk while preserving utility for analysis.
If a program has few clients with a particular characteristic (e.g., only one transgender client), that characteristic becomes identifying. Suppress or generalize such rare values.
Don't assume de-identification succeeds. After removing identifiers, review the data to ensure you can't reasonably re-identify individuals. When in doubt, remove more information rather than less.
Legal and regulatory frameworks create compliance requirements for data protection with AI tools. Key regulations include:
If your organization handles protected health information, HIPAA creates strict requirements. You cannot share protected health information with AI tools unless you have a Business Associate Agreement (BAA) in place. Most consumer AI tools don't offer BAAs. Enterprise healthcare AI tools typically do.
If your organization works with education records, FERPA restricts disclosure. You cannot share education records with third-party AI tools without appropriate agreements in place. School-focused AI tools typically have FERPA-compliant data agreements.
If your organization serves anyone in the EU or processes EU resident data, GDPR applies. GDPR creates strict requirements around data processing, consent, and accountability. Using consumer AI tools with EU resident data without appropriate data processing agreements violates GDPR.
States including California (CCPA), Virginia (VCDPA), and others have enacted privacy laws. These laws require organizations to disclose data uses and sometimes require consent for data sharing. Compliance varies by law and by how your organization uses data.
Your organization should establish clear data protection policies specific to AI tool use. Policies should cover:
Despite precautions, accidental data disclosures sometimes occur. Effective incident response minimizes harm and demonstrates accountability.
Regulators are generally more forgiving of incidents that organizations discover and manage responsibly than of incidents they learn about from other sources. Transparent, thorough incident response demonstrates accountability and reduces regulatory penalties.
Policies only work if staff understand them and internalize data protection as a value. Building a strong data protection culture requires:
Learn how to adapt AI governance policy templates for your organization's context and size.
Start Lesson 7