30 minutes • Create reports and documentation that demonstrate responsible AI use
Some funders explicitly ask about AI use in grant applications or reports. Others don't but might appreciate transparency. Proactive reporting demonstrates nothing to hide and positions your organization as sophisticated. Comprehensive reporting creates documentation that protects you if questions arise later.
Create annual report of AI use: How many grants used AI? What was AI used for? What results? Annual reports demonstrate scale and intentionality. Example format: "2024 AI Use Summary: 60 grants used AI-assisted research. 80 grants used AI for proposal drafting. Acceptance rate: 32%. No quality concerns reported. Staff completed AI ethics training." Clear, concise annual reporting is professional.
Rather than just statistics, explain what AI is actually used for. "We use Claude to research funder priorities and analyze past awards. This accelerates our research process while our team verifies all findings. We use Claude for initial proposal drafting, which our experienced writers then refine to match our voice and ensure accuracy." Narrative explains both use and safeguards.
Quantify benefits if possible. "AI-assisted research reduced research time by 25% while maintaining quality. Staff using AI have reported higher satisfaction due to reduced mechanical work and more time for strategy." Benefits demonstrate value. Quantified benefits are more credible than general claims.
Grant reports should mention AI use if relevant to the story being told. "Our team used AI-assisted research to accelerate our needs assessment, enabling faster response to emerging needs." Mention if it strengthens the narrative. Don't force it in if it's not relevant.
Mention AI casually, as you would any tool. "Using modern research tools including AI analysis, we identified..." Not defensive or apologetic. Not overstated. Natural, professional tone. AI is a tool you use competently, like databases or spreadsheets.
If funder report asks "Did you use AI?" answer directly and honestly. "Yes, we used AI for research and initial drafting. All content was reviewed by our team for accuracy and alignment with our approach." Honesty and openness demonstrate responsibility.
Keep records showing: which grants used AI, what specifically, how output was verified, who approved. Records aren't oppressive—they're protective. If auditors ask "How do we know your AI use was appropriate?" records provide answers. Records demonstrate systematic, thoughtful approach.
Document your AI policies and governance. How do you decide when to use AI? How do you ensure quality? How do you verify accuracy? Documented policies show systematic governance. Lack of documentation suggests ad-hoc approach, which undermines credibility.
Document staff training on AI use. Who trained? When? What was covered? Documentation demonstrates commitment to responsible use. "All staff using AI in grants completed our AI competency training" is credible. "Staff use AI" without training documentation is concerning.
Maintain: AI use policy document. Training completion records. Monthly governance meeting notes. Documented quality assurance processes. Records of tool selections and justifications. Incident logs (if any AI use caused problems and how addressed). This documentation is protective. If questions arise, you have comprehensive answers.
Visual displays of AI metrics are compelling. Dashboard showing: proposals using AI monthly, acceptance rates, funder feedback, staff training completion. Dashboards make data accessible and patterns visible. Funders reviewing your dashboard see systematic, measured approach.
Consider: percentage of proposals using AI, acceptance rates before/after AI adoption, quality scores, research time savings, cycle time improvements, staff satisfaction with AI tools, training completion rates. Choose metrics that demonstrate value and responsible use.
Share metrics with leadership, staff, and key funders. Transparency about metrics demonstrates confidence. Hiding metrics suggests something to hide. Open metrics demonstrate nothing to hide and systematic improvement culture.
Government funders increasingly ask about AI use. Answer their specific questions accurately. Some government grants explicitly permit AI; others restrict it. Know requirements. Comply fully. Government audits are serious; documentation is essential.
Foundation reports might briefly mention AI if relevant: "Our evaluation process was enhanced by AI-assisted data analysis, enabling more rapid insight generation." Mention if helpful. Most foundations won't require AI documentation, but mentioning shows sophistication.
Corporate funders might be particularly interested in AI use as innovation. "We're leveraging AI to maximize impact of your investment" appeals to corporate values. Corporations often want to know about innovation and efficiency gains from their grants.
If you didn't measure metrics before AI adoption, that's okay. Begin measuring now. "Going forward, we're tracking AI use and impact" shows commitment to measurement even if you lack historical comparison.
If AI adoption improved some metrics but not others, report honestly. "AI use improved research speed and proposal output. Quality metrics remained consistent with pre-AI levels." Honesty about mixed results is better than only reporting positives. Sophisticated organizations know results are complex.
If AI use caused problems (quality issue discovered, funder concern), report directly in your next communication. "We discovered an accuracy issue in AI-generated research on Grant X. We addressed it immediately. Here's what we're doing to prevent recurrence." Transparency about problems actually builds trust if handled well.
Consistent, honest reporting over time builds credibility. Year 1: "Here's our initial AI use and results." Year 2: "Here's how use has evolved and what we've learned." Year 3: "Here's our mature AI integration approach." Multi-year reporting shows evolution, learning, and sophistication.
Next, we'll explore customizing AI disclosure strategies for different types of funders.
Continue to Next Lesson