We are integrating AI across several workstreams on a federal health research information platform.

Publication discovery — using LLMs to surface relevant PubMed research, reducing manual literature review time and improving coverage across a high-volume publication landscape.

LLM comparative evaluations — running structured benchmarks across models to assess quality, consistency, and cost for specific content tasks on the platform. Evaluations are task-specific rather than general — we score against real outputs the platform needs to produce.

AI-assisted development workflows — incorporating AI tooling into the engineering workflow for code review, documentation, and implementation acceleration. All outputs are reviewed by the project team before use.

AI-powered CMS module vetting — evaluating third-party AI-powered Drupal modules before integration. Assessment criteria include data handling, output reliability, and compatibility with federal security requirements.

Communications and dissemination — using AI tools to support drafting and refinement of materials, with project team review and approval on all final outputs.