1. Purpose and Scope

Purpose

This Standard Operating Procedure establishes the framework, governance, and workflow for AI-assisted tool development by subject-matter experts within Department of War organizations. Expert-Driven Development enables domain experts to build, document, and sustain institutional tools using AI as a development accelerator -- without requiring a background in software engineering, contracting, or formal IT acquisition.

EDD IS / EDD ISN'T

EDD IS

  • A framework for building institutional tools using AI
  • A way to turn subject-matter expertise into working software
  • A governance model for responsible AI-assisted development
  • A training pipeline that produces AI-fluent builders
  • A documentation standard that ensures tool continuity
  • Compatible with existing DoW cybersecurity and privacy requirements

EDD ISN'T

  • A replacement for professional software development or IT acquisition
  • An authorization to bypass cybersecurity or privacy controls
  • A way to build systems that handle classified data
  • A shortcut that eliminates the need for documentation
  • A single AI tool or platform
  • Limited to any specific programming language or technology stack

Scope

This SOP covers the complete lifecycle of AI-assisted tool development: from problem identification and compliance review through prototyping, testing, documentation, deployment, and sustainment. It applies to tools built on approved platforms using approved AI assistants that handle unclassified data only.

Applicability

This SOP applies to all personnel who develop, review, supervise, or maintain tools built using the Expert-Driven Development framework. It is designed for adoption at the unit level and can be tailored to organization-specific requirements while maintaining compliance with DoW-wide cybersecurity and privacy directives.

Back to top

2. The Four-Layer Framework

Expert-Driven Development operates through four interdependent layers. Each layer builds on the previous one, and all four must be present for the framework to produce sustainable results.

Layer 1: Direct Tool Development

Subject-matter experts use AI to build tools that solve real problems in their organizations. The expert provides domain knowledge -- the understanding of the problem, the workflow, the users, and the requirements. AI provides the development capability -- translating that domain knowledge into working code, interfaces, and automation.

Key Principle

The expert drives the development. AI is the accelerator, not the decision-maker. Every design choice, workflow decision, and feature prioritization comes from the person who understands the problem.

Layer 2: Process Liberation

By removing the dependency on contractors, acquisition timelines, and IT development queues, EDD liberates the development process. Personnel can identify a problem and begin building a solution within the same week -- not the same fiscal year. This layer addresses the structural bottleneck that prevents organizations from responding to their own needs.

Layer 3: Capability Cultivation

Each build develops the builder. The five-course training curriculum and the 201-level skills framework cultivate six core competencies: Context Assembly, Quality Judgment, Task Decomposition, Iterative Refinement, Workflow Integration, and Frontier Recognition. Personnel who complete EDD projects gain transferable skills in AI-assisted development, systems thinking, user-centered design, and technical documentation. These capabilities persist after the specific tool is deployed and compound across subsequent projects.

Layer 4: Documentation and Replication

Every tool built under EDD is documented to a standard that enables another person to understand, maintain, modify, and replicate the tool without the original builder present. This layer ensures that institutional knowledge does not walk out the door when personnel transfer. Documentation is not an afterthought -- it is a core deliverable of every project.

Back to top

3. The Creator Mindset

EDD requires a fundamental shift in how personnel interact with AI tools. Most users operate as consumers -- they use AI to get answers, summarize text, or draft emails. EDD trains personnel to operate as creators -- using AI to build things that other people use.

Consumer vs Creator

Comparison of Consumer and Creator approaches to AI
Dimension Consumer Creator
Output Answers for yourself Tools for others
Interaction Single prompt, single response Iterative conversation toward a build
Scope Personal productivity Institutional capability
Duration One-time use Persistent, maintained artifact
Accountability None -- private use Documented, reviewed, and sustained
Knowledge transfer None Documentation package enables replication

The Artifact Test

The simplest way to determine whether someone is operating as a consumer or a creator:

Does your AI interaction produce an artifact that someone else can use?

If the answer is yes -- a tool, an application, a system, a workflow -- you are creating. If the answer is no -- a summary, an email draft, a personal answer -- you are consuming. Both are valid. EDD focuses on creating.

Back to top

4. Compliance and Security

All development under this SOP must comply with existing DoW cybersecurity, privacy, OPSEC, and records management requirements. AI-assisted development does not create new exceptions to these requirements -- it operates within them.

Data Classification

EDD tools handle unclassified data only. The following data types are prohibited from use in AI-assisted development:

Prohibited Data

  • Classified information (any level)
  • Controlled Unclassified Information (CUI) unless the platform is CUI-authorized
  • Personally Identifiable Information (PII) unless a Privacy Impact Assessment has been completed
  • Protected Health Information (PHI)
  • Operational plans, force movement data, or operational intelligence products

Privacy Requirements

Tools that collect, store, process, or display information about individuals must comply with DoW privacy directives. The following apply:

Privacy Impact Assessment (PIA)

Required for any tool that collects or maintains PII. A PIA threshold analysis must be completed during the Compliance Review phase (Phase 2) to determine whether a full PIA is required.

System of Records Notice (SORN)

If the tool creates a new system of records retrievable by a personal identifier, a SORN must be published in the Federal Register before the tool goes into production use.

Appeals and Redress

Any tool that makes decisions affecting individuals must include a mechanism for individuals to review, contest, and correct information about themselves.

Cybersecurity

Approved AI Tools

The following AI tools are approved for use in EDD development. Authorization levels indicate the scope of approved use:

Approved AI tools and their authorization levels
Tool Authorization Level Notes
M365 Copilot Enterprise-authorized Available through existing DoW M365 licensing; data stays within tenant
Azure OpenAI Enterprise-authorized Requires Azure Government subscription; FedRAMP High authorized
Copilot Studio Enterprise-authorized Low-code bot builder within M365 ecosystem
Gemini Approved with restrictions Unclassified use only; do not input PII or CUI
C3 AI Enterprise-authorized DoW Enterprise AI platform; available through enterprise agreement

Approved Platforms

Tools built under EDD must be deployed on approved platforms:

Approved deployment platforms
Platform Use Case Authorization
Microsoft Power Platform Low-code applications, automation, dashboards Enterprise-authorized within DoW M365
SharePoint Online Web-based tools, document management, portals Enterprise-authorized within DoW M365
Azure Government Cloud-hosted applications, APIs, databases FedRAMP High; requires subscription
GitHub (public) Open-source projects, static sites, documentation Unclassified public release only

When an ATO Is Required

An Authority to Operate (ATO) is required when a tool:

  • Connects to or operates on the DoW Information Network (DoWIN)
  • Processes, stores, or transmits CUI
  • Integrates with other authorized information systems
  • Operates outside the inherited authorization boundary of its host platform

Tools built entirely within the inherited authorization boundary of an approved platform (e.g., a Power App within DoW M365) typically do not require a separate ATO but must still comply with the platform's security controls.

Security Controls

All tools must implement security controls appropriate to their data sensitivity and deployment environment, including:

  • Authentication via CAC or enterprise identity provider
  • Role-based access control (RBAC) where applicable
  • Audit logging of user actions and data access
  • Input validation and output encoding
  • Encryption of data in transit (TLS 1.2+) and at rest

OPSEC

Developers must apply OPSEC principles throughout the development lifecycle. Do not include operational details, unit-specific procedures, force structure information, or vulnerability assessments in AI prompts, tool documentation, or publicly accessible code repositories. When in doubt, consult your organization's OPSEC officer before proceeding.

Records Management

Development artifacts (documentation packages, source code, deployment records) are official records and must be managed in accordance with DoW records management directives. Retain all project documentation for the duration specified by the applicable records schedule. The documentation package produced in Phase 5 serves as the primary record of each development effort.

Incident Response

If a security incident occurs involving an EDD tool (data breach, unauthorized access, malware, or compromise of an AI tool), personnel must:

  1. Immediately cease use of the affected tool
  2. Report the incident to the organization's Cybersecurity Office within 1 hour
  3. Preserve all logs, artifacts, and evidence related to the incident
  4. Notify the chain of command and the EDD Program Coordinator
  5. Cooperate with the incident response investigation
  6. Do not attempt to remediate the vulnerability without authorization
Back to top

5. Development Prerequisites

Required

  • Completed EDD training (minimum: AI Fluency Fundamentals)
  • Supervisor approval for the specific development project
  • Access to at least one approved AI tool
  • Access to at least one approved deployment platform
  • A defined problem that meets the criteria in Phase 1
  • Completed Problem Definition Template

Not Required

  • A background in software engineering or computer science
  • Prior programming experience
  • A contract, funding, or acquisition document
  • An IT service request or development ticket
  • Permission from a centralized IT authority (unless ATO is required)

Obtaining Access

Microsoft 365

Most DoW personnel already have M365 accounts. Power Platform access may require a license assignment through your organization's M365 administrator. Contact your local IT support to verify your license includes Power Apps, Power Automate, and Copilot Studio.

AI Tools

The following table provides recommendations for obtaining access to approved AI tools:

Recommendations for obtaining access to AI tools
Tool How to Obtain Access Recommendation
M365 Copilot Enterprise license assignment through M365 admin Start here if your organization has Copilot licenses
Azure OpenAI Azure Government subscription; request through cloud team Best for custom API-driven development
Copilot Studio Included with Power Platform licensing Best for building conversational AI assistants
Gemini Web access at gemini.google.com (unclassified only) Good general-purpose coding assistant
C3 AI Enterprise agreement; contact your AI/ML program office Best for data-heavy enterprise applications
Back to top

6. Development Workflow

The EDD development workflow consists of eight phases. Each phase has defined inputs, outputs, and time estimates. Phases are sequential, but iteration between adjacent phases is expected and encouraged.

Problem Definition

2-4 hours

Identify and document the problem you intend to solve. Complete the Problem Definition Template. Define the current state, desired state, affected users, and success criteria. A well-defined problem is the single strongest predictor of a successful build.

Output: Completed Problem Definition Template.

Compliance Review

1-4 hours

Review the proposed tool against compliance requirements. Complete the Compliance Checklist. Determine data classification, privacy requirements, and whether an ATO is needed. Identify the appropriate deployment platform. Obtain supervisor approval to proceed.

Output: Completed Compliance Checklist with supervisor signature.

Rapid Prototyping

8-20 hours

Build the tool using AI-assisted development. Start with the minimum viable feature set identified in Phase 1. Use iterative prompting to develop, test, and refine functionality. Maintain a development journal documenting key decisions, prompts that worked, and problems encountered.

Output: Working prototype with core functionality.

User Testing

4-8 hours

Put the prototype in front of actual users. Observe how they interact with it. Collect feedback on usability, functionality, and missing features. Iterate on the prototype based on user feedback. A minimum of three users should test the tool before proceeding to documentation.

Output: User-tested prototype with documented feedback and revisions.

Documentation

20-40 hrs (first) / 8-16 hrs (subsequent)

Produce the documentation package. This is the most critical phase for institutional value. The documentation must enable someone who has never seen the tool to understand what it does, how it works, how to maintain it, how to modify it, and how to replicate it. First-time builders should expect to spend significantly more time on documentation; the investment decreases with experience.

Output: Complete documentation package.

QA Review

2-4 hours

A designated QA Reviewer examines the tool and its documentation against the QA Checklist. The reviewer verifies functionality, security compliance, documentation completeness, CDAO GenAI RAI Toolkit alignment, and readiness for deployment. The reviewer must not be the same person who built the tool.

Output: Completed QA Checklist with reviewer sign-off or remediation requirements.

Deployment

1-2 hours

Deploy the tool to its production environment. Register the tool in the Tool Registry. Communicate availability to intended users. Establish the sustainment plan, including who is responsible for ongoing maintenance and how issues will be reported and resolved.

Output: Deployed tool, completed Tool Registry entry, and sustainment plan.

Sustainment

Ongoing

Maintain the tool in its operational environment. Monitor usage metrics, address user-reported issues, apply updates as needed, and update documentation to reflect changes. Plan for turnover: when the original builder departs, the documentation package must be sufficient for a successor to assume maintenance responsibility.

Output: Operational tool with current documentation and identified maintainer.

Back to top

7. Roles and Responsibilities

Developer

The individual who builds the tool. Responsible for completing all eight phases of the development workflow, maintaining compliance with this SOP, producing the documentation package, and sustaining the tool through its operational lifecycle. Any personnel member who has completed the required training and received supervisor approval may serve as a developer.

QA Reviewer

The individual who reviews the tool and documentation in Phase 6. The QA Reviewer must be a different person from the developer. Designation criteria:

  • Has completed EDD training (minimum: AI Fluency Fundamentals)
  • Has built at least one tool under EDD, or has been designated by the Program Coordinator
  • Is familiar with the compliance requirements in Section 4
  • Has no conflict of interest with the tool being reviewed

Supervisor

The developer's direct supervisor. Responsible for approving the development project, allocating duty time for development, and ensuring the tool meets organizational needs. Supervisors are required to complete the Supervisor Orientation before approving their first EDD project. This orientation ensures supervisors understand the framework, can evaluate proposals, and know what good AI-assisted output looks like.

Program Coordinator

The individual responsible for managing the EDD program at the organizational level. Maintains the Tool Registry, designates QA Reviewers, coordinates training, tracks program metrics, and reports program status to leadership. The Program Coordinator may also serve as a developer or QA Reviewer.

Cybersecurity Office

The organization's cybersecurity team. Provides guidance on security controls, reviews ATO requirements, responds to security incidents involving EDD tools, and validates that tools comply with cybersecurity directives. Consulted during Phase 2 (Compliance Review) for any tool that connects to the network or processes sensitive data.

Privacy Officer

The organization's designated privacy official. Reviews PIA threshold analyses, advises on SORN requirements, and ensures tools that handle PII comply with DoW privacy directives. Consulted during Phase 2 for any tool that collects, stores, or displays information about individuals.

Back to top

8. Training Requirements

The v5 curriculum comprises five courses organized by audience and prerequisite chain.

Required

The following course is the universal prerequisite and must be completed before beginning any EDD activity:

Required training course
Course Audience Duration Prerequisite
AI Fluency Fundamentals All personnel 2 hours None (universal prerequisite)

The following courses are recommended/elective for personnel who will build tools:

Recommended training courses for builders
Course Audience Duration Prerequisite
Builder Orientation Aspiring builders 2 hours AI Fluency Fundamentals
Platform Training Builders 4 hours Builder Orientation
Advanced Workshop Experienced builders 4 hours At least one deployed tool

Recommended for Leadership

The following course is recommended for supervisors and leadership personnel:

Recommended training course for leadership
Course Audience Duration Prerequisite
Supervisor Orientation Supervisors approving EDD projects 30 minutes None

Six 201-Level Skills

The training curriculum develops six core competencies defined in the 201-level skills framework. These skills represent the capabilities that separate effective AI-assisted builders from casual AI users:

  1. Context Assembly -- The ability to gather and structure the right information so AI can produce useful output.
  2. Quality Judgment -- The ability to evaluate AI-generated output for correctness, completeness, and fitness for purpose.
  3. Task Decomposition -- The ability to break complex problems into discrete, AI-addressable tasks.
  4. Iterative Refinement -- The ability to systematically improve AI output through structured feedback cycles.
  5. Workflow Integration -- The ability to embed AI-assisted processes into existing organizational workflows.
  6. Frontier Recognition -- The ability to identify where AI capability boundaries lie and adjust approach accordingly.

Instructor Certification

Instructors for each course must meet the following certification requirements:

Instructor certification requirements by course
Course Requirements
AI Fluency Fundamentals Completed course AND at least one builder course
Builder Orientation Completed all builder courses AND deployed at least one tool
Platform Training Completed all builder courses, deployed at least one tool, plus Power Platform proficiency
Advanced Workshop Completed all builder courses, deployed at least one tool, plus served as QA reviewer for 2+ tools
Supervisor Orientation Any qualified AI Fluency Fundamentals instructor

Certification process: Complete course as student → Build/deploy tool → Shadow instructor → Co-teach → Certification from Program Coordinator.

Back to top

9. Metrics and Assessment

Metrics serve two purposes: demonstrating the value of individual tools and assessing the health of the overall EDD program. All metrics should be collected consistently and reported to leadership quarterly.

Tool-Level Metrics

Collected for each deployed tool:

Tool-level metrics
Metric What It Measures Collection Method
Active users Adoption and utility Platform analytics or usage logs
Time saved per use Efficiency gain User survey or workflow comparison
Error rate reduction Quality improvement Before/after error tracking
Build time Development efficiency Developer-reported hours
Development cost Cost avoidance Hours multiplied by composite labor rate
User satisfaction Quality of the solution Post-deployment survey

Program-Level Metrics

Collected across all EDD activity within the organization:

Program-level metrics
Metric What It Measures Target
Number of trained personnel Pipeline capacity Growing quarter-over-quarter
Number of deployed tools Program output Growing quarter-over-quarter
Completion rate (started vs deployed) Process effectiveness >70% of started projects reach deployment
Documentation compliance rate Quality assurance 100% of deployed tools have complete packages
Total estimated cost avoidance Return on investment Reported quarterly to leadership
Security incidents Risk management Zero incidents per quarter
Back to top

10. Scaling and Adoption

90-Day Pilot Model

Organizations adopting EDD should begin with a 90-day pilot. The pilot provides a controlled environment to validate the framework, train initial personnel, produce the first tools, and establish organizational processes before expanding.

90-day pilot timeline
Phase Timeline Activities
Setup Days 1-14 Designate Program Coordinator, identify 2-4 pilot developers, conduct training, verify platform access
First builds Days 15-60 Pilot developers complete Phases 1-5 on their first projects with coaching support
QA and deployment Days 61-75 QA reviews, remediation, deployment, and Tool Registry entries
Assessment Days 76-90 Collect metrics, brief leadership, decide on expansion

Assessment guidance: The Assessment phase should include a review of individual tool and program metrics with a candid recommendation to leadership. Full compliance across all criteria on day one is rare; the recommendation should identify what is working, what needs remediation, and what adjustments are required before expansion. Leadership decisions should be based on trajectory and demonstrated value, not perfection.

Success / Failure Criteria

Success Criteria

  • At least 2 tools deployed and in active use
  • All deployed tools have complete documentation packages
  • Zero security incidents during the pilot
  • Measurable time savings or error reduction demonstrated
  • Leadership endorsement for expansion

Failure Indicators

  • No tools reach deployment within 90 days
  • Security or compliance violations occur
  • Developers cannot obtain platform access
  • Leadership does not allocate duty time for development
  • Documentation requirements are waived or ignored

Tool Registry

The Tool Registry is a centralized record of all tools built under EDD within the organization. Every deployed tool must have an entry. The registry tracks:

  • Tool name and description
  • Developer and current maintainer
  • Deployment platform and URL
  • Data classification
  • User count and status (active, deprecated, retired)
  • Documentation package location
  • Date deployed and last updated

The Program Coordinator maintains the Tool Registry and reviews it monthly to identify tools that need updates, lack a maintainer, or should be retired.

Expansion Model

After a successful pilot, expansion follows a train-the-trainer model:

  1. Pilot developers become QA Reviewers and mentors for the next cohort
  2. Each cohort is 2-4 new developers with defined projects
  3. New cohorts begin every 60-90 days
  4. The Program Coordinator scales training, QA capacity, and registry management as the program grows
  5. Cross-organizational sharing is encouraged through the Tool Registry and documentation standards
Back to top

11. References

Policy and Regulatory References

  • DoW Directive 8140.01 -- Cyberspace Workforce Management
  • DoW Instruction 5000.87 -- Operation of the Software Acquisition Pathway
  • DoW Instruction 5200.48 -- Controlled Unclassified Information (CUI)
  • DoW Instruction 5400.11 -- DoW Privacy and Civil Liberties Programs
  • DoW Instruction 8500.01 -- Cybersecurity
  • DoW Instruction 8510.01 -- Risk Management Framework (RMF) for DoW Systems
  • DoW Instruction 8582.01 -- Security of Non-DoW Information Systems
  • DoW Directive 3020.26 -- Department of War Continuity Policy
  • DoWM 5200.01 -- DoW Information Security Program
  • SECNAV M-5210.1 -- Department of the Navy Records Management Program
  • NIST SP 800-53 Rev. 5 -- Security and Privacy Controls for Information Systems and Organizations
  • OMB Circular A-130 -- Managing Information as a Strategic Resource
  • Executive Order 14179 -- Removing Barriers to American Leadership in Artificial Intelligence (January 2025)
  • DoW AI Strategy, January 2026

Research Foundation

  • Dell'Acqua, F., et al. (2023). "Navigating the Jagged Technological Frontier." Harvard Business School / BCG. Study of 758 consultants demonstrating the jagged frontier of AI capability.
  • Brynjolfsson, E., Li, D., & Raymond, L. (2025). "Generative AI at Work." Stanford / MIT. Study of 5,172 customer-support agents demonstrating skill-leveling effects of AI assistance.
  • Mollick, E. (2026). "Management as AI Superpower." Wharton School. Research on the delegation equation and management capabilities in AI-assisted work.
  • OpenAI (2025). "GDPval: Measuring Economic Value of AI." Study of 1,320 tasks establishing expert parity benchmarks.
  • UK Government Digital Services (2025). AI-assisted development deployment across 20,000 employees, demonstrating government-scale deployment patterns.
  • Microsoft Work Trend Index. Study of 300,000+ employees identifying 80% tool abandonment rate without structured training.
  • Stanford Digital Economy Lab. Research on the apprenticeship crisis and implications for AI-assisted skill development.
Back to top

12. Appendices

The following templates are used throughout the EDD development workflow. Download each template and complete it during the appropriate phase.

Downloadable templates for EDD development
Template Used In Download
Problem Definition Template Phase 1: Problem Definition problem-definition.md
QA Checklist Phase 6: QA Review qa-checklist.md
Tool Registry Entry Phase 7: Deployment tool-registry-entry.md
PIA Threshold Analysis Phase 2: Compliance Review documentation-package-outline.md
Compliance Checklist Phase 2: Compliance Review development-journal.md
Back to top