AI Risk

Developer vs. Deployer vs. Operator: Role-Specific Obligations Under NIST AI 600-1

April 25, 2026 Rebecca Leung
Table of Contents

Your compliance team is building an AI governance program. The CISO wants to know what you own. The vendor says it’s their problem. Your bank partner is asking for a completed questionnaire. And you’re sitting there trying to figure out: who’s actually responsible for what?

This is the core question NIST AI 600-1 tries to answer — and it answers it by splitting the AI actor landscape into distinct roles with distinct obligations. Most organizations are operating in more than one of these roles simultaneously, which is where the real complexity lives.

TL;DR

  • NIST AI 600-1 defines AI developers, deployers, and operators as distinct actors with different risk management obligations
  • Developers own model-level risks: training data, safety testing, documentation, output labeling
  • Deployers own deployment-context risks: fit-for-purpose evaluation, user disclosures, post-deployment monitoring
  • Operators run daily oversight: output monitoring, incident escalation, human-in-the-loop controls
  • Most fintechs and financial institutions are functioning as deployers — but fine-tuning a model triggers developer obligations

Why the Role Distinction Matters

NIST AI 600-1 isn’t just a checklist. It’s an accountability framework. Published in July 2024, it adapts the broader NIST AI RMF 1.0 to the specific risk profile of generative AI — hallucinations, training data poisoning, content provenance issues, harmful output, dual-use capability, and supply chain opacity.

The framework explicitly states: “Not all actions apply to all AI actors. The applicability of suggested actions to relevant AI actors should be determined based on organizational considerations and their unique uses of GAI systems.”

That’s a legal and operational escape hatch — but it’s also a trap. If you haven’t mapped yourself to an actor type, you can’t claim the right obligations apply (or don’t apply) to you. Examiners asking about your AI governance program are going to want to see that you’ve done this mapping.

Here’s the core framework:

RolePrimary PhaseCore Responsibility
AI DeveloperDesign, Development, TrainingModel architecture, training data governance, safety testing, output labeling, documentation
AI DeployerDeployment, IntegrationFit-for-purpose evaluation, user disclosures, regulatory compliance, monitoring setup
AI OperatorOperate and MonitorDaily output review, incident identification, human oversight, escalation

These aren’t mutually exclusive — they describe phases of the AI lifecycle and the actors responsible at each phase. A mid-size bank using a commercial GenAI API for document summarization is almost certainly a deployer and an operator, but not a developer. A fintech that fine-tunes GPT-4 on its proprietary loan data and then deploys it for underwriting decisions? That’s both.

The AI Developer: What You Own When You Build

Under NIST AI 600-1, an AI developer is any organization that “designs, codes, or produces AI systems.” This includes foundation model builders (OpenAI, Anthropic, Google), companies that fine-tune open-source models, and teams that train custom models from scratch.

The key phrase from the framework: developers “may be better positioned to manage risks related to the development and training of AI products and labeling of AI outputs, and identifying the product’s limitations and mitigation efforts that were undertaken to address potential risks.”

Developer Core Obligations

Training Data Governance Developers are responsible for the provenance, quality, and representativeness of training data. Under AI 600-1’s risk categories, this maps directly to data poisoning risk, harmful bias at the data layer, and privacy violations from ingesting personal data without appropriate controls. At minimum, developers should maintain a training data inventory with source documentation, consent or licensing basis, and data quality assessments.

Pre-Deployment Safety Testing AI 600-1’s MEASURE function requires developers to run TEVV (Test, Evaluation, Verification, and Validation) before deployment. For GenAI systems, this includes adversarial prompt testing, red-teaming for harmful output generation, and bias evaluation across demographic dimensions. The framework ties directly to the AI 600-1 TEVV requirements we covered in a prior post — safety testing isn’t optional, and it happens at the model layer, not just the application layer.

Intended Use Documentation Developers must document intended use cases, known limitations, and performance bounds. This is what the model card concept captures — and what regulators reviewing AI systems increasingly expect to see. If you can’t produce documentation showing what your model was designed to do and what it wasn’t tested for, that’s an immediate exam finding.

Output Labeling and Transparency AI 600-1 specifically addresses the need to identify and label AI-generated content. For GenAI developers, this means building labeling mechanisms (provenance metadata, watermarking, disclosure flags) into the system — not leaving it entirely to deployers to figure out after the fact.

Limitation Disclosure to Deployers Here’s the supply chain obligation that often gets missed: developers have a responsibility to communicate known model limitations and mitigation measures to downstream deployers. If you’re building a foundation model and you know it hallucinates in certain domains, that’s information that must flow to anyone using your model. The failure to disclose known limitations to deployers is a governance gap that NIST explicitly flags.

When a Deployer Becomes a Developer

This is the nuance that trips up most organizations. NIST AI 600-1 draws a distinction: “A deployer should only be considered a developer if it is changing the model in a transformative and extensive way.”

The practical interpretation:

ModificationRole Classification
System prompt engineering onlyDeployer
Retrieval-augmented generation (RAG)Deployer
Output filtering / guardrailsDeployer
Light fine-tuning (tone, style)Typically deployer
Domain-specific fine-tuning on proprietary dataDeveloper
Full model retrainingDeveloper
Distillation or model compressionDeveloper

If you’re doing domain-specific fine-tuning on your bank’s proprietary loan portfolio data, you’re a developer for that modified model — and you inherit the training data governance, safety testing, and limitation disclosure obligations that come with it.

The AI Deployer: What You Own When You Put It in Production

Most financial institutions and fintechs using commercial AI products are deployers. You’re taking a model that someone else built, integrating it into your environment, and using it to make or support business decisions.

Deployers “manage risks related to post-deployment evaluations and product monitoring and transparency notices for impacted individuals regarding the AI system’s use.” That’s a wide mandate.

Deployer Core Obligations

Fitness-for-Purpose Evaluation Before deploying any GenAI system, the deployer must independently assess whether the model is appropriate for the intended use case — in the deployment context, not the developer’s benchmark context. A model that performs well on general document summarization may perform poorly on regulatory filings. A model cleared for internal knowledge base queries may not be appropriate for customer-facing underwriting support.

This evaluation should be documented. It’s what the OCC and Fed expect to see under model risk management — “validation should be independent” applies to AI systems too, and the deployer can’t simply point to the developer’s testing documentation and call it a day.

User Transparency and Disclosure AI 600-1 requires deployers to implement transparency notices for impacted individuals. For a financial institution, this means customers who interact with AI-driven systems should know they’re doing so, and should understand the implications for decisions affecting them. This connects directly to ECOA adverse action notice requirements — if AI is influencing a credit decision, the disclosure has to be accurate about why.

Third-Party Risk Management for Developer Relationships Deployers who rely on commercial API providers (OpenAI, Anthropic, Google, Cohere, etc.) are responsible for due diligence on those developer relationships. GOVERN 6.1 under AI 600-1 requires “policies and procedures in place that address AI risks associated with third parties, including risks of infringement of a third party’s intellectual property or other rights.” GOVERN 6.2 requires contingency plans for third-party AI system failures.

This means your AI governance program needs to include a vendor risk component that goes beyond standard security questionnaires. For the AI vendor risk assessment piece, see our post on third-party AI vendor risk assessment.

Post-Deployment Monitoring Setup Deployers don’t just run the model — they’re responsible for establishing the monitoring infrastructure that operators will use. This includes defining what “normal” looks like for model outputs, setting alert thresholds for drift or performance degradation, and building the incident escalation path back to whoever can address a model-level problem.

Regulatory Compliance in the Deployment Context A developer may have built a model that’s technically sound and well-documented. But regulatory compliance in the deployment context — fair lending, ECOA, BSA/AML, privacy — is the deployer’s responsibility. The CFPB has been clear that algorithmic bias in lending is the institution’s problem, not the vendor’s. The deployer can’t transfer compliance responsibility through a contract.

The AI Operator: What You Own in the Daily Run

Operators — sometimes called “operation and monitoring” actors — are responsible for running the AI system after deployment and maintaining continuous oversight. In many organizations, this is the same team as the deployer. In larger institutions with more complex AI environments, operations may sit with a different team (an AI Ops function, model governance team, or line-of-business compliance).

Operator Core Obligations

Continuous Output Monitoring Operators are responsible for “regularly assess[ing] system output and impacts.” For GenAI systems, this means active monitoring of model outputs — not just uptime monitoring. Are outputs staying within acceptable quality bounds? Are edge cases emerging that weren’t caught in pre-deployment testing? Is the model producing harmful, biased, or factually incorrect outputs at rates that exceed tolerance thresholds?

This is harder than it sounds with generative models, because outputs are non-deterministic and high-volume. You need sampling strategies, automated detection for obvious violations, and human review processes for edge cases.

Incident Identification and Escalation Operators are responsible for identifying AI incidents — outputs that cause harm, create regulatory exposure, or indicate model failure — and escalating them appropriately. Under NIST AI 600-1’s incident disclosure requirements, there are specific expectations around what an AI incident is, how it should be documented, and who needs to be informed. The operator is the first line of that response chain.

Human-in-the-Loop Controls For high-stakes decisions — credit underwriting, fraud flagging, customer service escalations — AI 600-1 expects operators to maintain meaningful human oversight. The framework’s MANAGE 2.4 action specifically addresses the need for mechanisms to “supersede, disengage, or deactivate AI systems that demonstrate performance or outcomes inconsistent with intended use.”

In practice, this means operators need defined criteria for when to override model outputs, and those criteria need to be documented and followed consistently. Ad hoc overrides without documentation aren’t governance — they’re liability.

Performance and Drift Tracking Models degrade over time. Training data distributions shift. User behavior changes. Operators are responsible for tracking model performance against established baselines and flagging significant degradation. For financial institutions, this connects to the OCC’s expectations for ongoing model monitoring — the obligation doesn’t end at deployment.

So What? Mapping Your Organization

The practical exercise here is straightforward: for every AI system your organization uses, document which role you’re playing.

Step 1: AI Actor Inventory

For each AI system in use:

  • Are you building/training the model? → Developer
  • Are you taking a vendor’s model and deploying it for a specific use? → Deployer
  • Are you running and monitoring an active deployment? → Operator

Most organizations will be Deployer + Operator for most commercial AI. Some will be Developer + Deployer + Operator for in-house or heavily customized systems.

Step 2: Obligation Gap Assessment

Obligation AreaDeveloperDeployerOperator
Training data governance
Pre-deployment TEVV✓ (for context)
Fit-for-purpose eval
User disclosures✓ (design)✓ (implementation)✓ (ongoing)
Third-party vendor risk
Output monitoring✓ (setup)✓ (ongoing)
Incident escalation✓ (model issues)✓ (context issues)✓ (operational issues)
Human-in-the-loop✓ (design)✓ (implementation)✓ (enforcement)
Model deactivation criteria

Step 3: Governance Accountability Matrix

For each obligation area where you have responsibility, identify: who owns it, what documentation exists, and what monitoring controls are in place. This is what you’re building when you build an AI governance program.

For a complete framework covering all three roles — including model inventory, pre-deployment checklists, and vendor questionnaires — the AI Risk Assessment Template & Guide covers all of this with working templates.

30-Day Implementation Roadmap

  • Week 1: Inventory all AI systems in use; classify each as developer/deployer/operator (most will be deployer+operator)
  • Week 2: For deployer roles — review vendor agreements for limitation disclosure, incident notification, and audit rights
  • Week 3: For operator roles — document monitoring approach, thresholds, and escalation path for each system
  • Week 4: Produce an AI Actor Accountability Matrix and review with legal and compliance leadership

The Exam Reality

Regulators asking about AI governance aren’t looking for philosophy. They want to see that someone in your organization owns the risk at each stage of the AI lifecycle. The developer/deployer/operator framework from NIST AI 600-1 gives you the structure to answer that question clearly.

If you’re a deployer of a commercial LLM, you own the deployment context — the use case appropriateness, the user disclosures, the third-party vendor risk, and the post-deployment monitoring. You cannot point to the vendor and walk away. If you’ve fine-tuned that model on your own data, you’ve crossed into developer territory and you own additional obligations: training data governance, safety testing documentation, and limitation disclosure.

The organizations that will struggle in AI exams are the ones that haven’t done this mapping — that are still treating AI governance as something their vendors handle.


Need a working framework for AI actor accountability, model inventory, and vendor risk? The AI Risk Assessment Template & Guide includes templates for all three roles, pre-deployment checklists, and a third-party AI vendor questionnaire.

FAQ

What is the difference between an AI developer and an AI deployer under NIST AI 600-1?

An AI developer designs, builds, and trains the AI system — responsible for model architecture, training data governance, safety testing, and output labeling. An AI deployer takes that system and deploys it in a specific operational context — responsible for fit-for-purpose evaluation, user disclosures, ongoing monitoring, and regulatory compliance in the deployment environment. A single organization can be both.

What does NIST AI 600-1 require of AI operators?

Operators run the AI system day-to-day after deployment. Their obligations include continuous output monitoring, incident identification, human-in-the-loop oversight for high-stakes decisions, and flagging performance degradation or drift back to the responsible governance team.

Can one organization be both a developer and a deployer?

Yes. NIST AI 600-1 explicitly acknowledges this. A fintech that fine-tunes a foundation model and then deploys it in its own customer-facing product is functioning as both. If the modification is minor — system prompting only — it remains a deployer.

Which NIST AI 600-1 actions apply to deployers specifically?

Key deployer actions include: fitness-for-purpose evaluation in the deployment context, implementing transparency notices for end users, monitoring post-deployment outputs for bias and safety violations, establishing incident escalation procedures, and verifying that developer-supplied risk information is adequate for the intended use case.

How does NIST AI 600-1 address third-party AI supply chains?

GOVERN 6.1 requires policies for AI risks from third-party entities, including IP and data handling risks. GOVERN 6.2 requires contingency plans for third-party AI system failures. Deployers are responsible for due diligence on their developer counterparts, even when using commercial off-the-shelf GenAI APIs.

What happens when a deployer modifies a foundation model?

NIST AI 600-1 draws a line: a deployer becomes a developer if the modification is “transformative and extensive” — such as full fine-tuning on proprietary data that materially changes model behavior. System prompting, RAG, and output filtering alone don’t cross that threshold.

Frequently Asked Questions

What is the difference between an AI developer and an AI deployer under NIST AI 600-1?
An AI developer designs, builds, and trains the AI system—responsible for model architecture, training data governance, safety testing, and output labeling. An AI deployer takes that system and deploys it in a specific operational context—responsible for fit-for-purpose evaluation, user disclosures, ongoing monitoring, and regulatory compliance in the deployment environment. A single organization can be both.
What does NIST AI 600-1 require of AI operators?
Operators (or 'operation and monitoring' actors) run the AI system day-to-day after deployment. Their obligations under AI 600-1 include continuous output monitoring, incident identification, human-in-the-loop oversight for high-stakes decisions, and flagging performance degradation or drift back to the responsible governance team.
Can one organization be both a developer and a deployer?
Yes. NIST AI 600-1 explicitly acknowledges that a single organization may act as both. For example, a fintech that fine-tunes a foundation model and then deploys it in its own customer-facing product is functioning as both a developer (by transforming the model) and a deployer. If the modification is minor—like system prompting only—it remains a deployer.
Which NIST AI 600-1 actions apply to deployers specifically?
Key deployer actions include: ensuring fitness-for-purpose evaluation in the deployment context, implementing transparency notices for end users, monitoring post-deployment outputs for bias and safety violations, establishing incident escalation procedures, and verifying that developer-supplied risk information is adequate for the intended use case.
How does NIST AI 600-1 address third-party AI supply chains?
GOVERN 6.1 requires policies and procedures for AI risks from third-party entities, including IP and data handling risks. GOVERN 6.2 requires contingency plans for failures in third-party AI systems. Deployers are responsible for due diligence on their developer counterparts, even when using commercial off-the-shelf GenAI APIs.
What happens when a deployer modifies a foundation model?
NIST AI 600-1 draws a line: a deployer becomes a developer if the modification is 'transformative and extensive'—such as full fine-tuning on proprietary data that materially changes model behavior. System prompting, RAG, and output filtering alone don't cross that threshold. But if you're doing extensive fine-tuning, you inherit developer-level obligations.
Rebecca Leung

Rebecca Leung

Rebecca Leung has 8+ years of risk and compliance experience across first and second line roles at commercial banks, asset managers, and fintechs. Former management consultant advising financial institutions on risk strategy. Founder of RiskTemplates.

Related Framework

AI Risk Assessment Template & Guide

Comprehensive AI model governance and risk assessment templates for financial services teams.

Immaterial Findings ✉️

Weekly newsletter

Sharp risk & compliance insights practitioners actually read. Enforcement actions, regulatory shifts, and practical frameworks — no fluff, no filler.

Join practitioners from banks, fintechs, and asset managers. Delivered weekly.