Skip to main content

Documentation Index

Fetch the complete documentation index at: https://verdictweight.dev/llms.txt

Use this file to discover all available pages before exploring further.

What the NIST AI RMF is

The NIST AI Risk Management Framework (AI RMF 1.0, published January 2023) is a voluntary U.S. framework structured around four core functions:
FunctionPurpose
GovernCultivate a culture of risk management and accountability.
MapEstablish the context in which AI risks arise.
MeasureAnalyze, assess, and track AI risks.
ManageAllocate risk resources and respond to identified risks.
NIST AI RMF is the de facto baseline for federal AI procurement and a common reference point for U.S. enterprise governance programs. A clean mapping to AI RMF is table stakes for serious deployment in regulated environments.

Coverage summary

VERDICT WEIGHT addresses controls primarily within Measure and Manage, with supporting evidence for Map. Governance requirements (the Govern function) are organizational and largely outside the scope of any technical building block; the framework provides artifacts that support governance reporting but does not constitute governance itself.

Function: Map

The Map function establishes the deployment context, intended use, and known limitations of an AI system. VERDICT WEIGHT contributes to several Map subcategories:
AI RMF subcategoryHow VERDICT WEIGHT supports it
MAP 1.1 – intended purposes and contexts of useThe framework’s threat model (Threat model) explicitly enumerates the deployment contexts the framework is built for and the failure classes it targets.
MAP 2.3 – benefits and limitations of AI system componentKnown limitations is an honest enumeration of what the validation establishes and what it does not.
MAP 3.4 – AI system limitations communicatedThe “what is not claimed” sections throughout the documentation make limitations communicable to non-technical stakeholders.
MAP 5.1 – impacts to individuals, communitiesDocumented per use case in Use cases; operator-supplied for novel deployments.

Function: Measure

The Measure function is where VERDICT WEIGHT does the most work. The framework was built to make measurement of confidence-related risks tractable.
AI RMF subcategoryHow VERDICT WEIGHT supports it
MEASURE 1.1 – appropriate metrics identified and trackedReliability error (REL), AUC, Brier score, ECE — all four headline metrics are documented and reproducible. See Head-to-head comparison.
MEASURE 1.3 – AI system trustworthiness characteristics evaluatedCalibration (Stream 5), uncertainty decomposition (Stream 2), temporal stability (Stream 3), cross-source coherence (Stream 4) each address a distinct trustworthiness dimension.
MEASURE 2.1 – test sets, metrics, details documentedCVE dataset, NVD/KEV methodology, and the ablation table in Ablation studies provide the documentation.
MEASURE 2.4 – deployment performance monitoredThe audit chain (Stream 7) records every scoring event, enabling continuous monitoring of calibration drift, abstention rates, and kill-switch events.
MEASURE 2.6 – AI system safety risks assessedThe completeness proof is a structured safety-risk assessment relative to the documented failure-class taxonomy.
MEASURE 2.7 – AI system security and resilience evaluatedCurveball attack class addresses adversarial robustness; Stream 8 addresses scoring-layer compromise.
MEASURE 2.8 – risks of AI system documentedKnown limitations and the threat-model bounds throughout the documentation.
MEASURE 2.10 – privacy risks evaluatedField-hashing mode in Audit logging addresses sensitive-data handling in the audit record.
MEASURE 3.1 – risk responses identified and documentedThe composition rule routes each risk class to a specific response: graded contribution, abstention, or veto. See Eight-stream composition.
MEASURE 3.2 – risk tracking is robust to changesAudit-chain records preserve historical decisions and their evidence under any subsequent configuration change.
MEASURE 4.2 – AI system performance assessedThe four-metric benchmark suite is reproducible end-to-end. See Head-to-head comparison.

Function: Manage

The Manage function allocates resources and responds to identified risks. VERDICT WEIGHT contributes to the technical controls that enable response.
AI RMF subcategoryHow VERDICT WEIGHT supports it
MANAGE 1.2 – risk responses prioritizedThe framework’s veto-priority composition rule encodes a default risk-response prioritization (hardening signals override core scoring). Operators can override; overrides are recorded in the audit chain.
MANAGE 1.3 – high-priority AI risks responded toThe kill switch (Stream 8) is a binary, deterministic response to scoring-layer compromise — the highest-priority risk class.
MANAGE 2.2 – mechanisms for sustaining the AI system valueThe framework’s reproducibility primitives (canonical evidence records, versioned registry, signed audit chain) sustain value across operator changes.
MANAGE 2.3 – AI system performance monitoredContinuous via the audit chain; periodic re-validation of calibration is a documented operator responsibility.
MANAGE 2.4 – mechanisms for response, recovery, communicationDocumented procedures for kill-switch lower, audit-chain recovery, and incident escalation. See Audit logging.
MANAGE 3.1 – AI risk responses are trackedEvery kill-switch event, every abstention, every abort is recorded with reason in the audit chain.
MANAGE 4.1 – post-deployment monitoring plans implementedThe audit chain is the monitoring substrate. Operators wire dashboards on top of the structured event stream.
MANAGE 4.3 – incidents and errors communicatedAudit IDs and reason strings are designed to be propagated into operator incident-management systems.

Function: Govern

VERDICT WEIGHT does not constitute governance. Governance is organizational. What the framework provides are artifacts that support governance:
AI RMF subcategoryWhat VERDICT WEIGHT provides
GOVERN 1.1 – legal and regulatory requirements understoodMapping documents (this section) translate framework controls into the language of major regimes.
GOVERN 1.4 – risk management processes documentedThe framework’s documentation, source code, audit chain format, and reproducibility pipeline are themselves governance artifacts.
GOVERN 4.1 – organizational policies enforce trustworthy AIConfigurable thresholds and registry-protected configuration enforce policy at the technical layer.
GOVERN 4.2 – AI risks documented and trackedThe audit chain provides the immutable record on which organizational risk tracking is built.
GOVERN 5.2 – risk decisions are documentedLower-kill-switch and threshold-change events are recorded with operator identity and justification.
What the framework does not provide for Govern:
  • Risk management policy (a written statement of what the organization considers acceptable risk).
  • Roles and responsibilities (an organizational chart of who reviews what).
  • Training and awareness (workforce education programs).
  • Stakeholder engagement processes.
These are operator responsibilities. The framework provides substrate; the operator provides organization.

Audit artifacts produced

For each mapped subcategory, VERDICT WEIGHT produces concrete artifacts an auditor can review:
ArtifactWhat it evidences
Hash-chained audit logDecision provenance, integrity, reproducibility.
Per-stream contribution recordsMulti-dimensional risk assessment per decision.
Registry snapshots and hashesConfiguration governance and change tracking.
Calibration map fitting reportsCalibration validity over time.
Kill-switch event logRisk response actions and justifications.
Test suite results (673/673)Engineering-level quality evidence.
Reproducibility pipeline outputsIndependent verification of validation claims.

What the operator still owns

The mappings make this explicit. The operator owns:
  • Articulating organizational risk appetite and translating it into VERDICT WEIGHT thresholds.
  • Establishing review cadences and roles for audit-chain inspection.
  • Integrating audit events with the organization’s incident response.
  • Periodic re-validation of calibration on representative deployment data.
  • Custody of audit-chain signing keys.
  • Storage durability and retention policy for the audit log.
  • Stakeholder communication and training.
  • Legal and regulatory interpretation specific to the deployment.
The framework supports each of these activities; it does not replace them.

Reproducibility

Every claim in this mapping resolves to a specific page in the documentation, a specific module in the source code, or a specific record format in the audit chain. The mapping is intended to be read with the codebase open. Where it is unclear how the framework satisfies a control, that is a documentation defect; please report on GitHub.