Skip to content

NathanMaine/cmmc-compliance-ai-model

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

18 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

CMMC Compliance AI Model

A locally-hosted, fine-tuned language model specialized in CMMC 2.0, NIST 800-171, NIST 800-53, HIPAA, DFARS, and cybersecurity compliance frameworks.

Notice: Research models are provided for evaluation purposes only. The Gemma 4 31B flagship is not publicly available — enterprise deployment runs through the Memoriant Platform. For commercial licensing, pilot programs, or partnership inquiries, contact Nathan Maine or visit memoriant.ai.

Built to answer the question: Can a small team deploy a domain-specific AI compliance advisor that runs entirely on-premises — no cloud, no API fees, no CUI exposure?

Yes. Multiple model sizes from 4B edge models to 31B flagship — laptop to workstation to air-gapped appliance.

Current release: Q2 2026 — Gemma 4 31B flagship (private, enterprise-only, eval loss 0.4517). Two public Nemotron specialized models launching Q2 2026 (lookup specialist + POA&M generator). All Memoriant datasets and benchmarks refreshed quarterly. Published under Memoriant, Inc. organization.


Why This Exists

Organizations pursuing CMMC certification face a knowledge bottleneck. Compliance staff spend hours searching across NIST publications, CMMC assessment guides, and DFARS clauses to answer questions that a well-trained model can handle in seconds.

Commercial LLMs (GPT-4, Claude) are powerful but introduce data residency concerns for organizations handling CUI and ITAR-controlled information. These models run fully local — no data leaves the premises, no internet required, no per-token costs.


⚠️ AI Safety Disclaimer — Always Review Output

AI systems make mistakes. Always review AI-generated output before using it for any purpose.

Any AI system — including those trained on the Memoriant datasets — can produce factually incorrect information, hallucinated citations, outdated guidance, or plausible-sounding fabrications. Never submit AI output directly to DoD assessors, regulatory agencies, or formal compliance documentation without qualified human review.

This is especially critical for CMMC and defense compliance. Wrong answers can cause failed assessments. Failed assessments can cost DoD contracts. C3PAO assessors verify human understanding, not AI output. The DoD holds contractors accountable for their submissions, not the tools they used.

Memoriant's position: AI is a force multiplier for compliance professionals, not a replacement. The human stays accountable. The AI accelerates the work. Every output is a draft for human review.


Model Suite

Version 3.0 — Flagship & Legacy (Q2 2026)

Models trained across multiple base architectures on the v6.0 dataset (18,202 curated examples). Published under Memoriant, Inc. organization.

Flagship (private — enterprise deployment only):

Model Base Parameters Eval Loss GGUF Size Hardware Required
cmmc-expert-gemma4-31b Gemma 4 31B 31B 0.4517 20 GB 48 GB+ VRAM or NVIDIA DGX
cmmc-expert-gemma4-31b-it Gemma 4 31B IT 31B 0.5176 20 GB 48 GB+ VRAM or NVIDIA DGX

The Gemma 4 31B flagship is the best-performing compliance AI model Memoriant has trained. It is not publicly available — the weights remain private as product IP. Enterprise deployment through the Memoriant Platform includes the flagship model, RAG with 30+ authoritative government documents, governed LLM gateway with audit trails, and quarterly regulatory updates.

Contact Memoriant, Inc. for enterprise licensing.

Research Models (earlier training experiments):

Model Base Parameters Eval Loss GGUF Size Hardware Required
cmmc-expert-12b Gemma 3 12B 12B 0.5830 7.9 GB 16 GB VRAM
cmmc-expert-phi4-14b Phi-4 14B 14B 0.6059 ~9 GB 16 GB VRAM
cmmc-expert-olmo2-32b OLMo-2 32B 32B N/A ~20 GB 24 GB+ VRAM
cmmc-expert-llama3.1-8b Llama 3.1 8B 8B 0.8194 5.3 GB 8 GB VRAM

Coming Q2 2026 — Public Nemotron Specialized Models

Two specialized models trained on NVIDIA Nemotron 3 base architectures. Published publicly under memoriant/ on HuggingFace under the NVIDIA Nemotron Open Model License. Each model is highly specialized — excellent at one specific task rather than mediocre at many.

Model Base Specialty Parameters Expected Eval Loss Hardware Required Status
cmmc-expert-nemotron-3-nano-4b NVIDIA Nemotron 3 Nano 4B Lookup Specialist 4B dense TBD 8 GB VRAM (laptop/edge) Training Q2 2026
cmmc-expert-nemotron-3-nano-30b NVIDIA Nemotron 3 Nano 30B-A3B POA&M Generator 30B / 3.5B active (MoE) TBD 24 GB+ VRAM Training Q2 2026

Model 1: Lookup Specialist (Edge Tier — 4B)

Fine-tuned for fast, accurate CMMC control knowledge. Runs on a laptop, phone, or edge device.

What it does well:

  • Look up any CMMC control (AC.L2-3.1.1 through SI.L2-3.14.7) with full detail
  • Identify control families and levels correctly
  • Explain NIST 800-171 controls in plain English
  • Point to the right DFARS clauses for each requirement
  • Refuse to hallucinate (Level 4/5, non-existent controls)

What it does NOT do:

  • Draft SSPs
  • Generate POA&Ms
  • Perform gap analysis
  • Multi-turn compliance reasoning

Tagline: "The CMMC control reference that fits in your pocket. Offline. Free."

Model 2: POA&M Generator (Professional Tier — 30B MoE)

Fine-tuned for remediation planning — POA&M and SSP drafting. 30B total parameters but only 3.5B active per token (Mixture of Experts), so inference is fast despite the model size. 1M token context window.

What it does well:

  • Generate professional POA&M entries for failed controls
  • Draft SSP control descriptions
  • Identify remediation steps and milestones
  • Recommend evidence artifacts
  • Structured output with control IDs, timelines, resources
  • Basic gap analysis for common scenarios

What it does NOT do:

  • Track v6.0 regulatory updates (Phase 2 timeline, DFARS 7019 elimination, HIPAA NPRM)
  • Perform complex multi-framework mapping
  • Provide C3PAO-specific assessment guidance
  • RAG-grounded responses with source citations

Tagline: "Generate POA&M entries and SSP drafts in seconds. Human review required. 100x cheaper than a consultant."

Upgrade Path:

  1. Free — Lookup Specialist (4B) → Learn CMMC basics, reference tool
  2. Free — POA&M Generator (30B) → Draft compliance documentation
  3. Enterprise — Memoriant Platform → Full flagship model + RAG + audit trails + monthly updates

Both Nemotron models will be published under the NVIDIA Nemotron Open Model License (commercial-friendly, no layered licensing).


Training Method: QLoRA fine-tuning (4-bit NF4, rank 64, alpha 128) on NVIDIA B200, H200, and DGX Spark hardware

Runtime: Ollama (OpenAI-compatible API at localhost:11434/v1) or llama.cpp (6.5x faster for serving)


Download

Memoriant Organization (Q2 2026 Release)

All Memoriant assets are now on a quarterly versioning cadence with auto-gated access (login required, auto-approved). Each release is dated and expires at quarter end — new versions incorporate regulatory updates, DFARS amendments, and NIST revisions.

Training Data:

Resource HuggingFace Valid Through
CMMC Training Data memoriant/cmmc-training-data-2026-q2 June 30, 2026

Benchmarks (Tiered — use the right one for your purpose):

Tier Purpose HuggingFace Questions
v1 Preview Methodology sample (NOT for validation) memoriant/cmmc-benchmark-v1-preview-2026-q2 46
v2 Spot Check Triage tool (NOT comprehensive) memoriant/cmmc-benchmark-v2-spotcheck-2026-q2 454
v3 Comprehensive Authoritative evaluation standard memoriant/cmmc-benchmark-v3-comprehensive-2026-q2 1,273

If you are deploying compliance AI in a regulated environment, v3 is the only credible evaluation. v1 and v2 are preliminary tools.

Models:

Resource HuggingFace Type Status
CMMC Expert 12B memoriant/cmmc-expert-12b Model (GGUF) Available (gated)
CMMC Expert Nemotron 3 Nano 4B memoriant/cmmc-expert-nemotron-3-nano-4b Model (GGUF) Coming Q2 2026
CMMC Expert Nemotron 3 Nano 30B memoriant/cmmc-expert-nemotron-3-nano-30b Model (GGUF) Coming Q2 2026

The flagship Gemma 4 31B is not publicly available — enterprise deployment only through the Memoriant Platform. Contact Memoriant, Inc. for licensing.

⚠️ Quarterly Refresh Cycle

All Memoriant datasets and models follow a quarterly release cadence:

  • Q2 2026 (current) — Valid through June 30, 2026
  • Q3 2026 — Releases July 1, 2026
  • Q4 2026 — Releases October 1, 2026

Why quarterly: CMMC regulations, DFARS clauses, and NIST publications update continuously. A frozen benchmark becomes invalid within a quarter. A model trained on stale data produces stale answers. Quarterly refresh ensures evaluation and training stay current with the regulatory landscape.

Using assets after their expiration date produces incomplete evaluations and outdated model behavior.


Quick start with Ollama:

# Download and run the 12B model
ollama run memoriant/cmmc-expert-12b

# Or use the OpenAI-compatible API
curl http://localhost:11434/v1/chat/completions \
  -H "Content-Type: application/json" \
  -d '{
    "model": "memoriant/cmmc-expert-12b",
    "messages": [{"role": "user", "content": "What are the access control requirements for CMMC Level 2?"}]
  }'

Remember: Always review AI output before using it for compliance work. See the AI Safety Disclaimer at the top of this document.


Version 2.0 — Current Release

Released: February 2026

Version 2.0 significantly expands the training corpus with authoritative source material scraped directly from government APIs and official publications. An automated data pipeline handles scraping, conversion, quality filtering, deduplication, and versioning with full reproducibility.

What's New in v2.0

  • 40% more training data — 18,747 total examples (up from 16,906 in v1.0)
  • 6 new authoritative sources — NIST SP 800-53 Rev. 5, NIST CSF 2.0, eCFR regulations, Federal Register, DoD PDFs, NIST SP 800-171 Rev. 3
  • Expanded LoRA coverage — All 7 transformer modules targeted across all model sizes
  • Best eval loss: 1.048 — 72B flagship model achieves the lowest loss in the suite
  • Improved across the board — 7B eval loss improved from 1.241 (v1.0) to 1.142 (v2.0)

The regulatory landscape changed substantially since the v1.0 training data was assembled. Version 2.0 addresses these gaps:

Update Date Significance
NIST SP 800-171 Rev. 3 May 2024 Replaces Rev. 2. Consolidated from 110 to 97 controls. Adds 88 organization-defined parameters (ODPs) and 509 assessment objectives. This is the new CMMC Level 2 foundation
NIST CSF 2.0 Feb 2024 Major revision adding the GOVERN function (6 functions total). 34 categories, 174 subcategories with implementation examples
CMMC Final Rule (32 CFR 170) Dec 2024 The actual regulation establishing the CMMC program. Effective rule text, not just guidance
DFARS 252.204-7021 Nov 2025 Acquisition rule requiring CMMC certification in DoD contracts. Phase-in timeline and applicability
NIST SP 800-172 Rev. 3 2025 (FPD) Enhanced CUI requirements, final public draft. CMMC Level 3 delta
DoD Assessment Guides 2025 Official L2/L3 assessment procedures, scoping guides, ODP values document

Training Data

14,906 training + 3,841 validation examples (~4.5M tokens) assembled from 11 sources:

v1.0 Legacy Sources (13,434 examples)

Source Records % Coverage
NIST Cybersecurity Publications 6,372 33.9% SP 800-171, 800-172, 800-53, 800-37, CSF, and related guidance
CMMC Primary 4,787 25.5% CMMC 2.0 requirements, controls, implementation guidance
CMMC Balanced 994 5.3% Proportional coverage across all CMMC domains
HIPAA Compliance 961 5.1% Security Rule requirements and technical safeguards
CMMC Core 320 1.7% High-priority practices and assessment-critical requirements

v2.0 New Sources (1,841 examples via automated pipeline)

Source Scraper Records % Method
NIST SP 800-53 Rev. 5 nist_csrc 773 4.1% OSCAL JSON catalog. Full control catalog including enhancements
DoD Documents dod_documents 519 2.8% PDF extraction — Assessment Guide L2, Scoping Guides, ODP Values, SP 800-172 R3
Federal Register federal_register 350 1.9% Federal Register API — CMMC rulemakings, DFARS notices
eCFR Regulations ecfr 75 0.4% eCFR API — 32 CFR 170 (CMMC), DFARS cyber clauses, 45 CFR 164 (HIPAA)
NIST SP 800-171 Rev. 3 nist_sp800_171 63 0.3% OSCAL JSON — 97 controls with assessment objectives and ODPs
NIST CSF 2.0 nist_csf 61 0.3% OSCAL JSON — 34 categories + subcategories with implementation examples

Automated Pipeline

The v2.0 data pipeline (cmmc-data-pipeline) is fully automated and reproducible:

                Authoritative Sources
    ┌──────────────────────────────────────────┐
    │  NIST OSCAL (GitHub)    eCFR API         │
    │  DoD PDFs               Federal Register │
    └─────────────────┬────────────────────────┘
                      │
                      ▼
    ┌──────────────────────────────────────────┐
    │  Step 1: Scrape                          │
    │  Rate-limited, retry-enabled scrapers    │
    │  Raw JSON saved to data/raw/{source}/    │
    ├──────────────────────────────────────────┤
    │  Step 1b: Relevance Filter               │
    │  eCFR filtered to CMMC-relevant DFARS    │
    │  clauses only (252.204-70xx, 252.239)    │
    ├──────────────────────────────────────────┤
    │  Step 2: Convert                         │
    │  Source-specific templates generate       │
    │  chat-format instruction/response pairs  │
    ├──────────────────────────────────────────┤
    │  Step 3: Quality Filter                  │
    │  Min length, max length (8K), alpha ratio│
    ├──────────────────────────────────────────┤
    │  Step 4: Deduplicate                     │
    │  xxhash exact + MinHash LSH near-dedup   │
    │  (128 perms, Jaccard 0.8, 5-gram)        │
    ├──────────────────────────────────────────┤
    │  Step 5: Validate                        │
    │  Format checks, quality scoring, stats   │
    ├──────────────────────────────────────────┤
    │  Step 6: Version                         │
    │  Immutable snapshots with rollback       │
    ├──────────────────────────────────────────┤
    │  Step 7: Merge                           │
    │  Cross-version dedup against v1.0 data   │
    └──────────────────────────────────────────┘

The pipeline supports both full scrapes and incremental updates. Each run creates a versioned snapshot that can be inspected, diffed, or rolled back before merging into the training dataset.

Pipeline Results (Final — v004)

Step Result
Scraped 4,438 raw records from 6 new sources
Relevance filtered 4,100 kept (338 irrelevant DFARS clauses removed)
Converted 2,193 chat-format training records
Quality filtered 1,888 passed (305 rejected — too short or too long)
Deduplicated 1,841 unique (47 near-duplicates removed)
Validated PASSED — 0 format errors, avg answer length 1,322 chars
Merged with v1.0 18,747 total (14,906 train + 3,841 validation)

Training Configuration

All models trained using QLoRA (Quantized Low-Rank Adaptation) — base weights frozen in 4-bit NF4, trainable adapter layers injected into all 7 transformer projection modules. Trained on NVIDIA A100-SXM4-80GB via RunPod.

Parameter 7B 14B 32B 72B
GPU A100 80GB SXM A100 80GB SXM A100 80GB SXM A100 80GB SXM
LoRA rank 64 16 32 16
LoRA alpha 128 32 64 32
LoRA dropout 0.05 0.05 0.05 0.05
Target modules All 7 All 7 All 7 All 7
Effective batch size 32 16 16 16
Learning rate 2e-4 1e-4 1e-4 5e-5
Epochs 3 3 3 3
Max sequence length 2048 2048 2048 2048
Precision bf16 bf16 bf16 bf16
Optimizer AdamW 8-bit AdamW 8-bit AdamW 8-bit AdamW 8-bit
Packing Enabled Enabled Enabled Enabled
Training time ~3.1 hours ~6.5 hours ~9.6 hours ~13.0 hours
Final eval loss 1.142 1.144 1.073 1.048

The 72B model used Unsloth for memory-efficient 4-bit loading, enabling QLoRA fine-tuning on a single A100-80GB without multi-GPU setups.

Evaluation Results

All models showed continuous improvement across training with no overfitting observed.

Per-Model Training Metrics:

Metric 7B 14B 32B 72B
Final Eval Loss 1.142 1.144 1.073 1.048
Final Train Loss 1.030 1.009 1.005 0.966
Token Accuracy 76.5% 77.7% 77.9%
Total Steps 282 561 561 564
Training Time 3.1h 6.5h 9.6h 13.0h
GGUF Size 5.1 GB 9.8 GB 18.9 GB 45 GB

The 72B model achieves the lowest eval loss in the suite (1.048), demonstrating that model scale continues to improve compliance reasoning quality even with the same training data.

v1.0 vs v2.0 Comparison (7B):

Metric v1.0 v2.0 Change
Training Examples 13,434 14,906 +11%
Eval Loss 1.241 1.142 -8% (improved)
LoRA Target Modules 4 7 +75% coverage
Data Sources 5 11 +6 new

Compliance Framework Coverage

Full Coverage Map (v1.0 + v2.0 Combined)

Framework v1.0 v2.0 Adds
CMMC 2.0 (32 CFR Part 170) Level 1–3 practices, assessment methodology Actual regulatory text from eCFR, DoD assessment guide procedures, scoping guidance
NIST SP 800-171 Rev. 2 (110 requirements) Rev. 3 (97 controls with assessment objectives, methods, and 88 ODPs)
NIST SP 800-172 Original Rev. 3 Final Public Draft (enhanced CUI requirements)
NIST SP 800-53 Rev. 5 (from NIST pubs dataset) Rev. 5 full OSCAL catalog (1,016 controls + enhancements with structured statements and guidance)
NIST CSF 1.1 (5 functions) 2.0 (6 functions — adds GOVERN. 34 categories, 174 subcategories)
HIPAA Security Rule Q&A training pairs Full regulatory text from 45 CFR Part 164 (41 sections)
DFARS Clauses 7012, 7019, 7020, 7021 guidance Full regulatory text from 48 CFR Part 252 (348 sections)
Federal Register CMMC rulemakings, DFARS proposed/final rules, CUI policy notices
DoD Assessment Guides L2 Assessment Guide (332 chunks), L2/L3 Scoping Guides, ODP Values

DFARS and NIST Coverage Detail

The models understand the full chain from contract clause to technical implementation:

DFARS 252.204-7012 — Safeguarding Covered Defense Information

  • Adequate security requirements for CUI on contractor systems
  • Cyber incident reporting obligations (72-hour timeline)
  • Flow-down requirements to subcontractors
  • Relationship to NIST SP 800-171 compliance

DFARS 252.204-7019 — Notice of NIST SP 800-171 Assessment

  • SPRS (Supplier Performance Risk System) scoring methodology
  • Self-assessment requirements and documentation
  • DoD Assessment Methodology (Basic, Medium, High)

DFARS 252.204-7020 — NIST SP 800-171 DoD Assessment Requirements

  • Government assessment access and cooperation requirements
  • Relationship between SPRS scores and assessment levels

DFARS 252.204-7021 — Cybersecurity Maturity Model Certification

  • CMMC level requirements by contract type
  • Phase-in timeline and applicability
  • Relationship between CMMC levels and NIST SP 800-171/172

Use Cases

Application How It Works
SSP Generation Draft System Security Plan control descriptions with proper NIST/CMMC citations
Gap Analysis Identify which controls are required for specific CMMC levels and contract requirements
Assessment Prep Generate evidence checklists and assessment objective narratives
Cross-Framework Mapping Map controls between CMMC, NIST 800-53, HIPAA, and other frameworks
Policy Drafting Create initial policies and procedures aligned to specific CMMC practices
DFARS Clause Analysis Identify applicable requirements from contract language (7012, 7019, 7020, 7021)
Training & Education Always-available compliance reference — no waiting for SMEs

Architecture

┌──────────────────────────────────────────────────────────┐
│                     User Query                            │
│   "What access control requirements apply to              │
│    CMMC Level 2 for CUI handling?"                        │
└──────────────────────┬────────────────────────────────────┘
                       │
                       ▼
┌──────────────────────────────────────────────────────────┐
│              Ollama Runtime (Local)                        │
│  ┌────────────────────────────────────────────────────┐   │
│  │  Foundation Model (frozen, 4-bit quantized)         │   │
│  │  + QLoRA Adapters (compliance-tuned)                │   │
│  │                                                     │   │
│  │  7B  — quick lookups, day-to-day queries            │   │
│  │  14B — detailed analysis, multi-control reasoning   │   │
│  │  32B — deep gap assessments, SSP drafting           │   │
│  │  72B — complex multi-framework analysis             │   │
│  └────────────────────────────────────────────────────┘   │
│  System prompt: Compliance expert across CMMC/NIST/HIPAA  │
└──────────────────────┬────────────────────────────────────┘
                       │
                       ▼
┌──────────────────────────────────────────────────────────┐
│              Structured Response                          │
│   - Framework-specific references (SP 800-171 §3.1)      │
│   - Implementation guidance                               │
│   - Assessment evidence requirements                      │
│   - Cross-framework mappings (CMMC ↔ 800-53 ↔ HIPAA)    │
│   - DFARS clause applicability                            │
└──────────────────────────────────────────────────────────┘

Key Design Decisions

Decision Rationale
4 model sizes (7B–72B) Tiered deployment — laptop for quick lookups, workstation for deep analysis. Match the model to the hardware
QLoRA (not full fine-tune) Trainable params are <0.1% of total. 50x less compute, comparable domain accuracy
Security domain completeness Models are fine-tuned for complete security domain coverage, including vulnerability analysis, incident response scenarios, and access control failure modes required for professional SSP and POA&M generation. Behavioral guardrails and policy enforcement are handled at the governed-llm-gateway layer
q5_k_m / q4_k_m quantization 5-bit for smaller models (accuracy-sensitive), 4-bit for larger (size-constrained). Compliance is fact-heavy — extra bit preserves control IDs
Local-only deployment CUI/ITAR data cannot leave premises. Zero cloud dependency by design. Fully air-gappable
Multi-framework training Organizations rarely have single-framework obligations. Cross-mapping across CMMC, NIST, HIPAA, and DFARS is the real value
OSCAL-native scraping NIST's machine-readable OSCAL JSON provides structured control data far richer than PDF extraction
Automated pipeline with versioning Reproducible data updates with immutable snapshots, rollback, and dedup against existing training data

Security and Privacy

This model suite was designed for environments where data sovereignty matters:

  • Fully air-gappable — Runs entirely on local hardware after initial download. No internet required for inference
  • No telemetry — No data is transmitted to any external service
  • No third-party API dependency — No per-query costs, no rate limits, no data exposure to cloud providers
  • CUI-safe deployment — Suitable for use in environments processing Controlled Unclassified Information, as no data leaves the local system boundary
  • Customizable — Can be further fine-tuned with organization-specific policies, SSPs, and internal security documentation

Hardware Requirements

Inference (Running a Model)

Model GPU (VRAM) CPU-Only (System RAM) Storage
7B 8 GB 16 GB 10 GB
14B 12 GB 24 GB 15 GB
32B 24 GB 32 GB 25 GB
72B 48 GB 64 GB 50 GB

Training (Reproducing from Scratch)

Model GPU Required Approx. Time
7B 16 GB VRAM (e.g., RTX 5000 Ada) ~3.1 hours
14B 40+ GB VRAM (e.g., A100 40GB) ~6.5 hours
32B 80 GB VRAM (e.g., A100 80GB) ~9.6 hours
72B 80 GB VRAM (e.g., A100 80GB) ~13.0 hours

OS: Linux, macOS, Windows (WSL2)


Repository Structure

cmmc-compliance-ai-model/
├── README.md                    # This file
├── docs/
│   ├── training-methodology.md  # Detailed QLoRA configuration and rationale
│   ├── data-pipeline.md         # Full pipeline documentation with filtering logic
│   └── evaluation-results.md    # Eval metrics, example outputs, failure modes
├── pipeline/
│   ├── 01_format_converter.py   # Raw → chat-style instruction pairs
│   ├── 02_quality_filter.py     # Length, artifact, and fragment removal
│   ├── 03_relevance_filter.py   # NIST relevance scoring and sampling
│   ├── 04_deduplication.py      # xxhash exact + MinHash LSH near-dedup
│   └── 05_train_val_split.py    # Stratified split with source balancing
├── training/
│   ├── train_qlora.py           # QLoRA training script
│   ├── config.yaml              # Hyperparameters and training config
│   └── merge_and_quantize.py    # Adapter merge + GGUF quantization
├── evaluation/
│   ├── eval_compliance.py       # Framework-specific accuracy testing
│   └── eval_cross_mapping.py    # Cross-framework mapping validation
├── deployment/
│   ├── Modelfile                # Ollama model configuration
│   └── setup_ollama.sh          # Local deployment script
└── publishing/
    ├── huggingface/             # Model cards + upload scripts
    ├── ollama/                  # Ollama Library submission guide
    └── github-releases/         # GitHub Release creation script

Note: This repo contains the pipeline code, training configuration, and documentation. Pre-trained model weights (GGUF) are available on Hugging Face. Training data and checkpoints are excluded from the repository.


Roadmap

  • v1.0 - 4 models (7B-72B) trained on Qwen2.5, published on HuggingFace
  • v2.0 - Automated scraping pipeline, expanded to 18,747 examples from 11 sources
  • v3.0 - Base model migration to US-origin architectures (Gemma, Llama, Phi, OLMo, Granite)
  • v3.0 - 13 models trained across 8 architectures on NVIDIA B200 and DGX Spark
  • v3.0 - Flagship: Gemma 4 31B (eval loss 0.4517, day-zero fine-tuning)
  • v3.0 - Published under Memoriant, Inc. HuggingFace organization
  • v3.0 - Standardized compliance benchmark (46 questions, 9 tiers) published
  • v3.0 - CMMC Expert Platform v08 (orchestrator + gateway + RAG + PII + audit trails)
  • v3.0 - Open source contributions: huggingface/peft#3129, huggingface/transformers#45200
  • v3.1 - Platform v09 with Qdrant vector search (replacing ChromaDB)
  • v3.1 - Gemma 4 26B-A4B MoE fine-tuning (fast inference variant)
  • v3.2 - 500+ question production benchmark (full NIST 800-171 coverage)
  • v4.0 - Agent integration with CrewAI for multi-agent compliance workflows
  • v4.0 - FedRAMP baselines, CIS Controls, and ITAR coverage
  • v4.0 - Custom fine-tuning pipeline for organization-specific compliance data

Built With


Sources and References

All training data is derived from publicly available, authoritative government sources:

CMMC Program

Source URL
CMMC Final Rule (32 CFR Part 170) https://www.ecfr.gov/current/title-32/subtitle-A/chapter-I/subchapter-D/part-170
DoD CIO CMMC Portal https://dodcio.defense.gov/cmmc/
CMMC Level 2 Assessment Guide https://dodcio.defense.gov/Portals/0/Documents/CMMC/AGLevel2.pdf
CMMC Level 2 Scoping Guide https://dodcio.defense.gov/Portals/0/Documents/CMMC/ScopingGuideLevel2.pdf
CMMC Level 3 Scoping Guide https://dodcio.defense.gov/Portals/0/Documents/CMMC/ScopingGuideLevel3.pdf
CMMC ODP Values Document https://dodcio.defense.gov/Portals/0/Documents/CMMC/ODPValues.pdf

NIST Publications

Source URL
NIST SP 800-171 Rev. 3 https://csrc.nist.gov/pubs/sp/800/171/r3/final
NIST SP 800-171 Rev. 3 OSCAL Catalog https://github.com/usnistgov/oscal-content
NIST SP 800-172 Rev. 3 (Final Public Draft) https://csrc.nist.gov/pubs/sp/800/172/r3/fpd
NIST SP 800-53 Rev. 5 https://csrc.nist.gov/pubs/sp/800/53/r5/upd1/final
NIST SP 800-53 Rev. 5 OSCAL Catalog https://github.com/usnistgov/oscal-content
NIST SP 800-37 Rev. 2 (RMF) https://csrc.nist.gov/pubs/sp/800/37/r2/final
NIST Cybersecurity Framework 2.0 https://www.nist.gov/cyberframework
NIST CSF 2.0 OSCAL Catalog https://github.com/usnistgov/oscal-content

Federal Regulations

Source URL
eCFR — 32 CFR Part 170 (CMMC) https://www.ecfr.gov/current/title-32/subtitle-A/chapter-I/subchapter-D/part-170
eCFR — 48 CFR Part 252 (DFARS) https://www.ecfr.gov/current/title-48/chapter-2/subchapter-H/part-252
eCFR — 45 CFR Part 164 (HIPAA Security Rule) https://www.ecfr.gov/current/title-45/subtitle-A/subchapter-C/part-164
Federal Register (CMMC Rulemakings) https://www.federalregister.gov/

DFARS Clauses

Clause Title Reference
DFARS 252.204-7012 Safeguarding Covered Defense Information https://www.ecfr.gov/current/title-48/section-252.204-7012
DFARS 252.204-7019 Notice of NIST SP 800-171 Assessment https://www.ecfr.gov/current/title-48/section-252.204-7019
DFARS 252.204-7020 NIST SP 800-171 DoD Assessment Requirements https://www.ecfr.gov/current/title-48/section-252.204-7020
DFARS 252.204-7021 Cybersecurity Maturity Model Certification https://www.ecfr.gov/current/title-48/section-252.204-7021

Data Formats

Format Description Source
OSCAL JSON Machine-readable security control catalogs (NIST's Open Security Controls Assessment Language) https://pages.nist.gov/OSCAL/
eCFR API Electronic Code of Federal Regulations structured text https://www.ecfr.gov/developers/documentation/api/v1
Federal Register API Government policy documents and rulemakings https://www.federalregister.gov/developers/documentation/api/v1

Limitations

  • The model is trained on publicly available compliance framework documentation. It does not contain classified or controlled information.
  • Responses should be treated as expert-informed guidance, not as legal or regulatory determinations. All compliance decisions should be validated by qualified assessors.
  • The model's knowledge reflects the training data at time of creation. Regulatory updates published after the training cutoff should be incorporated through the automated pipeline or RAG.
  • Performance on highly specialized or edge-case compliance scenarios may vary. The model performs best on well-documented framework requirements.

Built by Nathan Maine — Solving compliance bottlenecks with purpose-built AI.

About

Suite of 4 fine-tuned LLMs (7B/14B/32B/72B) for CMMC 2.0, NIST 800-171, NIST 800-53, HIPAA, and DFARS compliance. Air-gappable, runs on Ollama with zero cloud dependency.

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages