The Artifact Set Behind Every Engagement

What you actually get.

Every CPP engagement ships a complete artifact set — interactive HTML dashboards your team can drill into, DOCX cover letters built for owner submission, XLSX workbooks that survive expert review, and PDF exhibits ready to drop into an arbitration binder. Browse the full set below.

The engine produces the files. The analyst produces the conclusion. Both ship.

Note: the cards below describe each deliverable's structure, sections, and methodology citations — not screenshots from real client engagements. To see one rendered live on your own data, drop your XER into the live demo.
Eight Deliverable Types

Every artifact a forensic engagement produces.

A schedule-health check ships one. An owner-submission EOT package ships six or seven, indexed and cross-referenced. Each one is keyed to a published recommended practice — DCMA 14-Point, AACE 29R-03, AACE 52R-06, AACE 122R-22, SCL Delay & Disruption Protocol 2nd Ed.

Schedule Health Dashboard

HTML · ~1.3 MB

Self-contained DCMA 14-Point assessment of a baseline-vs-current XER pair. One file, no dependencies, opens in any browser — drop it on a network share and your PMO sees the same dashboard you do.

What's inside
  • Executive summary with overall grade and the three to five findings most likely to surface in cross-examination
  • 14-criterion drill-downs — logic, lags, hard constraints, BEI, missed tasks, float distribution, critical-path reasonableness, each with pass/fail rationale
  • Baseline vs current variance — activity-level slip, added/removed activities, retroactive baseline edits
  • Slip register ranked by impact to project finish
  • BEI / CPLI / SPI metrics with industry benchmarks and contextual reading
  • GAO-16-89G + CPP Quality Overlay grade with embedded reproducibility manifest
Methodology: DCMA 14-Point Assessment · AACE RP 29R-03 · GAO-16-89G Schedule Assessment Guide

Forensic Delay Analysis Report

HTML DOCX CSV

Windows analysis (AACE MIP 3.3 — Observational / Dynamic / Contemporaneous As-Is) across the full update chain. The HTML dashboard is the analyst's working document; the DOCX narrative is the owner-deliverable executive read.

What's inside
  • Per-window cause attribution with critical-path-impact gate and per-party day buckets that sum to project drift
  • Concurrent delay matrix — per window × per party — with WOET 4-state day classifier
  • Notice-gate review per SCL Protocol §3 (paired with evidence ledger when claim-workbench is run alongside)
  • Baseline stability check flagging retroactive edits, added/removed activities, completion reversals
  • Dual-method validation against collapsed-as-built result with documented gap reconciliation
  • Per-window CSVs for downstream pivot work, expert workpapers, and exhibit preparation
Methodology: AACE RP 29R-03 §3.3 Observational/Dynamic/Contemporaneous As-Is · SCL Delay & Disruption Protocol 2nd Ed §11.5

Collapsed As-Built Validation

HTML

But-for analysis (AACE MIP 3.8 — Modeled / Subtractive / Single Simulation). Independent recomputation of project drift via subtractive removal of delay activities from the as-built schedule. The validation step opposing counsel demands.

What's inside
  • As-built finish vs but-for finish per delay event, with cumulative roll-up across the event population
  • Per-event float erosion showing which removals matter to project finish and which are absorbed
  • Subtractive simulation log recording every activity edit so the calculation is reproducible
  • Dual-method gap report against the windows-analysis result — flags any divergence that needs analyst reconciliation before submission
  • Methodology disclosure identifying what the technique can and cannot conclude
Methodology: AACE RP 29R-03 §3.8 Modeled/Subtractive/Single Simulation · SCL Delay & Disruption Protocol 2nd Ed §11.5

Time Impact Analysis Package

HTML DOCX

Prospective TIA with fragnet insertion against a contemporaneous baseline. The forward-looking method when a delay event has occurred but its impact on remaining work is the question.

What's inside
  • Fragnet insertion for each delay event — added activities, logic ties, calendar assignments, constraint treatment
  • Modeled / additive / multiple-base CPM forward-pass with pre- and post-impact finish dates
  • Per-event impact quantification in working days, with float-erosion accounting
  • Cumulative impact tracking when multiple events are modeled in sequence
  • Owner-vs-contractor delay attribution with concurrent-delay handling per SCL Protocol
  • DOCX narrative formatted for inclusion as a TIA exhibit in an EOT submission
Methodology: AACE RP 52R-06 Time Impact Analysis · AACE RP 29R-03 §3.7 Modeled/Additive/Multiple Base

Monte Carlo Schedule Risk Report

HTML DOCX

Probabilistic finish-date forecast with sensitivity analysis. Suitable for executive briefings, owner risk reviews, or as the quantitative spine of a contractor's schedule-acceleration case.

What's inside
  • P10 / P50 / P80 / P90 completion-date forecasts using linear-interpolation percentiles per AACE convention
  • Sensitivity tornado identifying the activities whose duration uncertainty drives finish-date variance
  • Scenario comparison — baseline risk profile vs mitigated profile, with side-by-side finish-date distributions
  • Convergence diagnostics confirming iteration count is sufficient for the reported percentile precision
  • QRAMM maturity badge per AACE 122R-22, naming the maturity tier the inputs support
  • DOCX narrative with input assumptions, methodology disclosure, and what the simulation does not establish
Methodology: AACE RP 122R-22 QRAMM-aligned · linear-interpolation percentiles

EOT Claim Submission Package

DOCX XLSX PDF MD

The full owner-submission-ready Extension of Time package. Cover letter, per-event exhibits, forensic workbook, mitigation report, supporting-documents index, and an internal QC pass — assembled as one indexed bundle.

What's inside
  • DOCX cover letter framed for owner submission, naming the contract clauses and notice provisions invoked
  • Per-event exhibits — one DOCX per delay event with cause, attribution, contemporaneous notice, impact quantification
  • Forensic workbook (XLSX) — per-window math, per-event roll-up, attribution buckets, formulas left visible for owner-side review
  • Mitigation report documenting steps taken to reduce or absorb impact, per SCL Protocol mitigation duty
  • Supporting documents index cross-referencing every cited RFI, change order, daily report, and correspondence item
  • Owner-submission QC markdown — internal pre-flight check for completeness, citation accuracy, exhibit cross-references
Methodology: SCL Delay & Disruption Protocol 2nd Ed §3 notice-gate · AACE RP 67R-11 entitlement framework

Counter-Claim / Defense Package

DOCX XLSX HTML

Defensive equivalent of the EOT package. Tests the opposing party's claim against ten defense theories, then assembles the rebuttal letter, defense workbook, and dashboard for response.

What's inside
  • DOCX rebuttal letter structured to match the opposing party's submission, theory by theory
  • XLSX defense workbook with per-event findings and the math behind each defense theory
  • HTML rebuttal dashboard for internal counsel review and deposition prep workpapers
  • Ten defense theories tested: no critical-path impact, concurrent delay, pacing, no notice, failure to mitigate, causation gap, float ownership, scope-change disguise, self-caused acceleration, LD cap
  • Per-theory evidence pull citing the contemporaneous record items each theory rests on
Methodology: SCL Delay & Disruption Protocol 2nd Ed · Wickwire and Pickavance treatise framing

Claim Workbench Bundle

HTML CSV JSON

Evidence-triage layer underneath the forensic report. Takes one folder of mixed correspondence (MSG / PDF / DOCX / XLSX) plus the full XER update chain and produces a unified analysis the analyst works from.

What's inside
  • Evidence ledger CSV — every correspondence item parsed, dated, sender-classified by analyst-supplied domain mapping
  • Schedule chain-diff CSV across the full update chain — TASKPRED add/remove, constraint flips, retroactive baseline edits, completion reversals
  • Rolling baseline — every activity carries the baseline date it had when it first appeared in the chain
  • Trust score JSON flagging statistical patterns (zero-duration-variance schedules, no-new-activities, every-activity-hits-baseline) — correlation, not causation
  • Slip-to-evidence cross-reference with NER-extracted recurring actors flagged for analyst follow-up review
  • HTML dashboard stitching all of the above into one navigable working surface
Methodology: Forensic-correctness audit framework · disclosure boilerplate inline · CORRELATION not CAUSATION
How an Engagement Maps to Deliverables

Four typical engagements. Different artifact sets.

Not every project needs the full bundle. A schedule-health check is one HTML file. An owner defense is the workbench plus the counter-claim package. The right deliverable mix is the first conversation in scoping.

Engagement Type 01

Schedule Health Check

Ships →

Schedule Health Dashboard — one self-contained HTML file with DCMA 14-Point findings, baseline-vs-current variance, and the BEI/CPLI/SPI snapshot. Used as a pre-bid check, monthly health pass, or pre-claim diagnostic.

Engagement Type 02

Forensic Delay Analysis

Ships →

Forensic Delay Analysis Report + Collapsed As-Built Validation + forensic workbook. Two methods (MIP 3.3 windows and MIP 3.8 collapsed as-built) computed independently, then reconciled. The dual-method standard opposing counsel cannot dismiss as single-method bias.

Engagement Type 03

EOT Claim Submission

Ships →

Full claim package — cover letter + per-event exhibits + forensic workbook + mitigation report + supporting-documents index + owner-submission QC markdown. Indexed as one bundle, ready to submit on the contract notice clause's required medium.

Engagement Type 04

Owner Defense (Counter-Claim)

Ships →

Counter-Claim Package + Claim Workbench Bundle for evidence triage. The workbench builds the contemporaneous record cross-reference; the counter-claim package tests the opposing claim against ten defense theories, theory by theory.

The deliverables are concrete. The analysis behind them is the work.

Every artifact above is reproducible from a CPP engagement — the engine produces the files, the analyst produces the conclusion. Try the engine on your own data, or scope an engagement directly.