DARPA program analysis sites: evidence you can audit
Per-program analytical sites where every quantitative claim is backed by a reproducible query and a confidence level.
Deepfield takes a strategic question, runs it through intake, deep research, entity extraction, knowledge-graph construction, MCTS reasoning, asymmetry and assumption analysis, course-of-action generation, wargame simulation, and output assembly — with every conclusion traceable back to its source.
Most strategic assessment work ends in a slide deck that's stale within a quarter. A program office wants to know whether to fund a bet, how a competitor is likely to react, what the evidence actually says. What they get is a one-shot memo. When the world shifts, they pay again for a fresh one.
We built Deepfield because we wanted to stop selling memos.
Deepfield is a nine-stage assessment pipeline. You feed it a strategic question and it moves through nine modular packages, each with a defined input and output contract:
The whole thing runs end-to-end from a single command. Twenty-thousand-plus lines of production TypeScript. Over two thousand tests.
Pick any recommendation Deepfield produces. You can trace it backward.
The recommendation came from a ranked course of action. That course came from a decision pipeline that scored it across six dimensions and seven risk categories. Those scores came from a reasoning tree whose rollouts used evidence from a knowledge graph. Each node in that graph came from extracted entities with a confidence score. Each entity came from specific source passages the extraction agents evaluated. Each source was retrieved by a research iteration that ran a specific query because a prior iteration flagged a coverage gap.
That chain is carried as W3C PROV-compliant lineage through every stage. It's not a nice-to-have we add when someone asks to see the work. It's how the system is built. Every module returns a Result<T,E>. If something fails, it fails explicitly. We don't let the pipeline fall back to templates or cached defaults to hide a broken call.
The default alternative is a consultancy engagement that delivers a PDF. Deepfield is not a PDF. It's the infrastructure that produces a fresh assessment when your inputs change, runs the same analysis repeatedly without re-paying for the analyst's learning curve, and lets you interrogate any conclusion down to the specific document that supports it.
Deepfield is a platform we embed. A typical engagement runs 12 to 16 weeks. We configure the pipeline for your domain, ingest your sources and credibility priors, tune the reasoning prompts and scoring weights, and run the first real assessment together with your team. At the end you keep the running system, the knowledge graph, the code, and your data.
It's a fit when you have a recurring assessment need, your own sources, and a requirement that every conclusion be traceable back to the evidence that produced it. If your decision is one-shot and never repeats, Deepfield is too much infrastructure. If it's ongoing and the stakes are high, a platform you run yourself costs less over the first year than two consulting engagements and produces work you can actually defend.
We've written a two-page business case for this engagement shape. Executive summary, problem statement, deliverables, risks, success metrics, investment range. Read it in the browser or print it to PDF and forward.
Read the business casePer-program analytical sites where every quantitative claim is backed by a reproducible query and a confidence level.
A knowledge-graph research console that opens Andrew Marshall's Office of Net Assessment tradition to a new generation of strategists.
A deployed teaming platform where an AI orchestrator dispatches specialized agents, but phase gates and a delegation contract keep every real judgment in student hands.
Tell us about the decision you're trying to improve. We'll schedule a briefing with our principals to understand your environment and explore a potential fit.
Schedule a Briefing