Most founders test demand. Far fewer pressure-test the math that says demand will pay the bills. The order of operations matters: a signed letter of intent does not save you from a category whose comp set churns at 50% in six months, and a viral waitlist does not rescue an AOV that cannot cover fulfillment.
Validation unit economics is the discipline of doing that pressure-test up front — before you write a line of code or raise a dollar — using public data on comparable businesses to figure out whether the contribution margin you are assuming is actually achievable in your category.
This is not "how to build a financial model in Excel." Excel models are for forecasting once you already believe the math works. Validation unit economics is the step before that, and most founders skip it.
What we mean by it
A validation unit-economics read asks one question: if everything in the operating plan goes roughly as well as it has gone for the most successful comparable companies in this category, does the per-unit math work?
That framing matters. The point is not to pencil in your best-case numbers and see if they look pretty. The point is to pencil in the best numbers any comparable business has ever publicly reported and ask whether even those make the unit profitable. If the answer is no, you have a category problem, not an execution problem — and no amount of grit will fix it.
Validation unit economics is a category test, not a plan test. It asks whether the math can work, not whether your specific plan will.
That distinction is what separates a research-backed validation report from a financial model. The model trusts your inputs. The validation report pressure-tests them.
Three signals visible from public data
The first signal is comp-set repeat-rate decay. Almost every consumer category has a public floor on customer frequency and a public ceiling on retention, set by the largest companies in the space and reported in their S-1s, investor decks, or earnings calls. If your plan assumes 2x-per-week purchase frequency in a category where the public leaders top out at 1x, that gap is pre-build-flaggable from public data. You don't need to launch to know the assumption is off; you need to read.
The second signal is density math against zip-code reality. Many delivery, last-mile, and physical-network businesses assume an order density that requires a population concentration their actual launch zips do not have. Census data, Google Maps drive-time isochrones, and the public route economics of incumbents (DoorDash, Instacart, Domino's franchise disclosures) tell you what density is achievable. If your contribution margin only works at densities the category has never sustained outside of three Manhattan zip codes, that is a structural finding, not a marketing problem.
The third signal is marginal cost per geographic unit. Most physical-footprint businesses underestimate the capex required per new city, and they underestimate the months to break-even utilization in that city. Public S-1s from Sweetgreen, Cava, Blue Apron, and the post-mortems on shut-down players all give you working ranges. If the plan assumes a city pays back in 9 months and the comp set says 18-24, the deck math collapses on contact with reality — and again, this is visible before you sign a single commercial lease.
None of these signals require proprietary data. They require an analyst willing to look at the comp set honestly.
The Munchery case
Munchery is the cleanest example. The prepared-meal-delivery startup raised roughly $125M and shut down in 2019. The autopsy is unambiguous: the unit economics never penciled, and the data to know that was public before the Series B closed.
Munchery's model required an AOV around $22 or higher to cover roughly $8-9 of per-order fulfillment cost (driver, packaging, last-mile). The achievable AOV in the category — what real customers actually put in real carts — was $14-18. That is a category ceiling, not a marketing failure.
Munchery's plan also assumed 1.5-2x weekly customer frequency. The comp set — Blue Apron and HelloFresh, both publicly reporting — showed 0.8-1.2x weekly frequency with 50%+ six-month churn. The frequency assumption was off by roughly 2x against the best comparables in the category.
Then there was the geography math. Each city required $1.5-2M in commissary kitchen and buildout capex, plus 6-9 months to reach break-even utilization. The deck math expanded faster than the unit math could compound.
And the kicker: Sprig had already shut down for the same reason in 2017. The post-mortems were public. A research-backed validation report on Munchery's premise in late 2017 would have flagged every one of these signals. Not as opinions — as numbers visible in S-1s, in shut-down post-mortems, and in census-grade density data.
This is what we mean by pre-build-flaggable from public data. Munchery did not need to lose $125M to discover its unit economics were broken. The discovery was sitting in the comp set the whole time.
For the longer worked example, see the Munchery autopsy, and for a related case where the unit math was equally visible up front, see Juicero.
How DimeADozen handles this
DimeADozen.AI's Financial Model section is built around exactly this kind of pressure test. We don't build your spreadsheet for you. We do something more useful: we tell you what the comp-set numbers actually are, where your stated assumptions sit relative to those numbers, and which assumptions would have to be true at the same time for the model to work. If three of them have to all hit category-record values simultaneously, that is the finding.
The Risk Analysis section then lifts the structural risks — category-ceiling AOV, frequency-assumption stretch, density requirements, capex-per-geography — into a single read so the founder can see them together rather than encountering each one separately mid-build.
The output is a structured downloadable decision document. Not a chatbot to argue with. Not a course to work through. A read you take into a Saturday morning with coffee, and at the end of it you have a sharper sense of whether the math can work in your category, full stop.
That is the build/don't-build read. It is meant to pressure-test the premise before you commit the next 18 months of your life to executing against it.
The shape of the artifact
A founder mid-decision does not need another tool subscription. They need one good document that pressure-tests the build/don't-build read against public data and gets out of the way.
That is what DimeADozen.AI ships. $59 once. No subscription. Credits don't expire. 1 credit = 1 full validation report. A structured downloadable decision document, not a chatbot or a course. One-shot artifact built to be read, decided on, and acted on.
If you are about to commit to a category, run the pressure-test first. The math is already public. Someone just has to look at it honestly.
For the broader frame on why validation has to come before everything else, see the question every founder gets wrong.