Methodology
How we test + score outdoor gear.
Every review on this site is built from a four-dimension scoring rubric — Overall, Value, Durability, and Comfort — pulled together from a fixed set of public, attributable data sources. Here's exactly what each score means, how it's calculated, and what we deliberately won't do.
The four scores
What Overall, Value, Durability, and Comfort actually measure.
Overall
Weighted average of the three sub-scores plus a category-specific 'critical-flaw' modifier. A single deal-breaker (a rain shell that wets out, a stove that won't simmer) caps Overall at 6.5 no matter how the sub-scores look.
Inputs
Value
Price-vs-performance against the three closest peers in the same category and price tier. A 9.5 means category-best for the dollar; a 5.0 means you're paying for a brand more than the gear.
Inputs
Durability
Materials and construction quality, plus how the gear is reported to age over multiple seasons. Heavily category-dependent — a 7.5 on a 14oz running shoe means something different than a 7.5 on a 4lb hardshell tent.
Inputs
Comfort & fit
How the gear feels in use over a representative window for its category — a few hours for a stove, a full multi-day trip for a backpack or sleeping bag. We also score how forgiving the fit is to body type and edge cases.
Inputs
Awards on roundup pages
How "Best Overall," "Best Value," and the rest are assigned.
On every Best of roundup, the at-a-glance matrix awards four badges — one per scoring dimension. The leader in each dimension wins that badge, full stop. There is no manual override, no editor's-choice tiebreaker, no "Best Overall" award shuffled because of brand familiarity. If two products tie within 0.1, we award both and leave the call to the reader.
The point of the matrix is to make the scoring legible. If our "Best Value" pick only scores 6.2 — that means the whole category is mid right now, and the matrix will say so. We won't manufacture a 9.0 to make the page feel like an endorsement.
Data sources
Where the numbers come from.
- Manufacturer specs
- Independent editorial reviews
- Public trail reports
- Subreddit field intelligence
- Hands-on testing
- Public repair / failure logs
What we won't do
The hard rules.
- Use Amazon customer reviews or Amazon star ratings as a data source. The Associates Operating Agreement prohibits it, and crowd ratings are heavily skewed by sample bias and review-gating.
- Inflate a score because a brand sponsors a podcast, advertises elsewhere, or sends review samples. We don't accept review samples in exchange for guaranteed coverage.
- Hide a critical flaw in the prose and bury it in 'cons.' If a product fails its core job, the verdict says so first, before anything else.
- Recommend a product because it's the cheapest, when the cheapest is wrong. Value is price ÷ performance, not price alone.
- Use stock photography of a product and imply we have one in hand. Hero images are clearly editorial; manufacturer product photos are clearly labeled.