Reliability Lab

Benchmark Reports Hub

This hub publishes monthly antidetect benchmark reports with repeatable test protocol, evidence quality labels, and decision caveats so teams can compare changes over time, not just one-time snapshots.

Updated: 2026-04-05 | Cadence: monthly | Scope: lawful QA and reliability-led procurement.

Coverage Snapshot

What This Hub Tracks

22Platforms in baseline universe
MonthlyReport publication cadence
v1.2Current protocol version
3+Minimum repeated runs per critical flow

Latest Release

May 2026 Preview Report

The May preview extends the April baseline with early-month trend deltas, open blocker tracking, and pre-close publication gates under methodology v1.2.

Protocol Chain

How Reports Stay Reproducible

Step 1: test protocol and scoring weights are locked in methodology.
Step 2: protocol updates are versioned and published in changelog.
Step 3: each monthly benchmark references the active protocol version.
Step 4: correction notes and impact statements are logged when material changes appear.

How To Use

Use Reports by Decision Stage

Stage What to read first Decision output
Shortlisting Latest monthly benchmark summary and top drift risks 3-5 candidate shortlist by reliability profile
Pilot planning Evidence-level notes and no-buy criteria Pilot scope, success metrics, and rollback thresholds
Procurement Methodology version notes and recent corrections Commit or pause decision with audit-ready rationale

FAQ

Benchmark Hub Questions

How often are benchmark reports published?

Benchmark reports are published monthly, with additional urgent notes when ecosystem changes can invalidate a key assumption.

Can benchmark ranking be used as a direct buy signal?

No. Rankings are directional. Procurement decisions should follow your own repeated-session validation and no-buy criteria.

Where can I audit protocol changes over time?

Use the methodology changelog to review version updates, scoring changes, and impact notes.