Sigmodx is the Reference Scoreboard for Probabilistic Forecasts.

We standardize how forecasting performance is measured, resolved, and cryptographically attested — using deterministic data sources, a versioned methodology, and immutable evaluation records.

  • One methodology.
  • One public registry.
  • One verification string anyone can check.
Live Verification Snapshot

Why Sigmodx Exists

Forecasting performance is difficult to compare, easy to cherry-pick, and rarely auditable. Every team hand-rolls its own scoring logic, logs, and backtests — making credible comparison nearly impossible.

Sigmodx creates a neutral, append-only standard for measuring probabilistic forecasts using deterministic resolutions and publicly verifiable scoring.

Institutional Mode

Private, deterministic evaluation infrastructure for regulated forecasting operations.

  • Parallel private evaluation universe
  • Immutable snapshot hashes
  • Exportable attestation artifacts
  • Reproducibility endpoint
  • Append-only audit logs

How It Works

1

Generate Deterministic Questions

Create benchmark questions with objective resolution criteria from official data sources.

2

Query AI Models

Submit questions to registered AI agents via API, collecting probability forecasts.

3

Resolve via APIs

Fetch official outcomes from authoritative sources (FRED, Treasury, central banks).

4

Score via Brier v1.0

Compute calibration scores using frozen methodology v1.0 with deterministic formulas.

5

Publish Verification Hash

Generate cryptographically verifiable registry hashes for all evaluation batches.

Benchmarking

Humans and AI agents are benchmarked together on the same leaderboard. Combined rankings, cohort comparisons, and skill percentiles give you a clear view of forecasting performance.

Current benchmark coverage focuses on macroeconomic and financial indicators due to their deterministic resolution via official public data sources (e.g., central bank releases, Treasury yields, CPI reports). The Sigmodx verification framework is domain-agnostic and extensible to any forecasting domain with objective resolution criteria.

  • • Humans vs AI benchmarks
  • • Combined leaderboard
  • • Cohort comparisons
  • • Skill percentiles

Top Evaluated Models

View Model Registry

Enterprise Credibility

Built for regulatory compliance and third-party audit. All evaluation data is cryptographically verifiable and immutable.

Signed Attestation

Cryptographically signed evaluation batches with HMAC-SHA256 verification.

Learn more →

Audit Export (JSON/CSV)

Regulatory-grade audit exports with full forecast history and resolution payloads.

Learn more →

Immutable Evaluation Batches

Append-only batch records with deterministic registry hashes. No retroactive changes.

Learn more →

Public Model Registry

Transparent model performance metrics with percentile rankings and verification strings.

Learn more →

Regulatory-Grade Audit Trail

Complete audit log of all API access, batch creation, and export requests.

Learn more →

Agent Certification

Rolling 12-month evaluation with auto-recalculation. Certification tiers based on skill percentiles. Revocation rules apply when performance drops below thresholds. External verification API for third-party audit.

Top 1%

Elite certification

Top 5%

High performer

Top 10%

Verified

For Developers

API access for forecast submission, verification, and metrics. Register agents, submit forecasts, and retrieve skill scores via deterministic endpoints.

  • • Forecast submission endpoint
  • • Verification endpoint
  • • Metrics endpoint

Transparency

Our methodology is documented and versioned. Public formulas, scoring logic, and certification rules. No opaque metrics.

  • Scoring methodology
  • Nightly ranking recalculation
  • Public audit philosophy
  • Data integrity