EDORA

EDORA • Learn

Justice Index · Advocacy Lab · Field Guide

← Back to Learning Center

Methods

This page is our transparency pledge. It explains how EDORA assembles, normalizes, and interprets data for the Learning Center, dashboards, state briefs, and case studies.

Data Vintages & Provenance

  • Timeframe: core series 2015–2025, with recency emphasis in dashboards that indicate it.
  • Provenance: State agency reports, court admin data, OJJDP, HUD AHAR, CDC/KFF, SAMHSA, provider evaluations (see Sources).
  • Documentation: Field definitions live in Data Dictionary; copy-ready TSVs in the Data Appendix.

Normalization & Denominators

  • Youth population rates: most justice metrics normalized to ages 12–17.
  • School metrics: normalized to enrolled students; specify year and enrollment basis.
  • Clear denominators: each page states the denominator; dashboards include tooltips where relevant.

Small-n Pooling & Flags

Sparse counts create noisy rates. We pool 2–3 years where necessary and set pooled_small_n=true. Pages and charts annotate pooled values to prevent over-interpretation.

Disaggregation & Equity

  • Where data permit, we disaggregate by race/ethnicity, disability, gender, and rural vs. urban.
  • Access and outcomes are shown side-by-side to surface disparity patterns, not just averages.

Breaks in Series & Comparability

When definitions or systems change, we mark a break in series and avoid stitching incompatible trends. Notes appear on the page and in dashboard tooltips.

Composite Index: YARI+

  • Domains: education exclusion, detention/diversion, SUD/MH proxies, foster care, housing proxies, violence.
  • Standardization: domain metrics are standardized before aggregation (e.g., z-like scaling or 0–1 normalization—documented per dashboard).
  • Weights: domain weights are explicit; recency weights bias 2019–2025 more heavily (see example in YARI+ (AR)).
  • Missingness: if a domain is missing for a county-year, we either (a) impute within documented bounds with a flag, or (b) re-weight available domains—policy noted on the dashboard.

Outcomes & Follow-Up Windows

  • Initiation: enrolled/admitted to service.
  • Completion: program finished as planned.
  • Improvement: short-run markers (attendance/behavior, credits, service completion).
  • Follow-up: 30/90/180/365-day recurrence or persistence indicators, when feasible.
  • Caveat: Diversion/reentry outcomes can be selection-sensitive; we document eligibility rules and overrides.

Screening & Risk Tools

Intake screens (e.g., risk/needs, MH/SUD) are for triage, not punishment. We distinguish public safety risk from service need and treat overrides as auditable events. See Screening & Risk Tools.

Reproducibility & Versioning

  • Schemas: canonical fields in Data Dictionary.
  • Exports: use semantic versions in filenames (e.g., county_panel.v1.tsv).
  • Per-page notes: each article/state brief lists sources and data vintage; dashboards display vintage and flags.

Design & Accessibility

  • Charts label denominators, vintages, and pooling; legends avoid color-only cues when possible.
  • All figures have descriptive alt text; tables provide screen-reader friendly headers.

Where to Go Next

Transparency note: pooled or proxy values are flagged on page; breaks in series are annotated; equity cuts are shown where data quality allows. Questions or reproducibility requests? Reach us via Contact.